Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Comparison Report: Companion Notebook 1 vs. DeepSeek Notebook

Date: 2026-02-16
Subject: Comparison of website/replication/companion-notebook-1.ipynb with Tom’s DeepSeek implementation tom/MMS_deepseek.ipynb
Paper: “Money as a Medium of Exchange in an Economy with Artificially Intelligent Agents,” Journal of Economic Dynamics and Control, 14, 329–373.


1. Faithfulness to the Paper

Our notebook closely follows the paper’s actual mechanisms:

DeepSeek’s notebook takes significant shortcuts:


2. Scale and Results

AspectOur NotebookDeepSeek
Agents per type50 (paper’s value)10
Periods (A1)1,000100
Periods (A1.2/GA)2,000
Economy BYesNo
Economy C (fiat money)YesNo
Classifier system per typeShared (as in paper)Per individual agent

Our notebook uses 50 agents per type and 1,000–2,000 periods as in the paper. DeepSeek uses only 10 agents and 100 periods, which is too short for meaningful convergence.


3. Architecture Decision: Shared vs. Individual Classifiers

Our notebook: All agents of the same type share one classifier system (as the paper specifies — “to economize on computation,” Section 5). This means 3 classifier systems total for 150 agents.

DeepSeek: Each of the 30 individual agents gets its own classifier system. This is 30× more classifier systems but doesn’t match the paper’s design. With only 10 agents per type and 100 periods, individual classifiers get too few learning episodes to converge.


4. Key Mechanism Differences

Bucket Brigade (strength updates)

Auction

Genetic Algorithm


5. Results Quality

Our notebook produces results matching the paper across all eight economies:

DeepSeek acknowledges its own limitations: “Average absolute difference” is reported but with only 100 periods and 10 agents, the results are noisy and don’t clearly demonstrate convergence.


6. Presentation and Exposition

DeepSeek does well in:

Our notebook does well in:


7. What We Could Learn from DeepSeek


Summary

Our notebook is a substantially more faithful and complete replication covering all eight economies from the paper (A1.1, A1.2, A2.1, A2.2, B.1, B.2, C, D), including the five-good Economy D which is the paper’s most complex demonstration. DeepSeek’s version is more of a sketch — it captures the general idea but takes shortcuts on every core mechanism (strength updates, GA, scale, shared classifiers) that materially affect whether the results actually replicate the paper’s findings.

The 10-agent, 100-period setup with pre-seeded “rational” initial strengths means the DeepSeek simulation is largely running on its initial conditions rather than demonstrating emergent learning, which is the entire point of the paper.