$6M AI Model Matches OpenAI o1—Then Goes Open Source
DeepSeek-R1 achieves competitive reasoning performance against OpenAI o1 and Claude 3.5 Sonnet using pure reinforcement learning at a fraction of typical training costs. The Chinese AI lab open-sourced the model under MIT license, though security vulnerabilities and political bias issues present real limitations for production use.

DeepSeek-R1 achieves competitive reasoning performance against OpenAI's o1 and Claude 3.5 Sonnet using pure reinforcement learning without supervised fine-tuning—a different approach to training reasoning models. The Chinese AI lab trained it for roughly $6 million, then released it under MIT license.
Where OpenAI and Anthropic blend supervised fine-tuning with reinforcement learning, DeepSeek applied RL directly to base models. The model learned chain-of-thought reasoning, self-verification, and reflection from the training process itself. Developers now have an open alternative to commercial reasoning APIs, though production use requires working through known security and bias issues.
Pure RL Without SFT: A Different Path to Reasoning
DeepSeek's technical bet: skip supervised fine-tuning entirely. Apply reinforcement learning directly to base models, and the system learns to generate long chains of thought, verify its own reasoning, and reflect on mistakes—behaviors that emerge from training rather than being taught through curated examples.
This differs from how OpenAI and Anthropic have approached reasoning models. Their methods layer SFT atop base models before applying RL, a hybrid approach optimized for enterprise-scale deployment. DeepSeek's pure RL path trades some polish for openness, showing that alternative training methods can reach similar reasoning capabilities.
Performance at a Fraction of the Cost
The $6 million training cost is a fraction of what frontier labs typically spend on reasoning models. DeepSeek-R1 shows improvements in step consistency across multi-hop reasoning problems, better context recall over long inputs, and higher execution fidelity in code and mathematical formulas.
The model competes directly with o1 and Claude 3.5 Sonnet on standard reasoning benchmarks while maintaining cost advantages that make experimentation accessible to smaller teams. Fireworks AI's technical analysis confirms the pure RL approach delivers reasoning capabilities without requiring the supervised fine-tuning infrastructure that typically gates access to this class of model.
MIT License: What Open Source Reasoning Means
DeepSeek-R1-0528 is available on GitHub Models. Teams can fine-tune, self-host, or modify the model for specific domains—options that don't exist with commercial reasoning APIs.
The MIT license removes friction between experimenting with reasoning models and deploying them in production environments where data sovereignty or cost predictability matter. For research teams and startups building reasoning-heavy applications, this shifts reasoning models from a paid service to an open tool.
The Challenges: Readability and Early RL Issues
DeepSeek-R1-Zero, the initial pure RL version, encounters challenges including endless repetition, poor readability, and language mixing in reasoning tasks. The model learns to reason but doesn't automatically learn to present that reasoning clearly.
The subsequent R1 release addresses many of these rough edges, though traces remain. For teams evaluating the model, these readability quirks represent the current state of pure RL approaches. The reasoning works; the presentation needs refinement.
Security Vulnerabilities and Political Bias
DeepSeek-R1 shows high susceptibility to jailbreaking techniques including Evil Jailbreak, Crescendo, Deceptive Delight, and Bad Likert Judge. Research shows the model produces up to 50% more security vulnerabilities in code when prompted with politically sensitive topics related to Tibet, Uyghurs, or Falun Gong.
Political bias from red teaming on topics flagged by the Chinese government affects response patterns in ways that matter for teams deploying the model globally. These aren't theoretical concerns—they're measurable limitations that require architectural mitigation for production systems on R1.
Momentum and What Comes Next
The repository hit 91,849 stars shortly after its January 20, 2025 release, with Hacker News discussions comparing it favorably to Claude 3.5 and o1. DeepSeek briefly became the world's most popular AI search term on Google Trends, driven by the combination of low development cost and competitive performance.
For developers choosing between commercial and open reasoning models, DeepSeek-R1 proves the technical viability of pure RL approaches while highlighting the security and bias work still required to match the production readiness of established alternatives. The diversity of approaches expands what's possible in reasoning AI.