Hugging Face
huggingface.co › deepseek-ai › DeepSeek-R1-Zero
deepseek-ai/DeepSeek-R1-Zero · Hugging Face
1 month ago - DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning.
arXiv
arxiv.org › abs › 2501.12948
[2501.12948] DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
January 22, 2025 - DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrates remarkable reasoning capabilities.
Deepseek R1 / R1 Zero
Open sourcing an o1 level model is incredible, already feared they might hide this beauty behind an api. More on reddit.com
An analysis of DeepSeek's R1-Zero and R1
While I think this is an interesting hypothesis, I'm skeptical. You might be lowering the cost of your training corpus by a few million dollars, but I highly doubt you are getting novel, high quality data · We are currently in a world where SOTA base model seems to be capped at around GPT4o levels. More on news.ycombinator.com
DeepSeek R1-zero
If I am reading the model summary on this page correctly this might be what you're looking for? https://huggingface.co/deepseek-ai/DeepSeek-R1-Zero "' reinforcement learning (RL) without supervised fine-tuning (SFT)" "This approach allows the model to explore chain-of-thought (CoT) for solving complex problems," More on reddit.com
[deleted by user]
Link the actual source don’t drop a context less img More on reddit.com
Videos
13:11
ArrrZero: Why DeepSeek R1 is less important than R1-Zero - YouTube
25:36
DeepSeek R1 Theory Overview | GRPO + RL + SFT - YouTube
03:31
DeepSeek-R1 vs DeepSeek-R1-Zero - YouTube
04:55
DeepSeek R1 Explained by AI Expert: How R1-Zero Led to an AI ...
01:07:43
DeepSeek R1 Theory Tutorial – Architecture, GRPO, KL Divergence ...
00:43
Mike Knoop on the difference between R1 and R1 zero - YouTube
Thelmbook
thelmbook.com › articles
DeepSeek R1 and R1-Zero Explained
This website requires Javascript to be enabled. Please turn on Javascript and reload the page
GitHub
github.com › deepseek-ai › DeepSeek-R1
GitHub - deepseek-ai/DeepSeek-R1
DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning.
Starred by 91.6K users
Forked by 11.8K users
Epoch AI
epoch.ai › gradient-updates › what-went-into-training-deepseek-r1
What went into training DeepSeek-R1? | Epoch AI
January 31, 2025 - The RL loop that produces R1-Zero is the core of the reasoning training, but it’s not the only step before the final R1 model is trained. Building on this checkpoint, DeepSeek curates a cold-start dataset (partly including cleaned up R1-Zero outputs) to fine-tune the base v3 model before ...
Nature
nature.com › articles › article
DeepSeek-R1 incentivizes reasoning in LLMs through reinforcement learning | Nature
September 17, 2025 - This design choice originates from our hypothesis that human-defined reasoning patterns may limit model exploration, whereas unrestricted RL training can better incentivize the emergence of new reasoning capabilities in LLMs. Through this process, detailed in the next section, our model (referred to as DeepSeek-R1-Zero) naturally developed diverse and sophisticated reasoning behaviours.
Wikipedia
en.wikipedia.org › wiki › DeepSeek
DeepSeek - Wikipedia
2 days ago - DeepSeek-R1-Zero was trained exclusively using GRPO RL without SFT. Unlike previous versions, it used no model-based reward. All reward functions were rule-based, "mainly" of two types (other types were not specified): accuracy rewards and format rewards. Accuracy reward was checking whether ...
Hacker News
news.ycombinator.com › item
An analysis of DeepSeek's R1-Zero and R1 | Hacker News
February 12, 2025 - While I think this is an interesting hypothesis, I'm skeptical. You might be lowering the cost of your training corpus by a few million dollars, but I highly doubt you are getting novel, high quality data · We are currently in a world where SOTA base model seems to be capped at around GPT4o levels.
BentoML
bentoml.com › blog › the-complete-guide-to-deepseek-models-from-v3-to-r1-and-beyond
The Complete Guide to DeepSeek Models: V3, R1, V3.1, V3.2 and Beyond
The result was R1, a model that not only keeps the reasoning power of R1-Zero but significantly improves accuracy, readability, and coherence. Unlike V3, which is optimized for general tasks, R1 is a true reasoning model. That means it doesn’t just give you an answer; it explains how it got there. Before responding, R1 generates a step-by-step chain of thought, making it especially useful for: ... According to the DeepSeek-R1 paper re-published in Nature and its supplementary information, R1’s training cost was the equivalent of just US$294K primarily on NVIDIA H800 chips.
DeepLearning.AI
deeplearning.ai › the-batch › deepseek-releases-r1-r1-zero-and-six-smaller-distilled-models
Data Points: DeepSeek releases R1, R1-Zero, and six smaller distilled models
January 20, 2025 - DeepSeek-R1 achieves performance comparable to OpenAI’s latest o1 model on reasoning tasks, including a 79.8 percent pass rate on AIME 2024 and 97.3 percent on MATH-500. The model, along with the reinforcement-learning-trained R1-Zero and smaller distilled versions, is now available under an MIT license, allowing open access for the community to use the model weights and outputs.
Sebastian Raschka
magazine.sebastianraschka.com › p › technical-deepseek
A Technical Tour of the DeepSeek Models from V3 to V3.2
3 weeks ago - While DeepSeek V3 wasn’t popular immediately upon release in December 2024, the DeepSeek R1 reasoning model (based on the identical architecture, using DeepSeek V3 as a base model) helped DeepSeek become one of the most popular open-weight models and a legit alternative to proprietary models ...
Towards AI
pub.towardsai.net › grpo-and-deepseek-r1-zero-9e81f15c6ba2
GRPO and DeepSeek-R1-Zero. 📚 Table of Contents | by Shakti Wadekar | Towards AI
March 15, 2025 - DeepSeek-R1-Zero is a cutting-edge large language model (LLM) developed by DeepSeek, a AI research company. What makes it stand out is its unique training method. While most language models rely heavily on supervised fine-tuning (SFT), where they learn from labeled datasets provided by humans, ...
Fptcloud
marketplace.fptcloud.com › en › ai-product › DeepSeek › DeepSeek-R1
DeepSeek-R1
A model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors.