🌐
GitHub
github.com › deepseek-ai › DeepSeek-R1
GitHub - deepseek-ai/DeepSeek-R1
DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen.
Starred by 91.6K users
Forked by 11.8K users
🌐
Prime Intellect
primeintellect.ai › blog › synthetic-1-release
SYNTHETIC-1 Release: Two Million Collaboratively Generated Reasoning Traces from Deepseek-R1
We are releasing SYNTHETIC-1, the largest open reasoning dataset generated from Deepseek-R1, collaboratively generated by compute contributors across the globe.
🌐
Kaggle
kaggle.com › models › deepseek-ai › deepseek-r1
DeepSeek R1
Checking your browser before accessing www.kaggle.com · Click here if you are not automatically redirected after 5 seconds
🌐
GitHub
github.com › huggingface › open-r1
GitHub - huggingface/open-r1: Fully open reproduction of DeepSeek-R1
We release Mixture-of-Thoughts--a curated reasoning dataset of 350k verified traces distilled from R1. The dataset spans tasks in mathematics, coding, and science, and is designed to teach language models to reason step-by-step.
Starred by 25.7K users
Forked by 2.4K users
Languages   Python 89.4% | Shell 10.0% | Makefile 0.6%
🌐
arXiv
arxiv.org › pdf › 2501.12948 pdf
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via
dataset is evaluated using problems from 10 Div.2 contests along with expert-crafted test cases, after which the expected ratings and percentages of competitors are calculated. SWE-Bench · verified results are obtained via the agentless framework (Xia et al., 2024). AIDER-related · benchmarks are measured using a "diff" format. DeepSeek-R1 ...
🌐
Hugging Face
huggingface.co › blog › sdiazlor › fine-tune-deepseek-with-a-synthetic-reasoning-data
Fine-tune Deepseek-R1 with a Synthetic Reasoning Dataset
In this blog post, we used the Synthetic Data Generator to create a custom and high-quality synthetic reasoning dataset for solving Python coding problems with DeepSeek-R1-Distill-Qwen-32B. We then fine-tuned a smaller model, DeepSeek-R1-Distill-Qwen-1.5B, using this dataset.
🌐
Hugging Face
huggingface.co › blog › open-r1
Open-R1: a fully open reproduction of DeepSeek-R1
The release of DeepSeek-R1 is an amazing boon for the community, but they didn’t release everything—although the model weights are open, the datasets and code used to train the model are not 😢.
🌐
MindSpore
mindspore.cn › mindformers › docs › en › master › example › distilled › distilled.html
Practice Case of Using DeepSeek-R1 for Model Distillation | MindSpore Transformers master documentation | MindSpore
If you want to generate a high-quality dataset, you are advised to refer to the dataset generation process in OpenR1-Math-220k. ... Deploy the DeepSeek-R1 inference service locally by referring to MindSpore-Lab/DeepSeek-R1 | Modelers or use the public API service.
🌐
Hugging Face
huggingface.co › deepseek-ai › DeepSeek-R1
deepseek-ai/DeepSeek-R1 · Hugging Face
DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen.
Find elsewhere
🌐
Camel-ai
camel-ai.org › blogs › distilling-math-reasoning-data-camel-ai
Distilling Mathematical Reasoning Data from DeepSeek R1 with CAMEL-AI
A step-by-step guide to generating high-quality mathematical reasoning datasets with CAMEL-AI and DeepSeek R1.
🌐
GitHub
github.com › FareedKhan-dev › train-deepseek-r1
GitHub - FareedKhan-dev/train-deepseek-r1: Building DeepSeek R1 from Scratch
Distillation takes the knowledge of a large, powerful “teacher” model (DeepSeek R1) and transfers it to smaller “student” models. Using a large dataset of reasoning examples, the outputs of DeepSeek R1 are used as the target answers.
Starred by 730 users
Forked by 118 users
Languages   Jupyter Notebook
🌐
DataCamp
datacamp.com › tutorial › fine-tuning-deepseek-r1-reasoning-model
Fine-Tuning DeepSeek R1 (Reasoning Model) | DataCamp
January 27, 2025 - In this tutorial, we will fine-tune the DeepSeek-R1-Distill-Llama-8B model on the Medical Chain-of-Thought Dataset from Hugging Face. This distilled DeepSeek-R1 model was created by fine-tuning the Llama 3.1 8B model on the data generated with DeepSeek-R1.
🌐
KDnuggets
kdnuggets.com › how-to-fine-tune-deepseek-r1-custom-dataset
How to Fine-Tune DeepSeek-R1 for Your Custom Dataset (Step-by-Step) - KDnuggets
By the end, you'll be able to fine-tune almost any large language model with a dataset of your choice. Before we begin, we need to install the Unsloth library along with its latest updates from GitHub. %�pture !pip install unsloth !pip install --force-reinstall --no-cache-dir --no-deps git+https://github.com/unslothai/unsloth.git · Now that Unsloth is installed, we can proceed to load our model and tokenizer. Now, we will load the DeepSeek model using Unsloth’s optimized methods. I am using the DeepSeek-R1-Distill-Llama-8B model.
🌐
NVIDIA
build.nvidia.com › deepseek-ai › deepseek-r1 › modelcard
deepseek-r1 Model by Deepseek-ai | NVIDIA NIM
DeepSeek-R1 achieves state-of-the-art results in various benchmarks and offers both its base models and distilled versions for community use.
🌐
Epoch AI
epoch.ai › gradient-updates › what-went-into-training-deepseek-r1
What went into training DeepSeek-R1? | Epoch AI
January 31, 2025 - Once this is done, DeepSeek creates a supervised fine-tuning dataset of around 600K reasoning samples from this last checkpoint and 200K samples that went into the post-training of v3 itself before fine-tuning v3-base on all of this data for two epochs. If the average length per sample is around 8K, which is in line with what R1-Zero achieves by the end of training, this SFT dataset has a total of 800K * 8K = 6.4B tokens and so the fine-tuning itself has a negligible cost.
🌐
Milvus
milvus.io › ai-quick-reference › what-is-the-training-dataset-size-for-deepseeks-r1-model
What is the training dataset size for DeepSeek's R1 model?
The exact training dataset size for DeepSeek’s R1 model has not been publicly disclosed by the developers. While specific numbers are unavailable, the scale of datasets for state-of-the-art language models typically ranges from hundreds of billions to trillions of tokens.
🌐
Opencompass
doc.opencompass.org.cn › user_guides › deepseek_r1.html
Tutorial for Evaluating Reasoning Models — OpenCompass 0.5.1 documentation
dataset version metric mode deepseek-r1-distill-qwen-7b-turbomind ---------------------------------- --------- ------------- ------ --------------------------------------- MATH - - - AIME2024-Aveage8 - naive_average gen 56.25
🌐
Firecrawl
firecrawl.dev › blog › fine-tuning-deepseek
Fine-tuning DeepSeek R1 on a Custom Instructions Dataset
Learn how to fine-tune DeepSeek R1 language models using custom instruction datasets.
🌐
Gitee
gitee.com › homer-1943 › train-deepseek-r1
FareedKhan-dev/train-deepseek-r1:
/homer-1943/train-deepseek-r1 · README · 0 Stars · 1 Watching · 0 Forks · Save · Cancel · No release · All · Jupyter Notebook 100.0% Load More · can not load any more · Edit · About · Homepage · Cancel Save · 马建仓 AI 助手 · 尝试更多 ·