🌐
GitHub
github.com › deepseek-ai › DeepSeek-R1
GitHub - deepseek-ai/DeepSeek-R1
DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen.
Starred by 91.6K users
Forked by 11.8K users
🌐
Prime Intellect
primeintellect.ai › blog › synthetic-1-release
SYNTHETIC-1 Release: Two Million Collaboratively Generated Reasoning Traces from Deepseek-R1
We are releasing SYNTHETIC-1, the largest open reasoning dataset generated from Deepseek-R1, collaboratively generated by compute contributors across the globe.
🌐
Kaggle
kaggle.com › models › deepseek-ai › deepseek-r1
DeepSeek R1
Checking your browser before accessing www.kaggle.com · Click here if you are not automatically redirected after 5 seconds
🌐
GitHub
github.com › huggingface › open-r1
GitHub - huggingface/open-r1: Fully open reproduction of DeepSeek-R1
We release Mixture-of-Thoughts--a curated reasoning dataset of 350k verified traces distilled from R1. The dataset spans tasks in mathematics, coding, and science, and is designed to teach language models to reason step-by-step.
Starred by 25.7K users
Forked by 2.4K users
Languages   Python 89.4% | Shell 10.0% | Makefile 0.6%
🌐
arXiv
arxiv.org › pdf › 2501.12948 pdf
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via
dataset is evaluated using problems from 10 Div.2 contests along with expert-crafted test cases, after which the expected ratings and percentages of competitors are calculated. SWE-Bench · verified results are obtained via the agentless framework (Xia et al., 2024). AIDER-related · benchmarks are measured using a "diff" format. DeepSeek-R1 ...
🌐
MindSpore
mindspore.cn › mindformers › docs › en › master › example › distilled › distilled.html
Practice Case of Using DeepSeek-R1 for Model Distillation | MindSpore Transformers master documentation | MindSpore
If you want to generate a high-quality dataset, you are advised to refer to the dataset generation process in OpenR1-Math-220k. ... Deploy the DeepSeek-R1 inference service locally by referring to MindSpore-Lab/DeepSeek-R1 | Modelers or use the public API service.
🌐
Hugging Face
huggingface.co › blog › sdiazlor › fine-tune-deepseek-with-a-synthetic-reasoning-data
Fine-tune Deepseek-R1 with a Synthetic Reasoning Dataset
In this blog post, we used the Synthetic Data Generator to create a custom and high-quality synthetic reasoning dataset for solving Python coding problems with DeepSeek-R1-Distill-Qwen-32B. We then fine-tuned a smaller model, DeepSeek-R1-Distill-Qwen-1.5B, using this dataset.
🌐
Hugging Face
huggingface.co › blog › open-r1
Open-R1: a fully open reproduction of DeepSeek-R1
The release of DeepSeek-R1 is an amazing boon for the community, but they didn’t release everything—although the model weights are open, the datasets and code used to train the model are not 😢.
Find elsewhere
🌐
Camel-ai
camel-ai.org › blogs › distilling-math-reasoning-data-camel-ai
Distilling Mathematical Reasoning Data from DeepSeek R1 with CAMEL-AI
A step-by-step guide to generating high-quality mathematical reasoning datasets with CAMEL-AI and DeepSeek R1.
🌐
GitHub
github.com › FareedKhan-dev › train-deepseek-r1
GitHub - FareedKhan-dev/train-deepseek-r1: Building DeepSeek R1 from Scratch
Distillation takes the knowledge of a large, powerful “teacher” model (DeepSeek R1) and transfers it to smaller “student” models. Using a large dataset of reasoning examples, the outputs of DeepSeek R1 are used as the target answers.
Starred by 730 users
Forked by 118 users
Languages   Jupyter Notebook
🌐
Opencompass
doc.opencompass.org.cn › user_guides › deepseek_r1.html
Tutorial for Evaluating Reasoning Models — OpenCompass 0.5.1 documentation
dataset version metric mode deepseek-r1-distill-qwen-7b-turbomind ---------------------------------- --------- ------------- ------ --------------------------------------- MATH - - - AIME2024-Aveage8 - naive_average gen 56.25
🌐
DataCamp
datacamp.com › tutorial › fine-tuning-deepseek-r1-reasoning-model
Fine-Tuning DeepSeek R1 (Reasoning Model) | DataCamp
January 27, 2025 - In this tutorial, we will fine-tune the DeepSeek-R1-Distill-Llama-8B model on the Medical Chain-of-Thought Dataset from Hugging Face. This distilled DeepSeek-R1 model was created by fine-tuning the Llama 3.1 8B model on the data generated with DeepSeek-R1.
🌐
KDnuggets
kdnuggets.com › how-to-fine-tune-deepseek-r1-custom-dataset
How to Fine-Tune DeepSeek-R1 for Your Custom Dataset (Step-by-Step) - KDnuggets
By the end, you'll be able to fine-tune almost any large language model with a dataset of your choice. Before we begin, we need to install the Unsloth library along with its latest updates from GitHub. %�pture !pip install unsloth !pip install --force-reinstall --no-cache-dir --no-deps git+https://github.com/unslothai/unsloth.git · Now that Unsloth is installed, we can proceed to load our model and tokenizer. Now, we will load the DeepSeek model using Unsloth’s optimized methods. I am using the DeepSeek-R1-Distill-Llama-8B model.
🌐
Theriseunion
theriseunion.com › en › blog › DeepSeek-r1-models-intro.html
DeepSeek-R1 Model Series: From Light Distillation to Full-Scale - RiseUnion
Comprehensive analysis of DeepSeek-R1 model series, spanning from 1.5B to 671B parameters. Explore version-specific features, use cases, and selection guidelines to help users optimize performance-cost balance. Includes detailed comparison between distilled and full versions for enterprise ...
🌐
Hugging Face
huggingface.co › deepseek-ai › DeepSeek-R1
deepseek-ai/DeepSeek-R1 · Hugging Face
DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen.
🌐
NVIDIA
build.nvidia.com › deepseek-ai › deepseek-r1 › modelcard
deepseek-r1 Model by Deepseek-ai | NVIDIA NIM
DeepSeek-R1 achieves state-of-the-art results in various benchmarks and offers both its base models and distilled versions for community use.
🌐
Gitee
gitee.com › homer-1943 › train-deepseek-r1
FareedKhan-dev/train-deepseek-r1:
/homer-1943/train-deepseek-r1 · README · 0 Stars · 1 Watching · 0 Forks · Save · Cancel · No release · All · Jupyter Notebook 100.0% Load More · can not load any more · Edit · About · Homepage · Cancel Save · 马建仓 AI 助手 · 尝试更多 ·
🌐
MLCommons
mlcommons.org › home › deepseek reasoning for mlperf inference v5.1
DeepSeek Reasoning for MLPerf Inference v5.1 - MLCommons
September 9, 2025 - With mean input and output sequence lengths of 800 and 3,880 tokens, respectively, the DS-R1 dataset highlights the model’s ability in parsing proficiency, contextual linking, and synthesizing insights from complex and lengthy inputs. The performance metrics chosen for the DeepSeek-R1 benchmark ...
🌐
Markaicode
markaicode.com › home › ollama › how to fine-tune deepseek-r1 with custom datasets: advanced tutorial 2025
How to Fine-tune DeepSeek-R1 with Custom Datasets: Advanced Tutorial 2025 | Markaicode
June 23, 2025 - This comprehensive guide demonstrates how to fine-tune DeepSeek-R1 distilled models with custom datasets using memory-efficient techniques like LoRA and Unsloth.