GitHub
github.com › deepseek-ai › DeepSeek-R1
GitHub - deepseek-ai/DeepSeek-R1
DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen.
Starred by 91.6K users
Forked by 11.8K users
Videos
11:52
Create Reasoning Dataset with DeepSeek R1 and Camel - YouTube
01:22:24
Gen AI Project | Log Classification System Using Deepseek R1 LLM, ...
19:31
Fine Tune DeepSeek Model on your Custom Dataset 🔥🚀 - YouTube
08:31
Create Synthetic Data to Train DeepSeek R1! (Reasoning Steps) - ...
00:55
What data is needed to train an AI model like Deepseek R1? #ai ...
DeepSeek-R1 Paper Explained - A New RL LLMs Era in AI?
Kaggle
kaggle.com › models › deepseek-ai › deepseek-r1
DeepSeek R1
Checking your browser before accessing www.kaggle.com · Click here if you are not automatically redirected after 5 seconds
GitHub
github.com › huggingface › open-r1
GitHub - huggingface/open-r1: Fully open reproduction of DeepSeek-R1
We release Mixture-of-Thoughts--a curated reasoning dataset of 350k verified traces distilled from R1. The dataset spans tasks in mathematics, coding, and science, and is designed to teach language models to reason step-by-step.
Starred by 25.7K users
Forked by 2.4K users
Languages Python 89.4% | Shell 10.0% | Makefile 0.6%
arXiv
arxiv.org › pdf › 2501.12948 pdf
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via
dataset is evaluated using problems from 10 Div.2 contests along with expert-crafted test cases, after which the expected ratings and percentages of competitors are calculated. SWE-Bench · verified results are obtained via the agentless framework (Xia et al., 2024). AIDER-related · benchmarks are measured using a "diff" format. DeepSeek-R1 ...
MindSpore
mindspore.cn › mindformers › docs › en › master › example › distilled › distilled.html
Practice Case of Using DeepSeek-R1 for Model Distillation | MindSpore Transformers master documentation | MindSpore
If you want to generate a high-quality dataset, you are advised to refer to the dataset generation process in OpenR1-Math-220k. ... Deploy the DeepSeek-R1 inference service locally by referring to MindSpore-Lab/DeepSeek-R1 | Modelers or use the public API service.
GitHub
github.com › FareedKhan-dev › train-deepseek-r1
GitHub - FareedKhan-dev/train-deepseek-r1: Building DeepSeek R1 from Scratch
Distillation takes the knowledge of a large, powerful “teacher” model (DeepSeek R1) and transfers it to smaller “student” models. Using a large dataset of reasoning examples, the outputs of DeepSeek R1 are used as the target answers.
Starred by 730 users
Forked by 118 users
Languages Jupyter Notebook
Opencompass
doc.opencompass.org.cn › user_guides › deepseek_r1.html
Tutorial for Evaluating Reasoning Models — OpenCompass 0.5.1 documentation
dataset version metric mode deepseek-r1-distill-qwen-7b-turbomind ---------------------------------- --------- ------------- ------ --------------------------------------- MATH - - - AIME2024-Aveage8 - naive_average gen 56.25
DataCamp
datacamp.com › tutorial › fine-tuning-deepseek-r1-reasoning-model
Fine-Tuning DeepSeek R1 (Reasoning Model) | DataCamp
January 27, 2025 - In this tutorial, we will fine-tune the DeepSeek-R1-Distill-Llama-8B model on the Medical Chain-of-Thought Dataset from Hugging Face. This distilled DeepSeek-R1 model was created by fine-tuning the Llama 3.1 8B model on the data generated with DeepSeek-R1.
KDnuggets
kdnuggets.com › how-to-fine-tune-deepseek-r1-custom-dataset
How to Fine-Tune DeepSeek-R1 for Your Custom Dataset (Step-by-Step) - KDnuggets
By the end, you'll be able to fine-tune almost any large language model with a dataset of your choice. Before we begin, we need to install the Unsloth library along with its latest updates from GitHub. %�pture !pip install unsloth !pip install --force-reinstall --no-cache-dir --no-deps git+https://github.com/unslothai/unsloth.git · Now that Unsloth is installed, we can proceed to load our model and tokenizer. Now, we will load the DeepSeek model using Unsloth’s optimized methods. I am using the DeepSeek-R1-Distill-Llama-8B model.
Theriseunion
theriseunion.com › en › blog › DeepSeek-r1-models-intro.html
DeepSeek-R1 Model Series: From Light Distillation to Full-Scale - RiseUnion
Comprehensive analysis of DeepSeek-R1 model series, spanning from 1.5B to 671B parameters. Explore version-specific features, use cases, and selection guidelines to help users optimize performance-cost balance. Includes detailed comparison between distilled and full versions for enterprise ...
MLCommons
mlcommons.org › home › deepseek reasoning for mlperf inference v5.1
DeepSeek Reasoning for MLPerf Inference v5.1 - MLCommons
September 9, 2025 - With mean input and output sequence lengths of 800 and 3,880 tokens, respectively, the DS-R1 dataset highlights the model’s ability in parsing proficiency, contextual linking, and synthesizing insights from complex and lengthy inputs. The performance metrics chosen for the DeepSeek-R1 benchmark ...
Markaicode
markaicode.com › home › ollama › how to fine-tune deepseek-r1 with custom datasets: advanced tutorial 2025
How to Fine-tune DeepSeek-R1 with Custom Datasets: Advanced Tutorial 2025 | Markaicode
June 23, 2025 - This comprehensive guide demonstrates how to fine-tune DeepSeek-R1 distilled models with custom datasets using memory-efficient techniques like LoRA and Unsloth.