Hugging Face
huggingface.co › deepseek-ai › DeepSeek-R1
deepseek-ai/DeepSeek-R1 · Hugging Face
1 month ago - For all our models, the maximum generation length is set to 32,768 tokens. For benchmarks requiring sampling, we use a temperature of $0.6$, a top-p value of $0.95$, and generate 64 responses per query to estimate pass@1.
DeepSeek Deep Dive R1 at Home! - #422 by ubergarm - Machine Learning, LLMs, & AI - Level1Techs Forums
So I finally have a moment this morning after drinking some coffee to look at the fine print of that intel sglang benchmark results. To keep it simple let’s look only at the DeepSeek-R1-671B INT8 quant report: MODEL DATA TYPE SOCKETS llama.cpp TTFT (ms) llama.cpp TPOT (ms) SGLang TTFT (ms) ... More on forum.level1techs.com
DeepSeek-R1-0528 Official Benchmark
And they called it a 'minor improvement' More on reddit.com
Deepseek R1 (Ollama) Hardware benchmark for LocalLLM : LocalLLaMA
Deepseek R1 was released and looks like one of the best models for local LLM. I tested it on some GPUs to see how many tps it can achieve. Tests... More on old.reddit.com
Deepseek R1 is the only one that nails this new viral benchmark
Zero context here, what is this benchmark about? More on reddit.com
Videos
06:04
DeepSeek R1 0528 in 6 Minutes - YouTube
15:10
DeepSeek R1 Fully Tested - Insane Performance - YouTube
DeepSeek R1 BLOWS AWAY The Competition - How Did ...
10:10
DeepSeek R1 just got a HUGE Update! (o3 Level Model) - YouTube
12:50
New DeepSeek R1 is Really, Really Good Coder - YouTube
09:09
Deepseek-R1-0528: BEST Opensource Reasoning Model! Powerful, Fast, ...
Level1Techs
forum.level1techs.com › high-performance computing › machine learning, llms, & ai
DeepSeek Deep Dive R1 at Home! - #422 by ubergarm - Machine Learning, LLMs, & AI - Level1Techs Forums
July 30, 2025 - Ahh thanks, that one is easy enough then even if TTFT is not so useful without knowledge of the prompt/context size. So I finally have a moment this morning after drinking some coffee to look at the fine print of that intel sglang benchmark results. To keep it simple let’s look only at the DeepSeek-R1-671B INT8 quant report: MODEL DATA TYPE SOCKETS llama.cpp TTFT (ms) llama.cpp TPOT (ms) SGLang TTFT (ms) SGLang TPOT (ms) Speedup TTFT Speedup TPOT DeepSeek-R1-671B INT8 2 24546.76 172.0...
arXiv
arxiv.org › html › 2501.12948v1
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
October 13, 2025 - For engineering-related tasks, ... achieves outstanding results, significantly outperforming DeepSeek-V3 with scores of 90.8% on MMLU, 84.0% on MMLU-Pro, and 71.5% on GPQA Diamond....
TIME
time.com › tech › best inventions 2025 › deepseek r1: the best inventions of 2025
DeepSeek R1: The Best Inventions of 2025 | TIME
October 9, 2025 - The abrupt appearance of DeepSeek’s R1 advanced reasoning model at the start of the year was akin to the “shot heard ‘round the world” in AI circles. Major tech companies had spent recent years pouring billions into generative AI projects, products, and infrastructure. Meanwhile, Chinese startup DeepSeek created in just months a model as good as OpenAI’s then-most advanced product on industry-standard benchmarks...
GitHub
github.com › deepseek-ai › DeepSeek-R1
GitHub - deepseek-ai/DeepSeek-R1
To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, ...
Starred by 91.6K users
Forked by 11.8K users
GeekyAnts
geekyants.com › blog › deepseek-r1-vs-openais-o1-the-open-source-disruptor-raising-the-bar
DeepSeek-R1 vs. OpenAI’s o1: The Open-Source Disruptor Raising the Bar - GeekyAnts
January 26, 2025 - What’s jaw-dropping is that DeepSeek-R1 not only talks the talk with transparency, but it also walks the walk in terms of performance. DeepSeek-R1 surpasses OpenAI’s o1 in critical benchmarks—including the math-heavy AIME, the MATH-500 dataset, and coding challenges on Codeforces.
Eqbench
eqbench.com › creative_writing.html
EQ-Bench Creative Writing v3 Leaderboard
Emotional Intelligence Benchmarks for LLMs · Github | Paper | | Twitter | About
Reddit
reddit.com › r/localllama › deepseek-r1-0528 official benchmark
r/LocalLLaMA on Reddit: DeepSeek-R1-0528 Official Benchmark
April 6, 2025 -
Source:https://mp.weixin.qq.com/s/U5fnTRW4cGvXYJER__YBiw
Top answer 1 of 5
154
And they called it a 'minor improvement'
2 of 5
43
For those looking for GGUFs for the large R1 - I'm still doing them, but temporarily there is a prelim 2bit, 3bit and 4bit ones using our dynamic method which increases accuracy. https://huggingface.co/unsloth/DeepSeek-R1-0528-GGUF Remember to also use -ot ".ffn_.*_exps.=CPU" to offload MoE layers to RAM / disk - you can fit Q2_K_XL in under 24GB of VRAM!