Step 1: Acquire ~25 4090's Step 2: Network them together Step 3: Start homebrew nuclear cold-fusion reactor Step 4: Plug into cluster Answer from tcarambat on reddit.com
🌐
Hugging Face
huggingface.co › deepseek-ai › DeepSeek-R1
deepseek-ai/DeepSeek-R1 · Hugging Face
1 month ago - DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. NOTE: Before running DeepSeek-R1 series models locally, we kindly recommend reviewing the Usage Recommendation section.
People also ask

Is DeepSeek-R1 open source?
Yes, DeepSeek is open source in that its model weights and training methods are freely available for the public to examine, use and build upon. However, its source code and any specifics about its underlying data are not available to the public.
🌐
builtin.com
builtin.com › artificial-intelligence › deepseek-r1
What Is DeepSeek-R1? | Built In
How to access DeepSeek-R1
DeepSeek’s chatbot (which can be powered by the R1 model) is free to use on the company’s website and is available for download on the Apple App Store. R1 is also available for use on Hugging Face and DeepSeek’s API.
🌐
builtin.com
builtin.com › artificial-intelligence › deepseek-r1
What Is DeepSeek-R1? | Built In
How many parameters does DeepSeek-R1 have?
DeepSeek-R1 has 671 billion parameters in total. But DeepSeek also released six “distilled” versions of R1, ranging in size from 1.5 billion parameters to 70 billion parameters. While the smallest can run on a laptop with consumer GPUs, the full R1 requires more substantial hardware.
🌐
builtin.com
builtin.com › artificial-intelligence › deepseek-r1
What Is DeepSeek-R1? | Built In
🌐
BentoML
bentoml.com › blog › the-complete-guide-to-deepseek-models-from-v3-to-r1-and-beyond
The Complete Guide to DeepSeek Models: V3, R1, V3.1, V3.2 and Beyond
Instead of training new models from scratch, DeepSeek took a smart shortcut: Started with 6 open-source models from Llama 3.1/3.3 and Qwen 2.5 · Generated 800,000 high-quality reasoning samples using R1 · Fine-tuned the smaller models on these synthetic reasoning data · Unlike R1, these distilled models rely solely on SFT and they do not include an RL stage. Despite their smaller size, these models perform remarkably well on reasoning tasks, proving that large-scale AI reasoning can be efficiently distilled.
🌐
Ollama
ollama.com › library › deepseek-r1
deepseek-r1
DeepSeek-R1 has received a minor version upgrade to DeepSeek-R1-0528 for the 8 billion parameter distilled model and the full 671 billion parameter model. In this update, DeepSeek R1 has significantly improved its reasoning and inference ...
🌐
Multiverse Computing
multiversecomputing.com › resources › deepseek-r1-uncensored-full-power-fraction-of-the-size
DeepSeek R1 Uncensored: Full Power, Fraction of the Size
November 19, 2025 - DeepSeek R1 Slim by CompactifAI has 300 billion fewer parameters than the original model, directly halving memory consumption and deployment costs while still maintaining accuracy on all deep reasoning tasks.
🌐
Fireworks AI
fireworks.ai › blog › deepseek-r1-deepdive
DeepSeek-R1 Overview: Features, Capabilities, Parameters
Operational expenses are estimated ... spend on OpenAI’s o1 model**.** Cost of running DeepSeek R1 on Fireworks AI is $8/ 1 M token (both input & output), whereas, running OpenAI o1 model costs $15/ 1M input tokens and $60/ ...
Find elsewhere
🌐
Built In
builtin.com › artificial-intelligence › deepseek-r1
What Is DeepSeek-R1? | Built In
DeepSeek-R1 has 671 billion parameters in total. But DeepSeek also released six “distilled” versions of R1, ranging in size from 1.5 billion parameters to 70 billion parameters.
Published   October 6, 2025
🌐
DeepSeek
api-docs.deepseek.com › models & pricing
Models & Pricing | DeepSeek API Docs
The prices listed below are in units of per 1M tokens. A token, the smallest unit of text that the model recognizes, can be a word, a number, or even a punctuation mark. We will bill based on the total number of input and output tokens by the model.
🌐
GeeksforGeeks
geeksforgeeks.org › websites & apps › deepseek-r1-vs-deepseek-v3
DeepSeek R1 vs DeepSeek V3: Benchmarking Speed, Accuracy, and Scalability - GeeksforGeeks
July 23, 2025 - Parameter Size: 671 billion parameters with 37 billion activated per token, thanks to its Mixture-of-Experts (MoE) architecture. Best For: Organizations and researchers needing robust, versatile AI across tasks.
🌐
Tenable®
tenable.com › blog › frequently-asked-questions-about-deepseek-large-language-model-llm-v3-r1
Frequently Asked Questions About DeepSeek Large Language Model (LLM) - Blog | Tenable®
August 5, 2025 - DeepSeek R1 has 671 billion parameters and requires multiple expensive high-end GPUs to run. There are distilled versions of the model starting at 1.5 billion parameters, going all the way up to 70 billion parameters.
🌐
Substack
simonw.substack.com › p › the-deepseek-r1-family-of-reasoning
The DeepSeek-R1 family of reasoning models
January 20, 2025 - It's over 650GB in size and, like most of their other releases, is under a clean MIT license. DeepSeek warn that "DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing."
🌐
APXML
apxml.com › posts › gpu-requirements-deepseek-r1
GPU System Requirements for Running DeepSeek-R1
GPU system requirements to run DeepSeek-R1 and its distilled models effectively, along with recommendations for choosing the right hardware for your needs.
🌐
Medium
medium.com › @alice.yang_10652 › how-to-choose-the-right-version-of-deepseek-r1-for-local-deployment-read-here-b24f4d0ec6cc
How to Choose the Right Version of DeepSeek-R1 for Local Deployment? Read Here! | by Alice Yang | Medium
February 11, 2025 - Selecting the right DeepSeek-R1 version depends on your available hardware and intended use case. If you’re working with basic text applications, a 1.5B or 7B model is sufficient. For advanced reasoning and text generation, consider 14B or larger.
🌐
Snowkylin
snowkylin.github.io › blogs › a-note-on-deepseek-r1.html
A Note on DeepSeek R1 Deployment
The original DeepSeek R1 671B model is 720GB in size, which is huge. Even a $200k monster like NVIDIA DGX H100 (with 8xH100) can barely hold it. Here I use Unsloth AI’s dynamically quantized version, which selectively quantize a few important layers to higher bits, while leaving most of the ...
🌐
Ollama
ollama.com › library › deepseek-r1 › blobs › 96c415656d37
deepseek-r1/model
deepseek-r1:latest · 71.3M Downloads Updated 4 months ago · tools thinking 1.5b 7b 8b 14b 32b 70b 671b · deepseek-r1:latest ... / model · 96c415656d37 · 4.7GB · Metadata · general.architecture · qwen2 · qwen2 · general.file_type · ...
🌐
Hugging Face
huggingface.co › deepseek-ai › DeepSeek-R1 › discussions › 19
deepseek-ai/DeepSeek-R1 · Hardware requirements?
vllm serve deepseek-ai/DeepSeek-R1-Distill-Qwen-32B --tensor-parallel-size 2 --max-model-len 32768 --enforce-eager
🌐
DeepSeek
api-docs.deepseek.com › deepseek-r1 release 2025/01/20
DeepSeek-R1 Release | DeepSeek API Docs
🔬 Distilled from DeepSeek-R1, 6 small models fully open-sourced · 📏 32B & 70B models on par with OpenAI-o1-mini · 🤝 Empowering the open-source community · 🌍 Pushing the boundaries of open AI!
🌐
Hugging Face
huggingface.co › deepseek-ai › DeepSeek-V3
deepseek-ai/DeepSeek-V3 · Hugging Face
1 month ago - Meanwhile, we also maintain a control over the output style and length of DeepSeek-V3. NOTE: The total size of DeepSeek-V3 models on HuggingFace is 685B, which includes 671B of the Main Model weights and 14B of the Multi-Token Prediction (MTP) ...