DeepSeek-R1 model weights

folder DeepSeek-R1 (163 files)
fileresolve/main/model-00163-of-000163.safetensors 6.58GB
fileresolve/main/model-00161-of-000163.safetensors 4.30GB
fileresolve/main/model-00162-of-000163.safetensors 4.30GB
fileresolve/main/model-00159-of-000163.safetensors 4.30GB
fileresolve/main/model-00160-of-000163.safetensors 5.23GB
fileresolve/main/model-00156-of-000163.safetensors 4.30GB
fileresolve/main/model-00157-of-000163.safetensors 4.30GB
fileresolve/main/model-00158-of-000163.safetensors 4.30GB
fileresolve/main/model-00152-of-000163.safetensors 4.30GB
fileresolve/main/model-00153-of-000163.safetensors 4.30GB
fileresolve/main/model-00154-of-000163.safetensors 4.30GB
fileresolve/main/model-00155-of-000163.safetensors 4.30GB
fileresolve/main/model-00150-of-000163.safetensors 4.30GB
fileresolve/main/model-00151-of-000163.safetensors 4.30GB
fileresolve/main/model-00149-of-000163.safetensors 4.30GB
fileresolve/main/model-00146-of-000163.safetensors 4.30GB
fileresolve/main/model-00147-of-000163.safetensors 4.30GB
fileresolve/main/model-00148-of-000163.safetensors 4.30GB
fileresolve/main/model-00143-of-000163.safetensors 4.30GB
fileresolve/main/model-00144-of-000163.safetensors 4.30GB
fileresolve/main/model-00145-of-000163.safetensors 4.30GB
fileresolve/main/model-00142-of-000163.safetensors 4.30GB
fileresolve/main/model-00140-of-000163.safetensors 4.30GB
fileresolve/main/model-00141-of-000163.safetensors 3.14GB
fileresolve/main/model-00139-of-000163.safetensors 4.30GB
fileresolve/main/model-00138-of-000163.safetensors 4.30GB
fileresolve/main/model-00136-of-000163.safetensors 4.30GB
fileresolve/main/model-00137-of-000163.safetensors 4.30GB
fileresolve/main/model-00134-of-000163.safetensors 4.30GB
fileresolve/main/model-00133-of-000163.safetensors 4.30GB
fileresolve/main/model-00135-of-000163.safetensors 4.30GB
fileresolve/main/model-00131-of-000163.safetensors 4.30GB
fileresolve/main/model-00132-of-000163.safetensors 4.30GB
fileresolve/main/model-00001-of-000163.safetensors 5.23GB
fileresolve/main/model-00002-of-000163.safetensors 4.30GB
fileresolve/main/model-00003-of-000163.safetensors 4.30GB
fileresolve/main/model-00004-of-000163.safetensors 4.30GB
fileresolve/main/model-00005-of-000163.safetensors 4.30GB
fileresolve/main/model-00006-of-000163.safetensors 4.37GB
fileresolve/main/model-00007-of-000163.safetensors 4.31GB
fileresolve/main/model-00008-of-000163.safetensors 4.30GB
fileresolve/main/model-00009-of-000163.safetensors 4.30GB
fileresolve/main/model-00010-of-000163.safetensors 4.30GB
fileresolve/main/model-00011-of-000163.safetensors 4.30GB
fileresolve/main/model-00012-of-000163.safetensors 1.32GB
fileresolve/main/model-00013-of-000163.safetensors 4.30GB
fileresolve/main/model-00014-of-000163.safetensors 4.30GB
fileresolve/main/model-00015-of-000163.safetensors 4.30GB
fileresolve/main/model-00016-of-000163.safetensors 4.30GB
Too many files! Click here to view them all.
Type: Dataset
Tags: weights, LLM

Bibtex:
@article{,
title= {DeepSeek-R1 model weights},
keywords= {LLM, weights},
author= {},
abstract= {Weights for DeepSeek-R1 from Huggingface

We introduce our first-generation reasoning models, DeepSeek-R1-Zero and DeepSeek-R1. DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors. However, DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. To address these issues and further enhance reasoning performance, we introduce DeepSeek-R1, which incorporates cold-start data before RL. DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

https://i.imgur.com/q6NKD6T.png

## License
This code repository and the model weights are licensed under the MIT License. DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:

DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from Qwen-2.5 series, which are originally licensed under Apache 2.0 License, and now finetuned with 800k samples curated with DeepSeek-R1.
DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under llama3.1 license.
DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under llama3.3 license.



```
@misc{deepseekai2025deepseekr1incentivizingreasoningcapability,
      title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning}, 
      author={DeepSeek-AI},
      year={2025},
      eprint={2501.12948},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.12948}, 
}
```
},
terms= {},
license= {https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE},
superseded= {},
url= {https://huggingface.co/deepseek-ai/DeepSeek-R1}
}

Hosted by users:

Send Feedback Start
   0.000006
DB Connect
   0.000357
Lookup hash in DB
   0.003346
Get torrent details
   0.002167
Get torrent details, finished
   0.000704
Get authors
   0.000007
Select authors
   0.001698
Parse bibtex
   0.000930
Write header
   0.000579
get stars
   0.000530
home tab
   0.014129
render right panel
   0.000016
render ads
   0.000072
fetch current hosters
   0.014244
Done