DeepSeek-R1 model weights

folder DeepSeek-R1 (163 files)
fileresolve/main/model-00163-of-000163.safetensors 6.58GB
fileresolve/main/model-00161-of-000163.safetensors 4.30GB
fileresolve/main/model-00162-of-000163.safetensors 4.30GB
fileresolve/main/model-00159-of-000163.safetensors 4.30GB
fileresolve/main/model-00160-of-000163.safetensors 5.23GB
fileresolve/main/model-00156-of-000163.safetensors 4.30GB
fileresolve/main/model-00157-of-000163.safetensors 4.30GB
fileresolve/main/model-00158-of-000163.safetensors 4.30GB
fileresolve/main/model-00152-of-000163.safetensors 4.30GB
fileresolve/main/model-00153-of-000163.safetensors 4.30GB
fileresolve/main/model-00154-of-000163.safetensors 4.30GB
fileresolve/main/model-00155-of-000163.safetensors 4.30GB
fileresolve/main/model-00150-of-000163.safetensors 4.30GB
fileresolve/main/model-00151-of-000163.safetensors 4.30GB
fileresolve/main/model-00149-of-000163.safetensors 4.30GB
fileresolve/main/model-00146-of-000163.safetensors 4.30GB
fileresolve/main/model-00147-of-000163.safetensors 4.30GB
fileresolve/main/model-00148-of-000163.safetensors 4.30GB
fileresolve/main/model-00143-of-000163.safetensors 4.30GB
fileresolve/main/model-00144-of-000163.safetensors 4.30GB
fileresolve/main/model-00145-of-000163.safetensors 4.30GB
fileresolve/main/model-00142-of-000163.safetensors 4.30GB
fileresolve/main/model-00140-of-000163.safetensors 4.30GB
fileresolve/main/model-00141-of-000163.safetensors 3.14GB
fileresolve/main/model-00139-of-000163.safetensors 4.30GB
fileresolve/main/model-00138-of-000163.safetensors 4.30GB
fileresolve/main/model-00136-of-000163.safetensors 4.30GB
fileresolve/main/model-00137-of-000163.safetensors 4.30GB
fileresolve/main/model-00134-of-000163.safetensors 4.30GB
fileresolve/main/model-00133-of-000163.safetensors 4.30GB
fileresolve/main/model-00135-of-000163.safetensors 4.30GB
fileresolve/main/model-00131-of-000163.safetensors 4.30GB
fileresolve/main/model-00132-of-000163.safetensors 4.30GB
fileresolve/main/model-00001-of-000163.safetensors 5.23GB
fileresolve/main/model-00002-of-000163.safetensors 4.30GB
fileresolve/main/model-00003-of-000163.safetensors 4.30GB
fileresolve/main/model-00004-of-000163.safetensors 4.30GB
fileresolve/main/model-00005-of-000163.safetensors 4.30GB
fileresolve/main/model-00006-of-000163.safetensors 4.37GB
fileresolve/main/model-00007-of-000163.safetensors 4.31GB
fileresolve/main/model-00008-of-000163.safetensors 4.30GB
fileresolve/main/model-00009-of-000163.safetensors 4.30GB
fileresolve/main/model-00010-of-000163.safetensors 4.30GB
fileresolve/main/model-00011-of-000163.safetensors 4.30GB
fileresolve/main/model-00012-of-000163.safetensors 1.32GB
fileresolve/main/model-00013-of-000163.safetensors 4.30GB
fileresolve/main/model-00014-of-000163.safetensors 4.30GB
fileresolve/main/model-00015-of-000163.safetensors 4.30GB
fileresolve/main/model-00016-of-000163.safetensors 4.30GB
Too many files! Click here to view them all.
Type: Dataset
Tags: weights, LLM

Bibtex:
@article{,
title= {DeepSeek-R1 model weights},
keywords= {LLM, weights},
author= {},
abstract= {Weights for DeepSeek-R1 from Huggingface

We introduce our first-generation reasoning models, DeepSeek-R1-Zero and DeepSeek-R1. DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors. However, DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. To address these issues and further enhance reasoning performance, we introduce DeepSeek-R1, which incorporates cold-start data before RL. DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

https://i.imgur.com/q6NKD6T.png

## License
This code repository and the model weights are licensed under the MIT License. DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:

DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from Qwen-2.5 series, which are originally licensed under Apache 2.0 License, and now finetuned with 800k samples curated with DeepSeek-R1.
DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under llama3.1 license.
DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under llama3.3 license.



```
@misc{deepseekai2025deepseekr1incentivizingreasoningcapability,
      title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning}, 
      author={DeepSeek-AI},
      year={2025},
      eprint={2501.12948},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.12948}, 
}
```
},
terms= {},
license= {https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE},
superseded= {},
url= {https://huggingface.co/deepseek-ai/DeepSeek-R1}
}

Hosted by users:

Send Feedback Start
   0.000005
DB Connect
   0.000359
Lookup hash in DB
   0.003405
Get torrent details
   0.000575
Get torrent details, finished
   0.000565
Get authors
   0.000005
Select authors
   0.000419
Parse bibtex
   0.000799
Write header
   0.000492
get stars
   0.000381
home tab
   0.012442
render right panel
   0.000015
render ads
   0.000056
fetch current hosters
   0.009482
Done