DeepSeek-R1 model weights

folder DeepSeek-R1 (163 files)
fileresolve/main/model-00163-of-000163.safetensors 6.58GB
fileresolve/main/model-00161-of-000163.safetensors 4.30GB
fileresolve/main/model-00162-of-000163.safetensors 4.30GB
fileresolve/main/model-00159-of-000163.safetensors 4.30GB
fileresolve/main/model-00160-of-000163.safetensors 5.23GB
fileresolve/main/model-00156-of-000163.safetensors 4.30GB
fileresolve/main/model-00157-of-000163.safetensors 4.30GB
fileresolve/main/model-00158-of-000163.safetensors 4.30GB
fileresolve/main/model-00152-of-000163.safetensors 4.30GB
fileresolve/main/model-00153-of-000163.safetensors 4.30GB
fileresolve/main/model-00154-of-000163.safetensors 4.30GB
fileresolve/main/model-00155-of-000163.safetensors 4.30GB
fileresolve/main/model-00150-of-000163.safetensors 4.30GB
fileresolve/main/model-00151-of-000163.safetensors 4.30GB
fileresolve/main/model-00149-of-000163.safetensors 4.30GB
fileresolve/main/model-00146-of-000163.safetensors 4.30GB
fileresolve/main/model-00147-of-000163.safetensors 4.30GB
fileresolve/main/model-00148-of-000163.safetensors 4.30GB
fileresolve/main/model-00143-of-000163.safetensors 4.30GB
fileresolve/main/model-00144-of-000163.safetensors 4.30GB
fileresolve/main/model-00145-of-000163.safetensors 4.30GB
fileresolve/main/model-00142-of-000163.safetensors 4.30GB
fileresolve/main/model-00140-of-000163.safetensors 4.30GB
fileresolve/main/model-00141-of-000163.safetensors 3.14GB
fileresolve/main/model-00139-of-000163.safetensors 4.30GB
fileresolve/main/model-00138-of-000163.safetensors 4.30GB
fileresolve/main/model-00136-of-000163.safetensors 4.30GB
fileresolve/main/model-00137-of-000163.safetensors 4.30GB
fileresolve/main/model-00134-of-000163.safetensors 4.30GB
fileresolve/main/model-00133-of-000163.safetensors 4.30GB
fileresolve/main/model-00135-of-000163.safetensors 4.30GB
fileresolve/main/model-00131-of-000163.safetensors 4.30GB
fileresolve/main/model-00132-of-000163.safetensors 4.30GB
fileresolve/main/model-00001-of-000163.safetensors 5.23GB
fileresolve/main/model-00002-of-000163.safetensors 4.30GB
fileresolve/main/model-00003-of-000163.safetensors 4.30GB
fileresolve/main/model-00004-of-000163.safetensors 4.30GB
fileresolve/main/model-00005-of-000163.safetensors 4.30GB
fileresolve/main/model-00006-of-000163.safetensors 4.37GB
fileresolve/main/model-00007-of-000163.safetensors 4.31GB
fileresolve/main/model-00008-of-000163.safetensors 4.30GB
fileresolve/main/model-00009-of-000163.safetensors 4.30GB
fileresolve/main/model-00010-of-000163.safetensors 4.30GB
fileresolve/main/model-00011-of-000163.safetensors 4.30GB
fileresolve/main/model-00012-of-000163.safetensors 1.32GB
fileresolve/main/model-00013-of-000163.safetensors 4.30GB
fileresolve/main/model-00014-of-000163.safetensors 4.30GB
fileresolve/main/model-00015-of-000163.safetensors 4.30GB
fileresolve/main/model-00016-of-000163.safetensors 4.30GB
Too many files! Click here to view them all.
Type: Dataset
Tags: weightsLLM

Bibtex:
@article{,
title= {DeepSeek-R1 model weights},
keywords= {LLM, weights},
author= {},
abstract= {Weights for DeepSeek-R1 from Huggingface

We introduce our first-generation reasoning models, DeepSeek-R1-Zero and DeepSeek-R1. DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors. However, DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. To address these issues and further enhance reasoning performance, we introduce DeepSeek-R1, which incorporates cold-start data before RL. DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

https://i.imgur.com/q6NKD6T.png

## License
This code repository and the model weights are licensed under the MIT License. DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:

DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from Qwen-2.5 series, which are originally licensed under Apache 2.0 License, and now finetuned with 800k samples curated with DeepSeek-R1.
DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under llama3.1 license.
DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under llama3.3 license.



```
@misc{deepseekai2025deepseekr1incentivizingreasoningcapability,
      title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning}, 
      author={DeepSeek-AI},
      year={2025},
      eprint={2501.12948},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.12948}, 
}
```
},
terms= {},
license= {https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE},
superseded= {},
url= {https://huggingface.co/deepseek-ai/DeepSeek-R1}
}

Hosted by users
No stats to report yet.

Send Feedback Start
   0.000005
DB Connect
   0.000406
Lookup hash in DB
   0.000369
Get torrent details
   0.000128
Get torrent details, finished
   0.000197
Get authors
   0.000001
Select authors
   0.000171
Parse bibtex
   0.000106
Write header
   0.000172
get stars
   0.000091
home tab
   0.000456
render right panel
   0.000006
render ads
   0.000340
fetch current hosters
   0.000263
Start get stats
   0.000337
End get stats
   0.000001
related datasets
   0.002921
Done