DeepSeek-R1 model weights

folder DeepSeek-R1 (163 files)
fileresolve/main/model-00163-of-000163.safetensors 6.58GB
fileresolve/main/model-00161-of-000163.safetensors 4.30GB
fileresolve/main/model-00162-of-000163.safetensors 4.30GB
fileresolve/main/model-00159-of-000163.safetensors 4.30GB
fileresolve/main/model-00160-of-000163.safetensors 5.23GB
fileresolve/main/model-00156-of-000163.safetensors 4.30GB
fileresolve/main/model-00157-of-000163.safetensors 4.30GB
fileresolve/main/model-00158-of-000163.safetensors 4.30GB
fileresolve/main/model-00152-of-000163.safetensors 4.30GB
fileresolve/main/model-00153-of-000163.safetensors 4.30GB
fileresolve/main/model-00154-of-000163.safetensors 4.30GB
fileresolve/main/model-00155-of-000163.safetensors 4.30GB
fileresolve/main/model-00150-of-000163.safetensors 4.30GB
fileresolve/main/model-00151-of-000163.safetensors 4.30GB
fileresolve/main/model-00149-of-000163.safetensors 4.30GB
fileresolve/main/model-00146-of-000163.safetensors 4.30GB
fileresolve/main/model-00147-of-000163.safetensors 4.30GB
fileresolve/main/model-00148-of-000163.safetensors 4.30GB
fileresolve/main/model-00143-of-000163.safetensors 4.30GB
fileresolve/main/model-00144-of-000163.safetensors 4.30GB
fileresolve/main/model-00145-of-000163.safetensors 4.30GB
fileresolve/main/model-00142-of-000163.safetensors 4.30GB
fileresolve/main/model-00140-of-000163.safetensors 4.30GB
fileresolve/main/model-00141-of-000163.safetensors 3.14GB
fileresolve/main/model-00139-of-000163.safetensors 4.30GB
fileresolve/main/model-00138-of-000163.safetensors 4.30GB
fileresolve/main/model-00136-of-000163.safetensors 4.30GB
fileresolve/main/model-00137-of-000163.safetensors 4.30GB
fileresolve/main/model-00134-of-000163.safetensors 4.30GB
fileresolve/main/model-00133-of-000163.safetensors 4.30GB
fileresolve/main/model-00135-of-000163.safetensors 4.30GB
fileresolve/main/model-00131-of-000163.safetensors 4.30GB
fileresolve/main/model-00132-of-000163.safetensors 4.30GB
fileresolve/main/model-00001-of-000163.safetensors 5.23GB
fileresolve/main/model-00002-of-000163.safetensors 4.30GB
fileresolve/main/model-00003-of-000163.safetensors 4.30GB
fileresolve/main/model-00004-of-000163.safetensors 4.30GB
fileresolve/main/model-00005-of-000163.safetensors 4.30GB
fileresolve/main/model-00006-of-000163.safetensors 4.37GB
fileresolve/main/model-00007-of-000163.safetensors 4.31GB
fileresolve/main/model-00008-of-000163.safetensors 4.30GB
fileresolve/main/model-00009-of-000163.safetensors 4.30GB
fileresolve/main/model-00010-of-000163.safetensors 4.30GB
fileresolve/main/model-00011-of-000163.safetensors 4.30GB
fileresolve/main/model-00012-of-000163.safetensors 1.32GB
fileresolve/main/model-00013-of-000163.safetensors 4.30GB
fileresolve/main/model-00014-of-000163.safetensors 4.30GB
fileresolve/main/model-00015-of-000163.safetensors 4.30GB
fileresolve/main/model-00016-of-000163.safetensors 4.30GB
Too many files! Click here to view them all.
Type: Dataset
Tags: weightsLLM

Bibtex:
@article{,
title= {DeepSeek-R1 model weights},
keywords= {LLM, weights},
author= {},
abstract= {Weights for DeepSeek-R1 from Huggingface

We introduce our first-generation reasoning models, DeepSeek-R1-Zero and DeepSeek-R1. DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors. However, DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. To address these issues and further enhance reasoning performance, we introduce DeepSeek-R1, which incorporates cold-start data before RL. DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

https://i.imgur.com/q6NKD6T.png

## License
This code repository and the model weights are licensed under the MIT License. DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:

DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from Qwen-2.5 series, which are originally licensed under Apache 2.0 License, and now finetuned with 800k samples curated with DeepSeek-R1.
DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under llama3.1 license.
DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under llama3.3 license.



```
@misc{deepseekai2025deepseekr1incentivizingreasoningcapability,
      title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning}, 
      author={DeepSeek-AI},
      year={2025},
      eprint={2501.12948},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.12948}, 
}
```
},
terms= {},
license= {https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE},
superseded= {},
url= {https://huggingface.co/deepseek-ai/DeepSeek-R1}
}


Hosted by users:

Send Feedback Start
   0.000010
DB Connect
   0.003785
Lookup hash in DB
   0.002301
Get torrent details
   0.000683
Get torrent details, finished
   0.007961
Get authors
   0.000003
Select authors
   0.000608
Parse bibtex
   0.000270
Write header
   0.000710
get stars
   0.000346
home tab
   0.005478
render right panel
   0.000016
render ads
   0.001221
fetch current hosters
   0.001276
Done