DeepSeek-R1 model weights

folder DeepSeek-R1 (163 files)
fileresolve/main/model-00163-of-000163.safetensors 6.58GB
fileresolve/main/model-00161-of-000163.safetensors 4.30GB
fileresolve/main/model-00162-of-000163.safetensors 4.30GB
fileresolve/main/model-00159-of-000163.safetensors 4.30GB
fileresolve/main/model-00160-of-000163.safetensors 5.23GB
fileresolve/main/model-00156-of-000163.safetensors 4.30GB
fileresolve/main/model-00157-of-000163.safetensors 4.30GB
fileresolve/main/model-00158-of-000163.safetensors 4.30GB
fileresolve/main/model-00152-of-000163.safetensors 4.30GB
fileresolve/main/model-00153-of-000163.safetensors 4.30GB
fileresolve/main/model-00154-of-000163.safetensors 4.30GB
fileresolve/main/model-00155-of-000163.safetensors 4.30GB
fileresolve/main/model-00150-of-000163.safetensors 4.30GB
fileresolve/main/model-00151-of-000163.safetensors 4.30GB
fileresolve/main/model-00149-of-000163.safetensors 4.30GB
fileresolve/main/model-00146-of-000163.safetensors 4.30GB
fileresolve/main/model-00147-of-000163.safetensors 4.30GB
fileresolve/main/model-00148-of-000163.safetensors 4.30GB
fileresolve/main/model-00143-of-000163.safetensors 4.30GB
fileresolve/main/model-00144-of-000163.safetensors 4.30GB
fileresolve/main/model-00145-of-000163.safetensors 4.30GB
fileresolve/main/model-00142-of-000163.safetensors 4.30GB
fileresolve/main/model-00140-of-000163.safetensors 4.30GB
fileresolve/main/model-00141-of-000163.safetensors 3.14GB
fileresolve/main/model-00139-of-000163.safetensors 4.30GB
fileresolve/main/model-00138-of-000163.safetensors 4.30GB
fileresolve/main/model-00136-of-000163.safetensors 4.30GB
fileresolve/main/model-00137-of-000163.safetensors 4.30GB
fileresolve/main/model-00134-of-000163.safetensors 4.30GB
fileresolve/main/model-00133-of-000163.safetensors 4.30GB
fileresolve/main/model-00135-of-000163.safetensors 4.30GB
fileresolve/main/model-00131-of-000163.safetensors 4.30GB
fileresolve/main/model-00132-of-000163.safetensors 4.30GB
fileresolve/main/model-00001-of-000163.safetensors 5.23GB
fileresolve/main/model-00002-of-000163.safetensors 4.30GB
fileresolve/main/model-00003-of-000163.safetensors 4.30GB
fileresolve/main/model-00004-of-000163.safetensors 4.30GB
fileresolve/main/model-00005-of-000163.safetensors 4.30GB
fileresolve/main/model-00006-of-000163.safetensors 4.37GB
fileresolve/main/model-00007-of-000163.safetensors 4.31GB
fileresolve/main/model-00008-of-000163.safetensors 4.30GB
fileresolve/main/model-00009-of-000163.safetensors 4.30GB
fileresolve/main/model-00010-of-000163.safetensors 4.30GB
fileresolve/main/model-00011-of-000163.safetensors 4.30GB
fileresolve/main/model-00012-of-000163.safetensors 1.32GB
fileresolve/main/model-00013-of-000163.safetensors 4.30GB
fileresolve/main/model-00014-of-000163.safetensors 4.30GB
fileresolve/main/model-00015-of-000163.safetensors 4.30GB
fileresolve/main/model-00016-of-000163.safetensors 4.30GB
Too many files! Click here to view them all.
Type: Dataset
Tags: weights, LLM

Bibtex:
@article{,
title= {DeepSeek-R1 model weights},
keywords= {LLM, weights},
author= {},
abstract= {Weights for DeepSeek-R1 from Huggingface

We introduce our first-generation reasoning models, DeepSeek-R1-Zero and DeepSeek-R1. DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors. However, DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. To address these issues and further enhance reasoning performance, we introduce DeepSeek-R1, which incorporates cold-start data before RL. DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

https://i.imgur.com/q6NKD6T.png

## License
This code repository and the model weights are licensed under the MIT License. DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:

DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from Qwen-2.5 series, which are originally licensed under Apache 2.0 License, and now finetuned with 800k samples curated with DeepSeek-R1.
DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under llama3.1 license.
DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under llama3.3 license.



```
@misc{deepseekai2025deepseekr1incentivizingreasoningcapability,
      title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning}, 
      author={DeepSeek-AI},
      year={2025},
      eprint={2501.12948},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.12948}, 
}
```
},
terms= {},
license= {https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE},
superseded= {},
url= {https://huggingface.co/deepseek-ai/DeepSeek-R1}
}

Hosted by users:

Send Feedback Start
   0.000004
DB Connect
   0.000866
Lookup hash in DB
   0.002084
Get torrent details
   0.000842
Get torrent details, finished
   0.001365
Get authors
   0.000003
Select authors
   0.000848
Parse bibtex
   0.000698
Write header
   0.000961
get stars
   0.000629
home tab
   0.014239
render right panel
   0.000012
render ads
   0.000042
fetch current hosters
   0.003977
Done