Skip to content

Issue converting DeepSeek-R1 to GGUF format. #11989

Answered by fairydreaming
bjodom asked this question in Q&A
Discussion options

You must be logged in to vote

The original model has fp8 weights and some extra tensors with dequantization scales. llama.cpp doesn't support this, you have to:

  1. Either Download bf16 DeepSeek R1, for example this: https://huggingface.co/unsloth/DeepSeek-R1-BF16. This will convert without any errors.
  2. Or convert the original fp8 model to bf16 by yourself. You can find instructions here: https://huggingface.co/huihui-ai/DeepSeek-R1-bf16. Note that the conversion script (it's bundled with DeepSeek V3 model) uses triton, so you need a GPU for this step.

Replies: 1 comment 5 replies

Comment options

You must be logged in to vote
5 replies
@fairydreaming
Comment options

Answer selected by CISC
@nickhuang99
Comment options

@fairydreaming
Comment options

@nickhuang99
Comment options

@fairydreaming
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
4 participants