• Extracted a 64 Rank Lora from DeepSeek-R1-Distill-Qwen-32B
  • Merged & Quantized into Q4_K_M

Note: The model seems to be somewhat working with the R1's weird template too but it repeats random Chinese characters and the quality seems to be consistently worse.

Maybe try using the R1 tokenizer.

Downloads last month
26
GGUF
Model size
32.8B params
Architecture
qwen2

4-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for Ba2han/qwen-coder-thinker-q4_k_m

Quantized
(107)
this model