DeepSeek: R1 Distill Qwen 32B

deepseek/deepseek-r1-distill-qwen-32b

Created Jan 29, 2025131,072 context
$0.12/M input tokens$0.18/M output tokens

DeepSeek R1 Distill Qwen 32B is a distilled large language model based on Qwen 2.5 32B, using outputs from DeepSeek R1. It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

Other benchmark results include:

  • AIME 2024 pass@1: 72.6
  • MATH-500 pass@1: 94.3
  • CodeForces Rating: 1691

The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

Recent activity on R1 Distill Qwen 32B

Tokens processed per day

Jan 31Feb 6Feb 12Feb 18Feb 24Mar 2Mar 8Mar 14Mar 20Mar 26Apr 1Apr 7Apr 13Apr 19Apr 25May 10200M400M600M800M
    DeepSeek: R1 Distill Qwen 32B – Recent Activity | OpenRouter