DeepSeek: DeepSeek R1 Zero (free)

deepseek/deepseek-r1-zero:free

Created Mar 6, 2025163,840 context
$0/M input tokens$0/M output tokens

DeepSeek-R1-Zero is a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step. It's 671B parameters in size, with 37B active in an inference pass.

It demonstrates remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors.

DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. See DeepSeek R1 for the SFT model.

Recent activity on DeepSeek R1 Zero (free)

Tokens processed per day

Mar 6Mar 10Mar 14Mar 18Mar 22Mar 26Mar 30Apr 3Apr 7Apr 11Apr 15Apr 19Apr 23Apr 27May 10300M600M900M1.2B
    DeepSeek: DeepSeek R1 Zero (free) – Recent Activity | OpenRouter