Model

deepseek-r1-distill-llama-70b

Public

Use cases

Minimum system memory

40GB

Tags

70B
llama

README

DeepSeek R1 Distill Llama 70B by deepseek-ai

Supports context length of 128k.

Distilled from DeepSeek's R1 reasoning model.

Tuned for reasoning and chain-of-thought.

Sources

The underlying model files this model uses