Models / DeepSeek
Chat
Reasoning

DeepSeek R1 Distilled Llama 70B

Llama 70B distilled with reasoning capabilities from Deepseek R1. Surpasses GPT-4o with 94.5% on MATH-500 & matches o1-mini on coding.

About model

DeepSeek R1 Distilled Llama 70B performs complex reasoning tasks, excelling in math, code, and reasoning benchmarks. It is designed for researchers and developers seeking advanced language models.

Performance benchmarks

Model

AIME 2025

GPQA Diamond

HLE

LiveCodeBench

MATH500

SWE-bench verified

65.0%

78.3%

Related open-source models

Competitor closed-source models

Claude Opus 4.6

90.5%

34.2%

78.7%

OpenAI o3

83.3%

24.9%

99.2%

62.3%

OpenAI o1

76.8%

96.4%

48.9%

GPT-4o

49.2%

2.7%

32.3%

89.3%

31.0%

    Related models
    • Model provider
      DeepSeek
    • Type
      Chat
      Reasoning
    • Main use cases
      Chat
      Reasoning
    • Features
      JSON Mode
    • Fine tuning
      Supported
    • Deployment
      On-Demand Dedicated
      Monthly Reserved
    • Parameters
      70B
    • Context length
      128K
    • Input price

      $2.00 / 1M tokens

    • Output price

      $2.00 / 1M tokens

    • Input modalities
      Text
    • Output modalities
      Text
    • Released
      January 20, 2025
    • Last updated
      December 22, 2025
    • Quantization level
      FP16
    • External link
    • Category
      Chat