Models / Qwen
Chat
Reasoning

Qwen3 235B A22B FP8 Throughput

Hybrid instruct + reasoning model (232Bx22B MoE) optimized for high-throughput, cost-efficient inference and distillation.

Performance benchmarks

Model

AIME 2025

GPQA Diamond

HLE

LiveCodeBench

MATH500

SWE-bench verified

70.7%

65.9%

Related open-source models

Competitor closed-source models

Claude Opus 4.6

90.5%

34.2%

78.7%

OpenAI o3

83.3%

24.9%

99.2%

62.3%

OpenAI o1

76.8%

96.4%

48.9%

GPT-4o

49.2%

2.7%

32.3%

89.3%

31.0%

This model is not available on Together’s Serverless API.

Deploy this model on an on-demand Dedicated Endpoint or pick a supported alternative from the Model Library.

Related models
  • Model provider
    Qwen
  • Type
    Chat
    Reasoning
  • Main use cases
    Chat
    Reasoning
    Medium General Purpose
    Function Calling
  • Features
    Function Calling
  • Parameters
    235.1B
  • Context length
    40k