Models / QwenQwen / / Qwen3-Next-80B-A3B-Instruct API
Qwen3-Next-80B-A3B-Instruct API

This model is not currently supported on Together AI.
Visit our Models page to view all the latest models.
Instruction-Optimized Efficiency:
Qwen3-Next Instruct features a highly sparse MoE structure that activates only 3B of its 80B parameters during inference. Supports only instruct mode without thinking blocks, delivering performance on par with Qwen3-235B-A22B-Instruct-2507 on certain benchmarks while using less than 10% training cost and providing 10x+ higher throughput on contexts over 32K tokens.
Qwen3-Next-80B-A3B-Instruct API Usage
Endpoint
How to use Qwen3-Next-80B-A3B-Instruct
Model details
Architecture Overview:
• 48 layers with 2048 hidden dimension and hybrid layout pattern
• 512 total experts with 10 activated and 1 shared expert per MoE layer
• Multi-token prediction mechanism for faster inference
Instruction Optimization:
• Supports only instruct mode without blocks
• Specialized post-training for task completion on 15T tokens
• Performance on par with Qwen3-235B while using significantly fewer resources
Performance Characteristics:
• 262K native context length, extensible to 1M tokens with YaRN scaling
• More than 10x higher throughput on contexts over 32K tokens
• SGLang and vLLM deployment support with Multi-Token Prediction
Prompting Qwen3-Next-80B-A3B-Instruct
Applications & Use Cases
Task Automation:
• Code generation and software development assistance with cost-effective processing
• Content creation and editing with specific instructions
• Data analysis and report generation following detailed guidelines
Business Applications:
• Customer service automation with instruction-based responses
• Technical documentation generation with specific formatting requirements
• Process automation and workflow optimization
Agentic Use Cases:
• Tool calling capabilities with MCP configuration support
• Multi-step task execution with built-in and custom tools
• Extended conversation and context-aware task completion up to 262K tokens