Models / Arcee AI
LLM
Chat

Arcee AI AFM-4.5B-Preview

4.5B-parameter foundation model trained on 6.58T curated tokens, achieving 200+ CPU tokens/sec with Western compliance standards, outperforming Qwen3-4B and Gemma3-4B across benchmarks.

This model is not available on Together’s Serverless API.

Deploy this model on an on-demand Dedicated Endpoint or pick a supported alternative from the Model Library.

Related models
  • Model provider
    Arcee AI
  • Type
    LLM
    Chat
  • Main use cases
    Chat
    Small & Fast
  • Parameters
    4.6B
  • Context length
    65k