Models / Arcee AI
LLM

Arcee AI AFM-4.5B

4.5B-parameter foundation model trained on 6.58T curated tokens, achieving 200+ CPU tokens/sec with Western compliance standards, outperforming Qwen3-4B and Gemma3-4B across benchmarks.

About model

Arcee AI AFM-4.5B is a 4.5 billion parameter instruction-tuned model for enterprise-grade performance, excelling in mathematical reasoning and code generation. It is designed for diverse deployment environments and suitable for users requiring robust model performance.

  • API usage

    • cURL
    • Python
    • Typescript

    Endpoint:

    arcee-ai/AFM-4.5B

    curl -X POST "https://api.together.xyz/v1/chat/completions" \
      -H "Authorization: Bearer $TOGETHER_API_KEY" \
      -H "Content-Type: application/json" \
      -d '{
        "model": "arcee-ai/AFM-4.5B",
        "messages": [
          {
            "role": "user",
            "content": "What are some fun things to do in New York?"
          }
        ]
    }'
    
    from together import Together
    
    client = Together()
    
    response = client.chat.completions.create(
      model="arcee-ai/AFM-4.5B",
      messages=[
        {
          "role": "user",
          "content": "What are some fun things to do in New York?"
        }
      ]
    )
    print(response.choices[0].message.content)
    
    import Together from 'together-ai';
    const together = new Together();
    
    const completion = await together.chat.completions.create({
      model: 'arcee-ai/AFM-4.5B',
      messages: [
        {
          role: 'user',
          content: 'What are some fun things to do in New York?'
         }
      ],
    });
    
    console.log(completion.choices[0].message.content);
    
  • Model card

    AFM-4.5B is Arcee AI's first foundation model, trained on 6.58T curated tokens via partnership with DatologyAI. Optimized for enterprise deployment with Western compliance standards and efficient CPU/GPU inference.

    Key Improvements:

    • 200+ tokens/sec CPU performance on 4-bit quantization
    • Enterprise-grade compliance for regulated industries
    • Multi-stage post-training with RL and KTO alignment
Related models
  • Model provider
    Arcee AI
  • Type
    LLM
  • Main use cases
    Chat
  • Deployment
    Serverless
    On-Demand Dedicated
  • Parameters
    4.6B
  • Context length
    65k
  • Input price

    $0.10 / 1M tokens

  • Output price

    $0.40 / 1M tokens

  • Input modalities
    Text
  • Output modalities
    Text