Arcee AI AFM-4.5B-Preview
4.5B-parameter foundation model trained on 6.58T curated tokens, achieving 200+ CPU tokens/sec with Western compliance standards, outperforming Qwen3-4B and Gemma3-4B across benchmarks.
About model
Arcee AI AFM-4.5B-Preview is an instruction-tuned model designed for enterprise-grade performance, developed for diverse deployment environments. It excels in mathematical reasoning and code generation. Suitable for enterprise users requiring robust model performance.
API usage
Endpoint:
Model card
AFM-4.5B-Preview is Arcee AI's first foundation model, trained on 6.58T curated tokens via partnership with DatologyAI. Optimized for enterprise deployment with Western compliance standards and efficient CPU/GPU inference.
Key Improvements:
- 200+ tokens/sec CPU performance on 4-bit quantization
- Enterprise-grade compliance for regulated industries
- Multi-stage post-training with RL and KTO alignment
Benchmark Performance
- Model providerArcee AI
- TypeLLM
- Main use casesChatSmall & Fast
- DeploymentServerless
- Endpoint
- Parameters4.6B
- Context length65k
- Input modalitiesText
- Output modalitiesText
- ReleasedJune 19, 2025
- External link
- CategoryChat
