Models / Deep Cogito
Chat
Code
Reasoning

Cogito V1 Preview Llama 70B

Best-in-class open-source LLM trained with IDA for alignment, reasoning, and self-reflective, agentic applications.

About model

Cogito V1 Preview Llama 70B is a hybrid reasoning model for text generation, optimized for coding, STEM, and instruction following, with high multilingual and tool calling capabilities, suitable for commercial use.

To run this model, you first need to deploy it on a Dedicated Endpoint.

  • Model card

    The Cogito LLMs are instruction tuned generative models (text in/text out). All models are released under an open license for commercial use.

    • Cogito models are hybrid reasoning models. Each model can answer directly (standard LLM), or self-reflect before answering (like reasoning models).
    • The LLMs are trained using Iterated Distillation and Amplification (IDA) - an scalable and efficient alignment strategy for superintelligence using iterative self-improvement.
    • The models have been optimized for coding, STEM, instruction following and general helpfulness, and have significantly higher multilingual, coding and tool calling capabilities than size equivalent counterparts.
      • In both standard and reasoning modes, Cogito v1-preview models outperform their size equivalent counterparts on common industry benchmarks.
    • Each model is trained in over 30 languages and supports a context length of 128k.

    Evaluations

    We compare our models against the state of the art size equivalent models in direct mode as well as the reasoning mode. For the direct mode, we compare against Llama / Qwen instruct counterparts. For reasoning, we use Deepseek's R1 distilled counterparts / Qwen's QwQ model.

    Logo

    Livebench Global Average:

    Logo

    For detailed evaluations, please refer to the Blog Post.

Related models
  • Model provider
    Deep Cogito
  • Type
    Chat
    Code
    Reasoning
  • Main use cases
    Chat
    Medium General Purpose
  • Deployment
    On-Demand Dedicated
    Monthly Reserved
  • Parameters
    70B
  • Context length
    128K
  • Input modalities
    Text
  • Output modalities
    Text