Deepcogito Models

Explore the Deepcogito language and embedding models available through our OpenAI Assistants API-compatible service.

Deepcogito logo

Deep Cogito: Cogito V2 Preview Llama 405B

Context Length:
32,768 tokens
Architecture:
text->text

Pricing:

Prompt: $0.0000035
Completion: $0.0000035

Cogito v2 405B is a dense hybrid reasoning model that combines direct answering capabilities with advanced self-reflection. It represents a significant step toward frontier intelligence with dense architecture delivering performance competitive with leading closed models. This advanced reasoning system combines policy improvement with massive scale for exceptional capabilities.

Deep Cogito: Cogito V2 Preview Llama 70B

Context Length:
32,768 tokens
Architecture:
text->text

Pricing:

Prompt: $0.00000088
Completion: $0.00000088

Cogito v2 70B is a dense hybrid reasoning model that combines direct answering capabilities with advanced self-reflection. Built with iterative policy improvement, it delivers strong performance across reasoning tasks while maintaining efficiency through shorter reasoning chains and improved intuition.

Cogito V2 Preview Llama 109B

Context Length:
32,767 tokens
Architecture:
text+image->text

Pricing:

Prompt: $0.00000018
Completion: $0.00000059

An instruction-tuned, hybrid-reasoning Mixture-of-Experts model built on Llama-4-Scout-17B-16E. Cogito v2 can answer directly or engage an extended “thinking” phase, with alignment guided by Iterated Distillation & Amplification (IDA). It targets coding, STEM, instruction following, and general helpfulness, with stronger multilingual, tool-calling, and reasoning performance than size-equivalent baselines. The model supports long-context use (up to 10M tokens) and standard Transformers workflows. Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs

Deep Cogito: Cogito V2 Preview Deepseek 671B

Context Length:
163,840 tokens
Architecture:
text->text

Pricing:

Prompt: $0.00000125
Completion: $0.00000125

Cogito v2 is a multilingual, instruction-tuned Mixture of Experts (MoE) large language model with 671 billion parameters. It supports both standard and reasoning-based generation modes. The model introduces hybrid reasoning via Iterated Distillation and Amplification (IDA)—an iterative self-improvement strategy designed to scale alignment with general intelligence. Cogito v2 has been optimized for STEM, programming, instruction following, and tool use. It supports 128k context length and offers strong performance in both multilingual and code-heavy environments. Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs

Ready to build with Deepcogito?

Start using these powerful models in your applications with our flexible pricing plans.