Cognitivecomputations Models
Explore the Cognitivecomputations language and embedding models available through our OpenAI Assistants API-compatible service.
Venice: Uncensored (free)
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
Pricing:
Venice Uncensored Dolphin Mistral 24B Venice Edition is a fine-tuned variant of Mistral-Small-24B-Instruct-2501, developed by dphn.ai in collaboration with Venice.ai. This model is designed as an “uncensored” instruct-tuned LLM, preserving user control over alignment, system prompts, and behavior. Intended for advanced and unrestricted use cases, Venice Uncensored emphasizes steerability and transparent behavior, removing default safety and alignment layers typically found in mainstream assistant models.
Dolphin3.0 Mistral 24B (free)
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
Pricing:
Dolphin 3.0 is the next generation of the Dolphin series of instruct-tuned models. Designed to be the ultimate general purpose local model, enabling coding, math, agentic, function calling, and general use cases.
Dolphin aims to be a general purpose instruct model, similar to the models behind ChatGPT, Claude, Gemini.
Part of the Dolphin 3.0 Collection Curated and trained by Eric Hartford, Ben Gitter, BlouseJury and Cognitive Computations
Dolphin3.0 Mistral 24B
- Context Length:
- 32,768 tokens
- Architecture:
- text->text
- Max Output:
- 32,768 tokens
Pricing:
Dolphin 3.0 is the next generation of the Dolphin series of instruct-tuned models. Designed to be the ultimate general purpose local model, enabling coding, math, agentic, function calling, and general use cases.
Dolphin aims to be a general purpose instruct model, similar to the models behind ChatGPT, Claude, Gemini.
Part of the Dolphin 3.0 Collection Curated and trained by Eric Hartford, Ben Gitter, BlouseJury and Cognitive Computations
Ready to build with Cognitivecomputations?
Start using these powerful models in your applications with our flexible pricing plans.