Mistral: Ministral 8B
Detailed specifications for implementing Mistral: Ministral 8B in your RAG applications.
Model Overview
Released:October 17, 2024
The Ministral 8B is a cutting-edge 8-billion-parameter model engineered for exceptional performance and efficiency. Its innovative interleaved sliding-window attention mechanism enables faster inference and reduced memory usage, making it ideal for edge computing applications. With support for context lengths of up to 128k tokens, the model excels in knowledge-intensive and reasoning tasks, setting a new benchmark in the sub-10B parameter category. Ministral 8B outperforms comparable models, delivering superior results for low-latency, privacy-focused use cases where speed and efficiency are critical.
Architecture
- Modality
- text->text
- Tokenizer
- Mistral
Pricing
Operation | Rate |
---|---|
Prompt | 0.0000001 |
Completion | 0.0000001 |
Image | 0 |
Request | 0 |
Provider Details
- Context Length
- 128,000 tokens
- Max Completion
- 0 tokens
- Moderation
- Not Enabled
Ready to implement Mistral: Ministral 8B?
Start building powerful RAG applications with our flexible pricing plans and developer-friendly API.