Model Comparison
Authormistralai
Context Length128K
Ministral 8B is an 8B parameter model featuring a unique interleaved sliding-window attention pattern for faster, memory-efficient inference. Designed for edge use cases, it supports up to 128k context length and excels in knowledge and reasoning tasks. It outperforms peers in the sub-10B category, making it perfect for low-latency, privacy-first applications.
Pricing
Input$0.10 / M tokens
Output$0.10 / M tokens
Images– –
Endpoint Features
Quantizationunknown
Max Tokens (input + output)128K
Max Output Tokens– –
Stream cancellation– –
Supports Tools– –
No Prompt Training
Reasoning– –