Mistral: Mixtral 8x7B Instruct


Updated Dec 1032,768 context
$0.3 / 1M input tokens$0.3 / 1M output tokens

A pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts (feed-forward networks) for a total of 47B parameters.

Instruct model fine-tuned by Mistral. If you send raw prompts, you should use [INST] and [/INST] tokens. Otherwise, with chat messages, the prompt will be formatted automatically for you.


OpenRouter first attempts the primary provider, and falls back to others if it encounters an error. Prices displayed per million tokens.