Mistral: Mixtral 8x7B Instruct

mistralai/mixtral-8x7b-instruct

Updated Dec 1032,768 context
$0.3 / 1M input tokens$0.3 / 1M output tokens

A pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts (feed-forward networks) for a total of 47B parameters.

Instruct model fine-tuned by Mistral. If you send raw prompts, you should use [INST] and [/INST] tokens. Otherwise, with chat messages, the prompt will be formatted automatically for you.

#moe

OpenRouter first attempts the primary provider, and falls back to others if it encounters an error. Prices displayed per million tokens.