- Anthropic: Claude 3 Haiku (self-moderated)
This is a lower-latency version of Claude 3 Haiku, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Claude 3 Haiku is Anthropic's fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance. See the launch announcement and benchmark results here #multimodal
by anthropic200K context$0.25/M input tkns$1.25/M output tkns$0.4/K input imgs38.9M tokens this week - Anthropic: Claude 3 Opus (self-moderated)
This is a lower-latency version of Claude 3 Opus, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Claude 3 Opus is Anthropic's most powerful model for highly complex tasks. It boasts top-level performance, intelligence, fluency, and understanding. See the launch announcement and benchmark results here #multimodal
by anthropic200K context$15/M input tkns$75/M output tkns$24/K input imgs25.1M tokens this week - Anthropic: Claude 3 Sonnet (self-moderated)
This is a lower-latency version of Claude 3 Sonnet, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Claude 3 Sonnet is an ideal balance of intelligence and speed for enterprise workloads. Maximum utility at a lower price, dependable, balanced for scaled deployments. See the launch announcement and benchmark results here #multimodal
by anthropic200K context$3/M input tkns$15/M output tkns$4.8/K input imgs17.6M tokens this week - Anthropic: Claude v2 (self-moderated)
This is a lower-latency version of Claude v2, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.
by anthropic200K context$8/M input tkns$24/M output tkns394K tokens this week - Anthropic: Claude v2.0 (self-moderated)
This is a lower-latency version of Claude v2.0, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Anthropic's flagship model. Superior performance on tasks that require complex reasoning. Supports hundreds of pages of text.
by anthropic100K context$8/M input tkns$24/M output tkns940K tokens this week - Anthropic: Claude v2.1 (self-moderated)
This is a lower-latency version of Claude v2.1, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.
by anthropic200K context$8/M input tkns$24/M output tkns2.02M tokens this week - Anthropic: Claude Instant v1 (self-moderated)
This is a lower-latency version of Claude Instant v1, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the model's side instead of OpenRouter's. It's in beta, and may change in the future. Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text.
by anthropic100K context$0.8/M input tkns$2.4/M output tkns299K tokens this week - RWKV v5 World 3B
RWKV is an RNN (recurrent neural network) with transformer-level performance. It aims to combine the best of RNNs and transformers - great performance, fast inference, low VRAM, fast training, "infinite" context length, and free sentence embedding. RWKV-5 is trained on 100+ world languages (70% English, 15% multilang, 15% code). RWKV 3B models are provided for free, by Recursal.AI, for the beta period. More details here. #rnn
by rwkv10K context$0/M input tkns$0/M output tkns24K tokens this week - RWKV v5 3B AI Town
This is an RWKV 3B model finetuned specifically for the AI Town project. RWKV is an RNN (recurrent neural network) with transformer-level performance. It aims to combine the best of RNNs and transformers - great performance, fast inference, low VRAM, fast training, "infinite" context length, and free sentence embedding. RWKV 3B models are provided for free, by Recursal.AI, for the beta period. More details here. #rnn
by recursal10K context$0/M input tkns$0/M output tkns5K tokens this week - Cinematika 7B (alpha) (free)
This model is under development. Check the OpenRouter Discord for updates. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by openrouter8K context$0/M input tkns$0/M output tkns173K tokens this week - Cinematika 7B (alpha)
This model is under development. Check the OpenRouter Discord for updates.
by openrouter8K context$0.18/M input tkns$0.18/M output tkns43K tokens this week - Auto (best for prompt)
Depending on their size, subject, and complexity, your prompts will be sent to Mistral Large, Claude 3 Sonnet or GPT-4o. To see which model was used, visit Activity.
by openrouter200K context