- Meta: Llama 3 8B Instruct (free)
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by meta-llama8K context$0/M input tkns$0/M output tkns1.84M tokens this week - Hugging Face: Zephyr 7B (free)
Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-β is the second model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by huggingfaceh44K context$0/M input tkns$0/M output tkns73K tokens this week - Cinematika 7B (alpha) (free)
This model is under development. Check the OpenRouter Discord for updates. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by openrouter8K context$0/M input tkns$0/M output tkns134K tokens this week - Mistral 7B Instruct (free)
A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length. This is v0.1 of Mistral 7B Instruct. For v0.2, use this model. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by mistralai33K context$0/M input tkns$0/M output tkns5.4M tokens this week - Nous: Capybara 7B (free)
The Capybara series is a collection of datasets and models made by fine-tuning on data created by Nous, mostly in-house. V1.9 uses unalignment techniques for more consistent and dynamic control. It also leverages a significantly better foundation model, Mistral 7B. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by nousresearch4K context$0/M input tkns$0/M output tkns750K tokens this week - Google: Gemma 7B (free)
Gemma by Google is an advanced, open-source language model family, leveraging the latest in decoder-only, text-to-text technology. It offers English language capabilities across text generation tasks like question answering, summarization, and reasoning. The Gemma 7B variant is comparable in performance to leading open source models. Usage of Gemma is subject to Google's Gemma Terms of Use. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by google8K context$0/M input tkns$0/M output tkns1.58M tokens this week - OpenChat 3.5 (free)
OpenChat is a library of open-source language models, fine-tuned with "C-RLFT (Conditioned Reinforcement Learning Fine-Tuning)" - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by openchat8K context$0/M input tkns$0/M output tkns3.24M tokens this week - MythoMist 7B (free)
From the creator of MythoMax, merges a suite of models to reduce word anticipation, ministrations, and other undesirable words in ChatGPT roleplaying data. It combines Neural Chat 7B, Airoboros 7b, Toppy M 7B, Zepher 7b beta, Nous Capybara 34B, OpenHeremes 2.5, and many others. #merge Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by gryphe33K context$0/M input tkns$0/M output tkns700K tokens this week - Toppy M 7B (free)
A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 - HuggingFaceH4/zephyr-7b-beta - lemonilia/AshhLimaRP-Mistral-7B - Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b - Undi95/Mistral-pippa-sharegpt-7b-qlora #merge #uncensored Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by undi954K context$0/M input tkns$0/M output tkns2.73M tokens this week - RWKV v5: Eagle 7B
Eagle 7B is trained on 1.1 Trillion Tokens across 100+ world languages (70% English, 15% multilang, 15% code). - Built on the RWKV-v5 architecture (a linear transformer with 10-100x+ lower inference cost) - Ranks as the world's greenest 7B model (per token) - Outperforms all 7B class models in multi-lingual benchmarks - Approaches Falcon (1.5T), LLaMA2 (2T), Mistral (>2T?) level of performance in English evals - Trade blows with MPT-7B (1T) in English evals - All while being an "Attention-Free Transformer" Eagle 7B models are provided for free, by Recursal.AI, for the beta period till end of March 2024 Find out more here rnn
by recursal10K context$0/M input tkns$0/M output tkns617 tokens this week - RWKV v5 World 3B
RWKV is an RNN (recurrent neural network) with transformer-level performance. It aims to combine the best of RNNs and transformers - great performance, fast inference, low VRAM, fast training, "infinite" context length, and free sentence embedding. RWKV-5 is trained on 100+ world languages (70% English, 15% multilang, 15% code). RWKV 3B models are provided for free, by Recursal.AI, for the beta period. More details here. #rnn
by rwkv10K context$0/M input tkns$0/M output tkns14K tokens this week - RWKV v5 3B AI Town
This is an RWKV 3B model finetuned specifically for the AI Town project. RWKV is an RNN (recurrent neural network) with transformer-level performance. It aims to combine the best of RNNs and transformers - great performance, fast inference, low VRAM, fast training, "infinite" context length, and free sentence embedding. RWKV 3B models are provided for free, by Recursal.AI, for the beta period. More details here. #rnn
by recursal10K context$0/M input tkns$0/M output tkns186 tokens this week