- Phind: CodeLlama 34B v2
A fine-tune of CodeLlama-34B on an internal dataset that helps it exceed GPT-4 on some benchmarks, including HumanEval.
by phind4.1K context$0.72/M input tkns$0.72/M output tkns2.94M tokens this week - Meta: CodeLlama 34B Instruct
Code Llama is built upon Llama 2 and excels at filling in code, handling extensive input contexts, and folling programming instructions without prior training for various programming tasks.
by meta-llama8.2K context$0.72/M input tkns$0.72/M output tkns538K tokens this week - Meta: Llama v2 13B Chat
A 13 billion parameter language model from Meta, fine tuned for chat completions
by meta-llama4.1K context$0.13/M input tkns$0.13/M output tkns19.1M tokens this week - Lynn: Llama 3 Soliloquy 8B
Soliloquy-L3 is a fast, highly capable roleplaying model designed for immersive, dynamic experiences. Trained on over 250 million tokens of roleplaying data, Soliloquy-L3 has a vast knowledge base, rich literary expression, and support for up to 24k context length. It outperforms existing ~13B models, delivering enhanced roleplaying capabilities. Usage of this model is subject to Meta's Acceptable Use Policy.
by lynn25K context$0.2/M input tkns$0.2/M output tkns123M tokens this week - Meta: Llama 3 8B Instruct (extended)
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy. Note: this is an extended-context version of this model. It may have higher prices and different outputs.
by meta-llama16K context$0.2751/M input tkns$2.826/M output tkns13.1M tokens this week - Meta: Llama 3 8B Instruct (nitro)
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy. Note: this is a higher-throughput version of this model, and may have higher prices and slightly different outputs.
by meta-llama8.2K context$0.2/M input tkns$0.2/M output tkns49.8M tokens this week - Meta: Llama 3 70B Instruct (nitro)
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy. Note: this is a higher-throughput version of this model, and may have higher prices and slightly different outputs.
by meta-llama8.2K context$0.9/M input tkns$0.9/M output tkns423M tokens this week - Meta: Llama 3 70B Instruct
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy.
by meta-llama8.2K context$0.81/M input tkns$0.81/M output tkns1.83B tokens this week - Meta: Llama 3 8B Instruct
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy.
by meta-llama8.2K context$0.1/M input tkns$0.1/M output tkns2.29B tokens this week - Databricks: DBRX 132B Instruct
DBRX is a new open source large language model developed by Databricks. At 132B, it outperforms existing open source LLMs like Llama 2 70B and Mixtral-8x7B on standard industry benchmarks for language understanding, programming, math, and logic. It uses a fine-grained mixture-of-experts (MoE) architecture. 36B parameters are active on any input. It was pre-trained on 12T tokens of text and code data. Compared to other open MoE models like Mixtral-8x7B and Grok-1, DBRX is fine-grained, meaning it uses a larger number of smaller experts. See the launch announcement and benchmark results here. #moe
by databricks33K context$0.6/M input tkns$0.6/M output tkns12.2M tokens this week - Chronos Hermes 13B v2
A 75/25 merge of Chronos 13b v2 and Nous Hermes Llama2 13b. This offers the imaginative writing style of Chronos while retaining coherency. Outputs are long and use exceptional prose. #merge
by austism4.1K context$0.13/M input tkns$0.13/M output tkns121M tokens this week - Psyfighter v2 13B
The v2 of Psyfighter - a merged model created by the KoboldAI community members Jeb Carter and TwistedShadows, made possible thanks to the KoboldAI merge request service. The intent was to add medical data to supplement the model's fictional ability with more details on anatomy and mental states. This model should not be used for medical advice or therapy because of its high likelihood of pulling in fictional data. It's a merge between: - KoboldAI/LLaMA2-13B-Tiefighter - Doctor-Shotgun/cat-v1.0-13b - Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged. #merge
by koboldai4.1K context$1/M input tkns$1/M output tkns22.4M tokens this week - Nous: Hermes 13B
A state-of-the-art language model fine-tuned on over 300k instructions by Nous Research, with Teknium and Emozilla leading the fine tuning process.
by nousresearch4.1K context$0.26/M input tkns$0.26/M output tkns692M tokens this week - Psyfighter 13B
A #merge model based on Llama-2-13B and made possible thanks to the compute provided by the KoboldAI community. It's a merge between: - KoboldAI/LLaMA2-13B-Tiefighter - chaoyi-wu/MedLLaMA_13B - Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged. #merge
by jebcarter4.1K context - MythoMax 13B (nitro)
One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge Note: this is a higher-throughput version of this model, and may have higher prices and slightly different outputs.
by gryphe4.1K context$0.2/M input tkns$0.2/M output tkns1.84B tokens this week - MythoMax 13B (extended)
One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge Note: this is an extended-context version of this model. It may have higher prices and different outputs.
by gryphe8.2K context$1.125/M input tkns$1.125/M output tkns46.8M tokens this week - MythoMax 13B
One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge
by gryphe4.1K context$0.18/M input tkns$0.18/M output tkns971M tokens this week - Mistral 7B Instruct (nitro)
A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length. This is v0.2 of Mistral 7B Instruct. For v0.1, use this model. Note: this is a higher-throughput version of this model, and may have higher prices and slightly different outputs.
by mistralai33K context$0.2/M input tkns$0.2/M output tkns173M tokens this week - Noromaid Mixtral 8x7B Instruct
This model was trained for 8h(v1) + 8h(v2) + 12h(v3) on customized modified datasets, focusing on RP, uncensoring, and a modified version of the Alpaca prompting (that was already used in LimaRP), which should be at the same conversational level as ChatLM or Llama2-Chat without adding any additional special tokens.
by neversleep8K context$8/M input tkns$8/M output tkns121M tokens this week - lzlv 70B
A Mythomax/MLewd_13B-style merge of selected 70B models. A multi-model merge of several LLaMA2 70B finetunes for roleplaying and creative work. The goal was to create a model that combines creativity with intelligence for an enhanced experience. #merge #uncensored
by lizpreciatior4.1K context$0.7/M input tkns$0.8/M output tkns412M tokens this week - Goliath 120B
A large LLM created by combining two fine-tuned Llama 70B models into one 120B model. Combines Xwin and Euryale. Credits to - @chargoddard for developing the framework used to merge the model - mergekit. - @Undi95 for helping with the merge ratios. #merge
by alpindale6.1K context$9.375/M input tkns$9.375/M output tkns65.1M tokens this week - Xwin 70B
Xwin-LM aims to develop and open-source alignment tech for LLMs. Our first release, built-upon on the Llama2 base models, ranked TOP-1 on AlpacaEval. Notably, it's the first to surpass GPT-4 on this benchmark. The project will be continuously updated.
by xwin-lm8.2K context$3.75/M input tkns$3.75/M output tkns10.9M tokens this week - Mistral 7B Instruct (free)
A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length. This is v0.1 of Mistral 7B Instruct. For v0.2, use this model. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by mistralai33K context$0/M input tkns$0/M output tkns60M tokens this week - Mistral 7B Instruct
A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length. This is v0.1 of Mistral 7B Instruct. For v0.2, use this model.
by mistralai33K context$0.1/M input tkns$0.25/M output tkns2.96B tokens this week