- LLaVA v1.6 34B
LLaVA Yi 34B is an open-source model trained by fine-tuning LLM on multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture. Base LLM: NousResearch/Nous-Hermes-2-Yi-34B It was trained in December 2023.
by liuhaotian4K context$0.9/M input tkns$0.9/M output tkns$0.5184/K input imgs5K tokens this week - Toppy M 7B (nitro)
A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 - HuggingFaceH4/zephyr-7b-beta - lemonilia/AshhLimaRP-Mistral-7B - Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b - Undi95/Mistral-pippa-sharegpt-7b-qlora #merge #uncensored Note: this is a higher-throughput version of this model, and may have higher prices and slightly different outputs.
by undi954K context$0.07/M input tkns$0.07/M output tkns18K tokens this week - Nous: Hermes 2 Mistral 7B DPO
This is the flagship 7B Hermes model, a Direct Preference Optimization (DPO) of Teknium/OpenHermes-2.5-Mistral-7B. It shows improvement across the board on all benchmarks tested - AGIEval, BigBench Reasoning, GPT4All, and TruthfulQA. The model prior to DPO was trained on 1,000,000 instructions/chats of GPT-4 quality or better, primarily synthetic data as well as other high quality datasets.
by nousresearch8K context$0.18/M input tkns$0.18/M output tkns12.1M tokens this week - Nous: Hermes 2 Mixtral 8x7B SFT
Nous Hermes 2 Mixtral 8x7B SFT is the supervised finetune only version of the Nous Research model trained over the Mixtral 8x7B MoE LLM. The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks. #moe
by nousresearch33K context$0.54/M input tkns$0.54/M output tkns115K tokens this week - Nous: Hermes 2 Mixtral 8x7B DPO
Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained over the Mixtral 8x7B MoE LLM. The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks. #moe
by nousresearch33K context$0.27/M input tkns$0.27/M output tkns1.52M tokens this week - Chronos Hermes 13B v2
A 75/25 merge of Chronos 13b v2 and Nous Hermes Llama2 13b. This offers the imaginative writing style of Chronos while retaining coherency. Outputs are long and use exceptional prose. #merge
by austism4K context$0.13/M input tkns$0.13/M output tkns4.12M tokens this week - Nous: Hermes 2 Yi 34B
Nous Hermes 2 Yi 34B was trained on 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape. Nous-Hermes 2 on Yi 34B outperforms all Nous-Hermes & Open-Hermes models of the past, achieving new heights in all benchmarks for a Nous Research LLM as well as surpassing many popular finetunes.
by nousresearch4K context$0.72/M input tkns$0.72/M output tkns90K tokens this week - Nous: Hermes 2 Vision 7B (alpha)
This vision-language model builds on innovations from the popular OpenHermes-2.5 model, by Teknium. It adds vision support, and is trained on a custom dataset enriched with function calling This project is led by qnguyen3 and teknium. #multimodal
by nousresearch4K context$10/M input tkns$10/M output tkns - Nous: Capybara 7B (free)
The Capybara series is a collection of datasets and models made by fine-tuning on data created by Nous, mostly in-house. V1.9 uses unalignment techniques for more consistent and dynamic control. It also leverages a significantly better foundation model, Mistral 7B. Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by nousresearch4K context$0/M input tkns$0/M output tkns554K tokens this week - Nous: Capybara 7B
The Capybara series is a collection of datasets and models made by fine-tuning on data created by Nous, mostly in-house. V1.9 uses unalignment techniques for more consistent and dynamic control. It also leverages a significantly better foundation model, Mistral 7B.
by nousresearch4K context$0.18/M input tkns$0.18/M output tkns5K tokens this week - MythoMist 7B (free)
From the creator of MythoMax, merges a suite of models to reduce word anticipation, ministrations, and other undesirable words in ChatGPT roleplaying data. It combines Neural Chat 7B, Airoboros 7b, Toppy M 7B, Zepher 7b beta, Nous Capybara 34B, OpenHeremes 2.5, and many others. #merge Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by gryphe33K context$0/M input tkns$0/M output tkns618K tokens this week - MythoMist 7B
From the creator of MythoMax, merges a suite of models to reduce word anticipation, ministrations, and other undesirable words in ChatGPT roleplaying data. It combines Neural Chat 7B, Airoboros 7b, Toppy M 7B, Zepher 7b beta, Nous Capybara 34B, OpenHeremes 2.5, and many others. #merge
by gryphe33K context$0.375/M input tkns$0.375/M output tkns107K tokens this week - Nous: Capybara 34B
This model is trained on the Yi-34B model for 3 epochs on the Capybara dataset. It's the first 34B Nous model and first 200K context length Nous model. Note: This endpoint currently supports 32k context.
by nousresearch33K context$0.9/M input tkns$0.9/M output tkns405K tokens this week - Toppy M 7B (free)
A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 - HuggingFaceH4/zephyr-7b-beta - lemonilia/AshhLimaRP-Mistral-7B - Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b - Undi95/Mistral-pippa-sharegpt-7b-qlora #merge #uncensored Note: this is a free, rate-limited version of this model. Outputs may be cached. Read about rate limits here.
by undi954K context$0/M input tkns$0/M output tkns2.33M tokens this week - Toppy M 7B
A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 - HuggingFaceH4/zephyr-7b-beta - lemonilia/AshhLimaRP-Mistral-7B - Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b - Undi95/Mistral-pippa-sharegpt-7b-qlora #merge #uncensored
by undi954K context$0.07/M input tkns$0.07/M output tkns1.33M tokens this week - Nous: Hermes 13B
A state-of-the-art language model fine-tuned on over 300k instructions by Nous Research, with Teknium and Emozilla leading the fine tuning process.
by nousresearch4K context$0.18/M input tkns$0.18/M output tkns7.96M tokens this week