Models
- Bagel 34B v0.2
An experimental fine-tune of Yi 34b 200k using bagel. This is the version of the fine-tune before direct preference optimization (DPO) has been applied. DPO performs better on benchmarks, but this version is likely better for creative writing, roleplay, etc.
by jondurbin8K context - Airoboros 70B
A Llama 2 70B fine-tune using synthetic data (the Airoboros dataset). Currently based on jondurbin/airoboros-l2-70b-2.2.1, but might get updated in the future.
by jondurbin4K context$0.7/M input tkns$0.9/M output tkns2.65M tokens this week