Search/
Skip to content
/
OpenRouter
© 2026 OpenRouter, Inc

Product

  • Chat
  • Rankings
  • Apps
  • Models
  • Providers
  • Pricing
  • Enterprise
  • Labs

Company

  • About
  • Announcements
  • CareersHiring
  • Privacy
  • Terms of Service
  • Support
  • State of AI
  • Works With OR
  • Data

Developer

  • Documentation
  • API Reference
  • SDK
  • Status

Connect

  • Discord
  • GitHub
  • LinkedIn
  • X
  • YouTube
Collections/Embedding Models

Text Embedding Models

Model rankings updated April 2026 based on real usage data.

Embedding models convert text into dense vector representations, enabling semantic search, retrieval-augmented generation (RAG), clustering, and similarity matching. OpenRouter provides access to leading embedding models through a single API gateway, so you can test models and compare performance and pricing without managing multiple provider integrations.

Whether you're building a knowledge base, powering search across documents, or feeding context into an LLM pipeline, these are the most popular embedding models available on OpenRouter today.

Embedding Models on OpenRouter

Favicon for qwen

Qwen: Qwen3 Embedding 8B

59.9B tokens

The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks. This series inherits the exceptional multilingual capabilities, long-text understanding, and reasoning skills of its foundational model. The Qwen3 Embedding series represents significant advancements in multiple text embedding and ranking tasks, including text retrieval, code retrieval, text classification, text clustering, and bitext mining.

by qwen32K context$0.01/M input tokens$0/M output tokens
Favicon for openai

OpenAI: Text Embedding 3 Small

54.9B tokens

text-embedding-3-small is OpenAI's improved, more performant version of the ada embedding model. Embeddings are a numerical representation of text that can be used to measure the relatedness between two pieces of text. Embeddings are useful for search, clustering, recommendations, anomaly detection, and classification tasks.

by openai8K context$0.02/M input tokens$0/M output tokens
Favicon for openai

OpenAI: Text Embedding 3 Large

16.1B tokens

text-embedding-3-large is OpenAI's most capable embedding model for both english and non-english tasks. Embeddings are a numerical representation of text that can be used to measure the relatedness between two pieces of text. Embeddings are useful for search, clustering, recommendations, anomaly detection, and classification tasks.

by openai8K context$0.13/M input tokens$0/M output tokens
Favicon for baai

BAAI: bge-m3

11.7B tokens

The bge-m3 embedding model encodes sentences, paragraphs, and long documents into a 1024-dimensional dense vector space, delivering high-quality semantic embeddings optimized for multilingual retrieval, semantic search, and large-context applications.

by baai8K context$0.01/M input tokens$0/M output tokens
Favicon for google

Google: Gemini Embedding 001

8.47B tokens

gemini-embedding-001 provides a unified cutting edge experience across domains, including science, legal, finance, and coding. This embedding model has consistently held a top spot on the Massive Text Embedding Benchmark (MTEB) Multilingual leaderboard since the experimental launch in March.

by google20K context$0.15/M input tokens$0/M output tokens
Favicon for openai

OpenAI: Text Embedding Ada 002

7.77B tokens

text-embedding-ada-002 is OpenAI's legacy text embedding model.

by openai8K context$0.10/M input tokens$0/M output tokens
Favicon for perplexity

Perplexity: Embed V1 4B

4.96B tokens

pplx-embed-v1 -4B is one of Perplexity's state-of-the-art text embedding models built for real-world, web-scale retrieval. pplx-embed-v1 is optimized for standard dense text retrieval with the 4B parameter model maximizing retrieval quality.

by perplexity32K context$0.03/M input tokens$0/M output tokens
Favicon for nvidia

NVIDIA: Llama Nemotron Embed VL 1B V2 (free)

4.24B tokens

The Llama Nemotron Embed VL 1B V2 embedding model is optimized for multimodal question-answering retrieval. The model can embed 'documents' in the form of image, text, or image and text combined. Documents can be retrieved given a user query in text form. The model supports images containing text, tables, charts, and infographics.

by nvidia131K context$0/M input tokens$0/M output tokens
Favicon for mistralai

Mistral: Mistral Embed 2312

3.23B tokens

Mistral Embed is a specialized embedding model for text data, optimized for semantic search and RAG applications. Developed by Mistral AI in late 2023, it produces 1024-dimensional vectors that effectively capture semantic relationships in text.

by mistralai8K context$0.10/M input tokens$0/M output tokens
Favicon for mistralai

Mistral: Codestral Embed 2505

2.95B tokens

Mistral Codestral Embed is specially designed for code, perfect for embedding code databases, repositories, and powering coding assistants with state-of-the-art retrieval.

by mistralai8K context$0.15/M input tokens$0/M output tokens