Search/
Skip to content
/
OpenRouter
© 2026 OpenRouter, Inc

Product

  • Chat
  • Rankings
  • Apps
  • Models
  • Providers
  • Pricing
  • Enterprise
  • Labs

Company

  • About
  • Announcements
  • CareersHiring
  • Privacy
  • Terms of Service
  • Support
  • State of AI
  • Works With OR
  • Data

Developer

  • Documentation
  • API Reference
  • SDK
  • Status

Connect

  • Discord
  • GitHub
  • LinkedIn
  • X
  • YouTube
Collections/Coding

Best AI Models for Coding

Model rankings updated April 2026 based on real usage data.

Compare the best AI models for coding, ranked by real usage from developers on OpenRouter. Whether you're generating code, debugging, refactoring or building an AI coding assistant, these LLMs deliver strong performance across popular languages and frameworks.

This collection features top coding models from Anthropic, Google, xAI, OpenAI and more, all accessible through a single API. From agentic coding workflows to one-off code generation, find the right model for your engineering needs.

LLM Leaderboard for Programming Models

1.
Kimi K2.6
by moonshotai
427B
29.2%
2.
Hy3 Preview (free)
by tencent
176B
12.0%
3.
Step 3.5 Flash
by stepfun
73B
5.0%
4.
Claude Opus 4.7
by anthropic
64.5B
4.4%
5.
Ling 2.6 1t (free)
by inclusionai
63.7B
4.3%
6.
Nemotron 3 Super 120B A12B (free)
by nvidia
60.3B
4.1%
7.
Claude Sonnet 4.6
by anthropic
54.2B
3.7%
8.
Minimax M2.7
by minimax
46.8B
3.2%
9.
GLM 5.1
by z-ai
44.3B
3.0%
10.
Others
by unknown
455B
31.1%

Top Coding Models on OpenRouter

Based on top weekly usage data from millions of users accessing AI models for coding through OpenRouter.

Favicon for moonshotai

MoonshotAI Kimi Latest

2.09T tokens

This model always redirects to the latest model in the MoonshotAI Kimi family.

by moonshotai256K context$0.7448/M input tokens$4.655/M output tokens
Favicon for moonshotai

MoonshotAI: Kimi K2.6

2.09T tokens

Kimi K2.6 is Moonshot AI's next-generation multimodal model, designed for long-horizon coding, coding-driven UI/UX generation, and multi-agent orchestration. It handles complex end-to-end coding tasks across Python, Rust, and Go, and can convert prompts and visual inputs into production-ready interfaces. Its agent swarm architecture scales to hundreds of parallel sub-agents for autonomous task decomposition - delivering documents, websites, and spreadsheets in a single run without human oversight.

by moonshotai256K context$0.7448/M input tokens$4.655/M output tokens
Favicon for anthropic

Anthropic Claude Sonnet Latest

1.51T tokens

This model always redirects to the latest model in the Anthropic Claude Sonnet family.

by anthropic1M context$3/M input tokens$15/M output tokens
Favicon for anthropic

Anthropic: Claude Sonnet 4.6

1.51T tokens

Sonnet 4.6 is Anthropic's most capable Sonnet-class model yet, with frontier performance across coding, agents, and professional work. It excels at iterative development, complex codebase navigation, end-to-end project management with memory, polished document creation, and confident computer use for web QA and workflow automation.

by anthropic1M context$3/M input tokens$15/M output tokens
Favicon for deepseek

DeepSeek: DeepSeek V3.2

1.35T tokens

DeepSeek-V3.2 is a large language model designed to harmonize high computational efficiency with strong reasoning and agentic tool-use performance. It introduces DeepSeek Sparse Attention (DSA), a fine-grained sparse attention mechanism that reduces training and inference cost while preserving quality in long-context scenarios. A scalable reinforcement learning post-training framework further improves reasoning, with reported performance in the GPT-5 class, and the model has demonstrated gold-medal results on the 2025 IMO and IOI. V3.2 also uses a large-scale agentic task synthesis pipeline to better integrate reasoning into tool-use settings, boosting compliance and generalization in interactive environments.

Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs

by deepseek131K context$0.252/M input tokens$0.378/M output tokens
Favicon for anthropic

Anthropic: Claude Opus Latest

1.28T tokens

This model always redirects to the latest model in the Claude Opus family.

by anthropic1M context$5/M input tokens$25/M output tokens
Favicon for anthropic

Anthropic: Claude Opus 4.7

1.28T tokens

Opus 4.7 is the next generation of Anthropic's Opus family, built for long-running, asynchronous agents. Building on the coding and agentic strengths of Opus 4.6, it delivers stronger performance on complex, multi-step tasks and more reliable agentic execution across extended workflows. It is especially effective for asynchronous agent pipelines where tasks unfold over time - large codebases, multi-stage debugging, and end-to-end project orchestration.

Beyond coding, Opus 4.7 brings improved knowledge work capabilities - from drafting documents and building presentations to analyzing data. It maintains coherence across very long outputs and extended sessions, making it a strong default for tasks that require persistence, judgment, and follow-through.

For users upgrading from earlier Opus versions, see our official migration guide here

by anthropic1M context$5/M input tokens$25/M output tokens
Favicon for google

Google Gemini Flash Latest

1.13T tokens

This model always redirects to the latest model in the Google Gemini Flash family.

by google1.05M context$0.50/M input tokens$3/M output tokens$1/M audio tokens
Favicon for google

Google: Gemini 3 Flash Preview

1.13T tokens

Gemini 3 Flash Preview is a high speed, high value thinking model designed for agentic workflows, multi turn chat, and coding assistance. It delivers near Pro level reasoning and tool use performance with substantially lower latency than larger Gemini variants, making it well suited for interactive development, long running agent loops, and collaborative coding tasks. Compared to Gemini 2.5 Flash, it provides broad quality improvements across reasoning, multimodal understanding, and reliability.

The model supports a 1M token context window and multimodal inputs including text, images, audio, video, and PDFs, with text output. It includes configurable reasoning via thinking levels (minimal, low, medium, high), structured output, tool use, and automatic context caching. Gemini 3 Flash Preview is optimized for users who want strong reasoning and agentic behavior without the cost or latency of full scale frontier models.

by google1.05M context$0.50/M input tokens$3/M output tokens$1/M audio tokens
Favicon for stepfun

StepFun: Step 3.5 Flash

850B tokens

Step 3.5 Flash is StepFun's most capable open-source foundation model. Built on a sparse Mixture of Experts (MoE) architecture, it selectively activates only 11B of its 196B parameters per token. It is a reasoning model that is incredibly speed efficient even at long contexts.

by stepfun262K context$0.10/M input tokens$0.30/M output tokens