Skip to content
  1.  
  2. © 2023 – 2025 OpenRouter, Inc
    Favicon for OpenInference

    OpenInference

    Browse models provided by OpenInference (Terms of Service)

    2 models

    Tokens processed on OpenRouter

    • Qwen: Qwen3 Coder 480B A35BQwen3 Coder 480B A35BFree variant

      Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over repositories. The model features 480 billion total parameters, with 35 billion active per forward pass (8 out of 160 experts). Pricing for the Alibaba endpoints varies by context length. Once a request is greater than 128k input tokens, the higher pricing is used.

    by qwen1.05M context$0/M input tokens$0/M output tokens
  3. MoonshotAI: Kimi K2 0711Kimi K2 0711Free variant

    Kimi K2 Instruct is a large-scale Mixture-of-Experts (MoE) language model developed by Moonshot AI, featuring 1 trillion total parameters with 32 billion active per forward pass. It is optimized for agentic capabilities, including advanced tool use, reasoning, and code synthesis. Kimi K2 excels across a broad range of benchmarks, particularly in coding (LiveCodeBench, SWE-bench), reasoning (ZebraLogic, GPQA), and tool-use (Tau2, AceBench) tasks. It supports long-context inference up to 128K tokens and is designed with a novel training stack that includes the MuonClip optimizer for stable large-scale MoE training.

    by moonshotai131K context$0/M input tokens$0/M output tokens