Skip to content
  1.  
  2. © 2023 – 2025 OpenRouter, Inc
    Favicon for Google AI Studio

    Google AI Studio

    Browse models provided by Google AI Studio (Terms of Service)

    17 models

    Tokens processed on OpenRouter

    • Google: Gemini 2.5 Flash Image (Nano Banana)Gemini 2.5 Flash Image (Nano Banana)

      Gemini 2.5 Flash Image, a.k.a. "Nano Banana," is now generally available. It is a state of the art image generation model with contextual understanding. It is capable of image generation, edits, and multi-turn conversations. Aspect ratios can be controlled with the image_config API Parameter

      by google33K context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs
    $0.03/K output imgs
  3. Google: Gemini 2.5 Flash Preview 09-2025Gemini 2.5 Flash Preview 09-2025

    Gemini 2.5 Flash Preview September 2025 Checkpoint is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in "thinking" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. Additionally, Gemini 2.5 Flash is configurable through the "max tokens for reasoning" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).

    by google1.05M context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs$1/M audio tokens
  4. Google: Gemini 2.5 Flash Lite Preview 09-2025Gemini 2.5 Flash Lite Preview 09-2025

    Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the Reasoning API parameter to selectively trade off cost for intelligence.

    by google1.05M context$0.10/M input tokens$0.40/M output tokens
  5. Google: Gemini 2.5 Flash Image Preview (Nano Banana)Gemini 2.5 Flash Image Preview (Nano Banana)

    Gemini 2.5 Flash Image Preview, a.k.a. "Nano Banana," is a state of the art image generation model with contextual understanding. It is capable of image generation, edits, and multi-turn conversations.

    by google33K context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs$0.03/K output imgs
  6. Google: Gemini 2.5 Flash LiteGemini 2.5 Flash Lite

    Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the Reasoning API parameter to selectively trade off cost for intelligence.

    by google1.05M context$0.10/M input tokens$0.40/M output tokens$0.30/M audio tokens
  7. Google: Gemma 3n 2BGemma 3n 2BFree variant

    Gemma 3n E2B IT is a multimodal, instruction-tuned model developed by Google DeepMind, designed to operate efficiently at an effective parameter size of 2B while leveraging a 6B architecture. Based on the MatFormer architecture, it supports nested submodels and modular composition via the Mix-and-Match framework. Gemma 3n models are optimized for low-resource deployment, offering 32K context length and strong multilingual and reasoning performance across common benchmarks. This variant is trained on a diverse corpus including code, math, web, and multimodal data.

    by google8K context$0/M input tokens$0/M output tokens
  8. Google: Gemini 2.5 Flash Lite Preview 06-17Gemini 2.5 Flash Lite Preview 06-17

    Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the Reasoning API parameter to selectively trade off cost for intelligence.

    by google1.05M context$0.10/M input tokens$0.40/M output tokens$0.30/M audio tokens
  9. Google: Gemini 2.5 FlashGemini 2.5 Flash

    Gemini 2.5 Flash is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in "thinking" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. Additionally, Gemini 2.5 Flash is configurable through the "max tokens for reasoning" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).

    by google1.05M context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs$1/M audio tokens
  10. Google: Gemini 2.5 ProGemini 2.5 Pro

    Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.

    by google1.05M context$1.25/M input tokens$10/M output tokens$5.16/K input imgs$2.50/M audio tokens
  11. Google: Gemma 3n 4BGemma 3n 4BFree variant

    Gemma 3n E4B-it is optimized for efficient execution on mobile and low-resource devices, such as phones, laptops, and tablets. It supports multimodal inputs—including text, visual data, and audio—enabling diverse tasks such as text generation, speech recognition, translation, and image analysis. Leveraging innovations like Per-Layer Embedding (PLE) caching and the MatFormer architecture, Gemma 3n dynamically manages memory usage and computational load by selectively activating model parameters, significantly reducing runtime resource requirements. This model supports a wide linguistic range (trained in over 140 languages) and features a flexible 32K token context window. Gemma 3n can selectively load parameters, optimizing memory and computational efficiency based on the task or device capabilities, making it well-suited for privacy-focused, offline-capable applications and on-device AI solutions. Read more in the blog post

    by google32K context$0/M input tokens$0/M output tokens
  12. Google: Gemini 2.5 Pro Preview 05-06Gemini 2.5 Pro Preview 05-06

    Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.

    by google1.05M context$1.25/M input tokens$10/M output tokens$5.16/K input imgs
  13. Google: Gemma 3 4BGemma 3 4BFree variant

    Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling.

    by google131K context$0/M input tokens$0/M output tokens
  14. Google: Gemma 3 12BGemma 3 12BFree variant

    Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 12B is the second largest in the family of Gemma 3 models after Gemma 3 27B

    by google131K context$0/M input tokens$0/M output tokens
  15. Google: Gemma 3 27BGemma 3 27BFree variant

    Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 27B is Google's latest open source model, successor to Gemma 2

    by google131K context$0/M input tokens$0/M output tokens
  16. Google: Gemini 2.0 Flash LiteGemini 2.0 Flash Lite

    Gemini 2.0 Flash Lite offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5, all at extremely economical token prices.

    by google1.05M context$0.075/M input tokens$0.30/M output tokens$0.075/M audio tokens
  17. Google: Gemini 2.0 FlashGemini 2.0 Flash

    Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5. It introduces notable enhancements in multimodal understanding, coding capabilities, complex instruction following, and function calling. These advancements come together to deliver more seamless and robust agentic experiences.

    by google1M context$0.10/M input tokens$0.40/M output tokens$0.0258/K input imgs$0.70/M audio tokens
  18. Google: Gemini 2.0 Flash ExperimentalGemini 2.0 Flash ExperimentalFree variant

    Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5. It introduces notable enhancements in multimodal understanding, coding capabilities, complex instruction following, and function calling. These advancements come together to deliver more seamless and robust agentic experiences.

    by google1.05M context$0/M input tokens$0/M output tokens