Multimodal Capabilities

Send images, PDFs, audio, and video to OpenRouter models

OpenRouter supports multiple input modalities beyond text, allowing you to send images, PDFs, audio, and video files to compatible models through our unified API. This enables rich multimodal interactions for a wide variety of use cases.

Supported Modalities

Images

Send images to vision-capable models for analysis, description, OCR, and more. OpenRouter supports multiple image formats and both URL-based and base64-encoded images.

Learn more about image inputs →

Image Generation

Generate images from text prompts using AI models with image output capabilities. OpenRouter supports various image generation models that can create high-quality images based on your descriptions.

Learn more about image generation →

PDFs

Process PDF documents with any model on OpenRouter. Our intelligent PDF parsing system extracts text and handles both text-based and scanned documents.

Learn more about PDF processing →

Audio

Send audio files to speech-capable models for transcription, analysis, and processing. OpenRouter supports common audio formats with automatic routing to compatible models.

Learn more about audio inputs →

Video

Send video files to video-capable models for analysis, description, object detection, and action recognition. OpenRouter supports multiple video formats for comprehensive video understanding tasks.

Learn more about video inputs →

Getting Started

All multimodal inputs use the same /api/v1/chat/completions endpoint with the messages parameter. Different content types are specified in the message content array:

  • Images: Use image_url content type
  • PDFs: Use file content type with PDF data
  • Audio: Use input_audio content type
  • Video: Use input_video content type

You can combine multiple modalities in a single request, and the number of files you can send varies by provider and model.

Model Compatibility

Not all models support every modality. OpenRouter automatically filters available models based on your request content:

  • Vision models: Required for image processing
  • File-compatible models: Can process PDFs natively or through our parsing system
  • Audio-capable models: Required for audio input processing
  • Video-capable models: Required for video input processing

Use our Models page to find models that support your desired input modalities.

Input Format Support

OpenRouter supports both direct URLs and base64-encoded data for multimodal inputs:

  • Images: https://example.com/image.jpg
  • PDFs: https://example.com/document.pdf
  • Audio: Not supported via URL (base64 only)
  • Video: Provider-specific (e.g., YouTube links for Gemini on AI Studio)

Base64 Encoding (Required for local files)

  • Images: data:image/jpeg;base64,{base64_data}
  • PDFs: data:application/pdf;base64,{base64_data}
  • Audio: Raw base64 string with format specification
  • Video: data:video/mp4;base64,{base64_data}

URLs are more efficient for large files as they don’t require local encoding and reduce request payload size. Base64 encoding is required for local files or when the content is not publicly accessible.

Note for video URLs: Video URL support varies by provider. For example, Google Gemini on AI Studio only supports YouTube links. See the video inputs documentation for provider-specific details.

Frequently Asked Questions

Yes! You can send text, images, PDFs, audio, and video in the same request. The model will process all inputs together.

  • Images: Typically priced per image or as input tokens
  • PDFs: Free text extraction, paid OCR processing, or native model pricing
  • Audio: Priced as input tokens based on duration
  • Video: Priced as input tokens based on duration and resolution

Video support varies by model. Use the Models page to filter for video-capable models. Check each model’s documentation for specific video format and duration limits.