Skip to content
  •  
  • © 2023 – 2025 OpenRouter, Inc
      Favicon for Perplexity

      Perplexity

      Browse models provided by Perplexity (Terms of Service)

      8 models

      Tokens processed

      • Perplexity: Sonar Reasoning Pro

        Note: Sonar Pro pricing includes Perplexity search pricing. See details here Sonar Reasoning Pro is a premier reasoning model powered by DeepSeek R1 with Chain of Thought (CoT). Designed for advanced use cases, it supports in-depth, multi-step queries with a larger context window and can surface more citations per search, enabling more comprehensive and extensible responses.

        by perplexity128K context$2/M input tokens$8/M output tokens
      • Perplexity: Sonar Pro

        Note: Sonar Pro pricing includes Perplexity search pricing. See details here For enterprises seeking more advanced capabilities, the Sonar Pro API can handle in-depth, multi-step queries with added extensibility, like double the number of citations per search as Sonar on average. Plus, with a larger context window, it can handle longer and more nuanced searches and follow-up questions.

        by perplexity200K context$3/M input tokens$15/M output tokens
      • Perplexity: Sonar Deep Research

        Sonar Deep Research is a research-focused model designed for multi-step retrieval, synthesis, and reasoning across complex topics. It autonomously searches, reads, and evaluates sources, refining its approach as it gathers information. This enables comprehensive report generation across domains like finance, technology, health, and current events. Notes on Pricing (Source) - Input tokens comprise of Prompt tokens (user prompt) + Citation tokens (these are processed tokens from running searches) - Deep Research runs multiple searches to conduct exhaustive research. Searches are priced at $5/1000 searches. A request that does 30 searches will cost $0.15 in this step. - Reasoning is a distinct step in Deep Research since it does extensive automated reasoning through all the material it gathers during its research phase. Reasoning tokens here are a bit different than the CoTs in the answer - these are tokens that we use to reason through the research material prior to generating the outputs via the CoTs. Reasoning tokens are priced at $3/1M tokens

        by perplexity128K context$2/M input tokens$8/M output tokens
      • Perplexity: R1 1776

        R1 1776 is a version of DeepSeek-R1 that has been post-trained to remove censorship constraints related to topics restricted by the Chinese government. The model retains its original reasoning capabilities while providing direct responses to a wider range of queries. R1 1776 is an offline chat model that does not use the perplexity search subsystem. The model was tested on a multilingual dataset of over 1,000 examples covering sensitive topics to measure its likelihood of refusal or overly filtered responses. Evaluation Results Its performance on math and reasoning benchmarks remains similar to the base R1 model. Reasoning Performance Read more on the Blog Post

        by perplexity128K context$2/M input tokens$8/M output tokens
      • Perplexity: Sonar Reasoning

        Sonar Reasoning is a reasoning model provided by Perplexity based on DeepSeek R1. It allows developers to utilize long chain of thought with built-in web search. Sonar Reasoning is uncensored and hosted in US datacenters.

        by perplexity127K context$1/M input tokens$5/M output tokens$5/K reqs
      • Perplexity: Sonar

        Sonar is lightweight, affordable, fast, and simple to use — now featuring citations and the ability to customize sources. It is designed for companies seeking to integrate lightweight question-and-answer features optimized for speed.

        by perplexity127K context$1/M input tokens$1/M output tokens$5/K reqs
      • Perplexity: Llama 3.1 Sonar 70B Online

        Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the offline chat model. It is focused on delivering helpful, up-to-date, and factual responses. #online

        by perplexity127K context$1/M input tokens$1/M output tokens$5/K reqs
      • Perplexity: Llama 3.1 Sonar 8B Online

        Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the offline chat model. It is focused on delivering helpful, up-to-date, and factual responses. #online

        by perplexity127K context$0.20/M input tokens$0.20/M output tokens$5/K reqs