Skip to content
  •  
  • © 2023 – 2025 OpenRouter, Inc
      Favicon for Phala

      Phala

      Browse models provided by Phala (Terms of Service)

      2 models

      Tokens processed

      • DeepSeek: R1 Distill Llama 70B

        DeepSeek R1 Distill Llama 70B is a distilled large language model based on Llama-3.3-70B-Instruct, using outputs from DeepSeek R1. The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: - AIME 2024 pass@1: 70.0 - MATH-500 pass@1: 94.5 - CodeForces Rating: 1633 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

        by deepseek128K context$0.20/M input tokens$0.70/M output tokens
      • Meta: Llama 3.3 70B Instruct

        The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction tuned text only model is optimized for multilingual dialogue use cases and outperforms many of the available open source and closed chat models on common industry benchmarks. Supported languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. Model Card

        by meta-llama131K context$0.12/M input tokens$0.35/M output tokens