ModelContext LengthModel Type
sonar-reasoning-pro128kChat Completion
sonar-reasoning128kChat Completion
sonar-pro200kChat Completion
sonar128kChat Completion
r1-1776128kChat Completion
  1. sonar-reasoning-pro and sonar-pro have a max output token limit of 8k
  2. The reasoning models output CoTs in their responses as well
  3. r1-1776 is an offline chat model that does not use our search subsystem

Legacy Models

These models will be deprecated and will no longer be available to use after 2/22/2025

ModelContext LengthModel Type
llama-3.1-sonar-small-128k-online127kChat Completion
llama-3.1-sonar-large-128k-online127kChat Completion
llama-3.1-sonar-huge-128k-online127kChat Completion