ModelContext LengthModel Type
sonar-reasoning127kChat Completion
sonar-pro200kChat Completion
sonar127kChat Completion
  1. sonar-pro has a max output token limit of 8k
  2. sonar-reasoning outputs CoT in its response as well

Legacy Models

These models will be deprecated and will no longer be available to use after 2/22/2025

ModelContext LengthModel Type
llama-3.1-sonar-small-128k-online127kChat Completion
llama-3.1-sonar-large-128k-online127kChat Completion
llama-3.1-sonar-huge-128k-online127kChat Completion