Supported Models
Perplexity Models
Model | Parameter Count | Context Length | Model Type |
---|---|---|---|
llama-3-sonar-small-32k-online * | 8B | 28,000 | Chat Completion |
llama-3-sonar-small-32k-chat | 8B | 32,768 | Chat Completion |
llama-3-sonar-large-32k-online * | 70B | 28,000 | Chat Completion |
llama-3-sonar-large-32k-chat | 70B | 32,768 | Chat Completion |
*Note that the search subsystem of the Online LLMs do not attend to the system prompt. You can use the system prompt to provide instructions related to style, tone, and language of the response.
Access to citations and images via API is in closed beta. To request access to citations, fill out this form and send an email describing your use case to [email protected].
Open-Source Models
Where possible, we try to match the Hugging Face implementation.
Model | Parameter Count | Context Length | Model Type |
---|---|---|---|
llama-3-8b-instruct | 8B | 8,192 | Chat Completion |
llama-3-70b-instruct | 70B | 8,192 | Chat Completion |
mixtral-8x7b-instruct | 8x7B | 16,384 | Chat Completion |
Special Tokens
We do not raise any exceptions if your chat inputs contain messages with special tokens. If avoiding prompt injections is a concern for your use case, it is recommended that you check for special tokens prior to calling the API. For more details, read Meta's recommendations for Llama.
Updated 11 days ago