Perplexity home pagelight logodark logo
  • Playground
  • Playground
Changelog
Changelog
Home
Models
Guides
API Reference
Changelog
System Status
FAQ
Roadmap
Discussions
Changelog
  • Overview
Changelog

Changelog

We are excited to announce the release of our new API portal, designed to help you better manage your organization and API usage.

With this portal, you can:

  • Organize and manage your API keys more effectively.
  • Gain insights into your API usage and team activity.
  • Streamline collaboration within your organization.

Check it out here:
https://www.perplexity.ai/account/api/group

Looking to narrow down your search results based on users’ locations?
We now support user location filtering, allowing you to retrieve results only from a particular user location.

Check out the guide:
https://docs.perplexity.ai/guides/user-locaiton-filter-guide

You can now upload images to Sonar and use them as part of your multimodal search experience.
Give it a try by following our image upload guide:
https://docs.perplexity.ai/guides/image-guide

Looking to narrow down your search results to specific dates?
We now support date range filtering, allowing you to retrieve results only from a particular timeframe.

Check out the guide:
https://docs.perplexity.ai/guides/date-range-filter-guide

We’ve fully transitioned to our new pricing model: citation tokens are no longer charged.
If you were already using the search_context_size parameter, you’ve been on this model already.

This change makes pricing simpler and cheaper for everyone — with no downside.

View the updated pricing:
https://docs.perplexity.ai/guides/pricing

We’ve removed all feature gating based on tiered spending. These were previously only available to users of Tier 3 and above.

That means every user now has access to all API capabilities, regardless of usage volume or spend. Rate limits are still applicable.
Whether you’re just getting started or scaling up, you get the full power of Sonar out of the box.

We’re excited to announce that structured outputs are now available to all Perplexity API users, regardless of tier level. Based on valuable feedback from our developer community, we’ve removed the previous Tier 3 requirement for this feature.

What’s available now:

  • JSON structured outputs are supported across all models
  • Both JSON and Regex structured outputs are supported for sonar and sonar-reasoning models

Coming soon:

  • Full Regex support for all models

This change allows developers to create more reliable and consistent applications from day one. We believe in empowering our community with the tools they need to succeed, and we’re committed to continuing to improve accessibility to our advanced features.

Thank you for your feedback—it helps us make Perplexity API better for everyone.

We’re excited to announce significant improvements to our Sonar models that deliver superior performance at lower costs. Our latest benchmark testing confirms that Sonar and Sonar Pro now outperform leading competitors while maintaining more affordable pricing.

Key updates include:

  • Three new search modes across most Sonar models:

    • High: Maximum depth for complex queries
    • Medium: Balanced approach for moderate complexity
    • Low: Cost-efficient for straightforward queries (equivalent to current pricing)
  • Simplified billing structure:

    • Transparent pricing for input/output tokens
    • No charges for citation tokens in responses (except for Sonar Deep Research)

The current billing structure will be supported as the default option for 30 days (until April 18, 2025). During this period, the new search modes will be available as opt-in features.

Important Note: After April 18, 2025, Sonar Pro and Sonar Reasoning Pro will not return Citation tokens or number of search results in the usage field in the API response.

Please note that as of February 22, 2025, several models and model name aliases will no longer be accessible. The following model names will no longer be available via API:

llama-3.1-sonar-small-128k-online

llama-3.1-sonar-large-128k-online

llama-3.1-sonar-huge-128k-online

We recommend updating your applications to use our recently released Sonar or Sonar Pro models – you can learn more about them here. Thank you for being a Perplexity API user.

We are expanding API offerings with the most efficient and cost-effective search solutions available: Sonar and Sonar Pro.

Sonar gives you fast, straightforward answers

Sonar Pro tackles complex questions that need deeper research and provides more sources

Both models offer built-in citations, automated scaling of rate limits, and public access to advanced features like structured outputs and search domain filters. And don’t worry, we never train on your data. Your information stays yours.

You can learn more about our new APIs here - http://sonar.perplexity.ai/

We are excited to announce the public availability of citations in the Perplexity API. In addition, we have also increased our default rate limit for the sonar online models to 50 requests/min for all users.

Effective immediately, all API users will see citations returned as part of their requests by default. This is not a breaking change. The return_citations parameter will no longer have any effect.

If you have any questions or need assistance, feel free to reach out to our team at api@perplexity.ai

We are excited to announce the launch of our latest Perplexity Sonar models:

Online Models - llama-3.1-sonar-small-128k-online llama-3.1-sonar-large-128k-online

Chat Models - llama-3.1-sonar-small-128k-chat llama-3.1-sonar-large-128k-chat

These new additions surpass the performance of the previous iteration. For detailed information on our supported models, please visit our model card documentation.

[Action Required] Model Deprecation Notice Please note that several models will no longer be accessible effective 8/12/2024. We recommend updating your applications to use models in the Llama-3.1 family immediately.

The following model names will no longer be available via API - llama-3-sonar-small-32k-online llama-3-sonar-large-32k-online llama-3-sonar-small-32k-chat llama-3-sonar-large-32k-chat llama-3-8b-instruct llama-3-70b-instruct mistral-7b-instruct mixtral-8x7b-instruct

We recommend switching to models in the Llama-3.1 family:

Online Models - llama-3.1-sonar-small-128k-online llama-3.1-sonar-large-128k-online

Chat Models - llama-3.1-sonar-small-128k-chat llama-3.1-sonar-large-128k-chat

Instruct Models - llama-3.1-70b-instruct llama-3.1-8b-instruct

If you have any questions, please email support@perplexity.ai. Thank you for being a Perplexity API user.

Stay curious,

Team Perplexity

Please note that as of May 14, several models and model name aliases will no longer be accessible. We recommend updating your applications to use models in the Llama-3 family immediately. The following model names will no longer be available via API:

codellama-70b-instruct mistral-7b-instruct mixtral-8x22b-instruct pplx-7b-chat pplx-7b-online sonar-small-chat sonar-small-online pplx-70b-chat pplx-70b-online pplx-8x7b-chat pplx-8x7b-online sonar-medium-chat sonar-medium-online

In lieu of the above, we recommend switching to models from the Llama 3 family:

llama-3-sonar-small-32k-chat llama-3-sonar-small-32k-online llama-3-sonar-large-32k-chat llama-3-sonar-large-32k-online llama-3-8b-instruct llama-3-70b-instruct

Effective immediately, input and output tokens are now charged with the same price. Previously, output tokens were more expensive than input tokens. Prices have generally gone down as a result.

Announcing Our Newest Model

We are excited to announce the launch of our latest Perplexity models: sonar-small-chat and sonar-medium-chat, along with their search-enhanced versions, sonar-small-online and sonar-medium-online. These new additions surpass our earlier models in cost-efficiency, speed, and performance. For detailed information on our supported models, please visit our model card documentation.

Expanded Context Windows

The context window length for several models has been doubled from 8k to 16k, including mixtral-8x7b-instruct and all Perplexity models. 4k tokens are reserved for search results in online models.

Model Deprecation Notice

Please note that as of March 15, the pplx-70b-chat, pplx-70b-online, llama-2-70b-chat, and codellama-34b-instruct models will no longer be available through the Perplexity API. We will gradually phase out less frequently used models in favor of our newer and more performant offerings.

Revised Pricing Structure for 8x7b Models

The pricing for the mixtral-8x7b-instruct model will be adjusted. Previously charged at 0.14/0.14 / 0.14/0.58 per million input and output tokens, the rates will change to 0.60/0.60 / 0.60/1.80 per million input and output tokens moving forward.

Increased Public Rate Limits

Public limits for all models have increased by ~2x. Find the current rate limits here.

Access to Citations and Elevated Rate Limits

Responding to popular demand in our API discussion forum, we are introducing URL citation access for our Online LLMs to approved users. For access to citations, or to request a rate limit increase, please complete this form.

Terms of Service and Data Processing Addendum

We wish to reiterate our commitment to data privacy for commercial application developers using the Perplexity API. The updated Terms of Service and Data Processing Addendum can be found here. Thank you for being a Perplexity API user.

We’re excited to announce that pplx-api is now serving the latest open-source mixture-of-experts model, mixtral-8x7b-instruct, at the blazingly fast speed of inference you are accustomed to.

We’re excited to share two new PPLX models: pplx-7b-online and pplx-70b-online. These first-of-a-kind models are integrated with our in-house search technology for factual grounding. Read our blog post for more information!

https://blog.perplexity.ai/blog/introducing-pplx-online-llms

We’re also announcing general availability for our API. We’ve rolled out usage-based pricing, which enables us to gradually relax the rate limits on our models. Follow the updated steps for getting started.

We have removed support for replit-code-v1.5-3b and openhermes-2-mistral-7b. There are no immediate plans to add these models back. If you were a user who enjoyed openhermes-2-mistral-7b, try instead using our in-house models, pplx-7b-chat, pplx-70b-chat!

The Perplexity AI API is currently in beta release v0. Clients are not protected from backwards incompatible changes and cannot specify their desired API version. Examples of backwards incompatible changes include…

Removing support for a given model. Renaming a response field. Removing a response field. Adding a required request parameter. Backwards incompatible changes will be documented here.

Generally, the API is currently designed to be compatible with OpenAI client libraries. This means that, given the same request body, swapping the API base URL and adding your Perplexity API key will yield a response that can be parsed in the same way as the response OpenAI would yield, except for certain explicitly unsupported body parameters documented in the reference (link to reference).

twitterlinkedindiscordgithubwebsite