Home Hub Features Use Cases How-To Guides Platform Pricing Login



Perplexity Features Deep Dive

How Perplexity Works:
Deep Research, Spaces, Pages,
Model Council, Comet, and More

Perplexity ships eleven distinct user-facing features split across four categories: research and reasoning (Deep Research, Model Council, Labs), workspace and content creation (Spaces, Pages), browser and shopping (Comet, Shopping, Finance), and core capabilities (Citations, File Uploads, Memory).

This guide covers what each feature actually does, how it works mechanically, when to use it, when not to, and the documented limitations and structural risks. For tier requirements, see the Perplexity Pricing Guide. For comparisons against ChatGPT, Claude, Gemini, and Grok equivalents, see Perplexity vs Other AI Models.

Last verified May 10, 2026. Next refresh due August 10, 2026.

See how Perplexity Works With other Four Frontier AI Models in Multi-AI Orchestrated Business Discussion




How multi-step research works
with the variable cost structure.

Deep Research is the consumer feature, and sonar-deep-research is the API model that exposes the same capability programmatically. Both run an iterative agentic loop. The system decomposes the user query into sub-queries, performs parallel web searches, reads and evaluates source documents, updates its research plan based on findings, and iterates until a synthesis threshold is reached. The output is a structured cited report.

In the consumer interface, Deep Research queries take 2 to 4 minutes per execution. The user enters a query, selects Deep Research mode, and waits for the synthesis output. The report includes numbered citations linked to source URLs, can be exported to PDF or converted to a Perplexity Page, and contains roughly 30 searches per typical complex query.

The API exposes the same capability through sonar-deep-research with one important difference. The API has a 300-second timeout requirement that developers must configure. The default request timeout in many client libraries is shorter than this, and queries can fail silently if the timeout is not extended. The recommended pattern is to set request timeout to 300 seconds for any sonar-deep-research call.

The Variable Cost Structure

sonar-deep-research does not have a fixed per-query price. Total cost is the sum of five components.

  • Input tokens at $2.00 per million.
  • Output tokens at $8.00 per million.
  • Citation tokens at $2.00 per million.
  • Reasoning tokens at $3.00 per million.
  • Search queries at $5.00 per 1,000 searches.

A single complex query (21 searches, 193,947 reasoning tokens, 19,028 citation tokens, 11,395 output tokens) costs approximately $0.82 per request based on official sample metadata. Cost projections built only on standard input and output token rates will significantly underestimate actual cost on long research queries. The reasoning tokens dominate cost on most complex queries.

Tier Availability

  • Free tier: 5 Deep Research queries per day.
  • Pro tier: approximately 500 Deep Research queries per day at average usage.
  • Enterprise Pro: 50 per month.
  • Enterprise Max: 500 per month.
  • API: pay-per-use with the variable cost structure described above.

Documented Limitations

The HLE score for Deep Research at 21.1% (announced 2025-02-14) is now markedly stale. As of May 2026, the HLE leaderboard shows Gemini 3.1 Pro Preview at 44.7% and GPT-5.4 at 41.6% at the top. Perplexity has not published an updated HLE score for current Deep Research. The original benchmark claim is accurate at publication but the position has deteriorated significantly relative to the current frontier.

The CoT (chain-of-thought) tokens are not exposed in the API response for sonar-deep-research, unlike sonar-reasoning-pro which exposes reasoning in a <think> block. This is a deliberate design decision but limits debugging visibility for developers integrating Deep Research.



Reasoning with exposed chain-of-thought.
The <think> block is a parsing consideration.

Sonar Reasoning Pro is the current premier reasoning model in the Sonar family, replacing the deprecated sonar-reasoning as of 2025-12-15. It uses enhanced multi-step chain-of-thought reasoning over real-time web search and outputs a <think> section containing reasoning tokens before the final response.

The exposed <think> block is a developer integration consideration. The response_format parameter does not strip these reasoning tokens, so developers requesting structured JSON output must implement custom parsers to extract the JSON portion of the response after the <think> block. This is a documented JSON parsing failure mode that affects integrations expecting clean structured output.

Sonar Reasoning Pro achieved a 1,143 score on the Search Arena leaderboard as of May 2026, ranking 11th globally with 29,825 votes. SimpleQA F-score is 0.858, the highest of any model at the time of testing per Suprmind’s AI Hallucination Rates and Benchmarks reference. GPQA Diamond is 62.3% per third-party leaderboard data, AIME 2025 is 77%, and MATH-500 is 92.1%.

The model uses 128K context window. Tool use and function calling are supported via the JSON output structure with the <think> prefix caveat. Cerebras inference is not used for the reasoning variants.



Persistent workspaces for related threads,
files, and custom AI instructions.

Spaces are workspace containers for related threads, files, and custom AI instructions. Users create named workspaces with optional custom instructions that apply to all threads within the space. Files can be uploaded directly (PDF, DOCX, and other formats) or pulled from connectors (SharePoint, OneDrive, Google Drive). The AI retrieves relevant sections from uploaded files at query time rather than loading the entire document into context.

Thread queries within a Space can toggle web search on or off and set recency filters. Custom AI instructions configured at the Space level apply to all threads inside that Space, so users can configure a “research assistant” Space with specific behavioral instructions and have those instructions apply to every conversation within it.

The persistence model differs from standard threads. Standard threads use a 7-day auto-purge for uploaded files. Spaces files persist until explicitly deleted. For workflows that require long-term reference document retention, Spaces is the structural fit.

Tier Availability

  • Free: no Spaces access.
  • Pro: up to 50 files per Space.
  • Enterprise Pro: up to 500 files per Space.
  • Enterprise Max: up to 5,000 files per Space.
  • File size limits: 40 MB per file in consumer Spaces, 50 MB per file in Enterprise. Enterprise customers also get a 500-file organization repository.

Documented limitation: context window can fill with project files in long sessions, leaving limited space for conversation. Users running multi-thread Spaces with large file sets may hit context constraints inside individual threads even if the Space file count is below the tier cap.



Knowledge creation with inline citations.
The export limitation is the headline gap.

Pages is Perplexity’s knowledge-creation feature. Users enter a topic and select an audience level (beginner or expert). Perplexity generates a multi-section article with inline citations and images sourced from current web data. The resulting Page is published on perplexity.ai with a shareable URL. Sections can be reordered, previewed, and unpublished by the creator. Pages can be added to Spaces.

The output format is a structured article with H1, H2, and H3 headings, image embeds, and inline numbered citations linking to source URLs. The article is automatically formatted for web reading and includes a publication URL for sharing.

Tier Availability and Limitations

Free tier provides basic Pages access (limited features). Pro tier provides full Pages including expert-level content generation, customizable sections, and media addition. Available via web and mobile.

Documented limitation: Pages cannot be exported to PDF, WordPress, or external CMS as of early 2026. This is the most cited Pages limitation in user feedback. For workflows that require export to external content management systems, Pages is structurally limited and the workflow either ends at perplexity.ai or requires manual content reconstruction in the destination CMS.



Multi-model dispatch with synthesis.
Architecturally distinct from shared-thread orchestration.

Model Council is Perplexity’s multi-model orchestration feature, launched 2026-02-05 and available exclusively at the Max tier ($200 per month) and Enterprise Max tier ($325 per seat per month). The feature dispatches a single user query to three frontier models simultaneously and produces a synthesis output.

The current configuration runs Claude Opus 4.6, GPT-5.2, and Gemini 3 Pro. The user submits one query. The query is sent to all three models in parallel. Each model produces an independent response. A separate synthesis model (the chair model) processes the three responses and produces a comparison output that explicitly surfaces points of agreement, points of disagreement, and unique insights from each model. The user sees both the individual model outputs and the synthesis.

Architectural Distinction

Worth flagging because the positioning overlaps with multi-model orchestration platforms. Model Council is parallel dispatch with synthesis: three models receive the same query independently, do not see each other’s responses, and the chair model summarizes after the fact.

Multi-model orchestration platforms run models in a shared conversation thread where each model reads what the others said before responding. The architectural difference produces measurably different outputs. Per the Suprmind Multi-Model Divergence Index, April 2026 Edition (n=1,324 production turns), 99.1% of multi-model turns produce at least one contradiction, correction, or unique insight that single-model use would miss. Shared-thread orchestration captures cross-model corrections embedded in the response sequence rather than reported in a separate synthesis layer.

Both patterns have legitimate use cases. Model Council is the structural fit when three independent perspectives on a single question is the desired output. Shared-thread orchestration is the structural fit when models challenge each other and produce a refined answer through iteration.

Tier Availability and Limits

Max ($200 per month) and Enterprise Max ($325 per seat per month) only. Not available on Pro. Plans announced for future Pro tier expansion but not yet implemented as of the research date. Web only at launch. Three models fixed in current configuration. Users cannot select which three models participate. The $200 per month Max tier creates a high cost barrier for professional users who want multi-model validation.



Self-supervised project assistant
with a 10-minute work cycle.

Perplexity Labs is a self-supervised work cycle that runs approximately 10 minutes per task. Labs combines web browsing, code execution, chart and image generation, and asset creation to deliver structured outputs. Unlike Deep Research, which primarily synthesizes sources into a report, Labs executes code and creates interactive or exportable files.

Output deliverables include reports, spreadsheets, dashboards, and web apps. Labs positions closer to an agentic project assistant than a research tool. The user describes a project objective, Labs runs the work cycle, and the output is a set of files or a deployed asset.

Tier Availability

Pro, Max, Enterprise Pro, Enterprise Max. Free tier does not have Labs access. The “Create files and apps” query quota differentiates tiers: Pro gets monthly average-use limits, Enterprise Max gets 500 per month.

Documented limitation: 10-minute maximum self-supervised work cycle per task is the documented hard limit. Output format constraints are not publicly detailed beyond the general structured-deliverable category. Labs is relatively new and user feedback is still developing.



AI-native desktop browser.
Sidecar AI assistant in every tab.

Comet is Perplexity’s AI-native desktop browser for Mac and Windows, built on Chromium with a sidecar AI assistant embedded in every tab. The assistant can answer questions about the current page, summarize content, perform cross-tab tasks, manage email, handle shopping, and execute background agentic tasks.

Comet launched 2025-07-09 as Max-only and was made free for all users worldwide on 2025-10-01. The browser download is free. Feature access within Comet is gated by Perplexity subscription tier, with Free users getting limited daily queries and Max users getting full access including the background “Comet Assistant” agentic capability.

Comet Plus Add-On

$5 per month add-on available to all Comet users. Bundles premium publisher content access from CNN, Washington Post, Fortune, LA Times, and Condé Nast properties. Works independently of the Perplexity subscription tier, making it the cheapest paid component in Perplexity’s pricing surface.

Comet Assistant

The background agentic capability. Available at Max and Enterprise Max tiers. Enterprise Pro includes 80 Comet Assistant tasks per month. Enterprise Max includes 800. Tasks include cross-tab research, email triage, scheduled web monitoring, and similar agentic workflows that run without continuous user input.

Tier Availability

Comet browser download is free for all users worldwide. Feature access within Comet is gated by Perplexity subscription tier. Comet Plus is a $5 per month add-on for the publisher content bundle, available to all Comet users including Free.



Conversational product search
with PayPal-backed checkout.

Perplexity Shopping interprets purchase intent from conversational queries, retrieves live product data (pricing, availability, specifications, reviews) from integrations with Shopify, Amazon, BigCommerce, and other marketplaces, and presents curated product cards.

Three core capabilities differentiate Shopping from standard web search. “Snap to Shop” allows image uploads to find visually similar products. “Instant Buy” is a checkout button (built with PayPal, supporting 5,000+ merchants) that enables in-session checkout without leaving Perplexity. “Buy with Pro” is a direct purchase mechanism for supported merchants available to Pro subscribers.

The launch timeline: late 2024 saw the initial “Buy with Pro” button. November 2025 expanded the feature with free AI shopping for all US users.

Tier Availability and Geographic Scope

Free for all US users (web and desktop) as of 2025-11. Instant Buy and checkout features require Pro. US-only initially. Amazon checkout redirects to Amazon rather than completing in Perplexity.



Market data partnerships and personalized news.
Two surfaces leveraging the citation system.

Finance: Market Data and Enterprise Partnerships

The Perplexity Finance hub at perplexity.ai/finance combines real-time web search with structured financial data endpoints to answer queries about stocks, markets, earnings, and economic indicators. Basic financial queries are available across all tiers.

Enterprise Max specifically partners with PitchBook and Wiley for expanded financial and academic data access. The PitchBook partnership is a meaningful differentiator for enterprise customers in private capital markets and venture intelligence workflows.

Discover: Personalized News Feed

Discover is a curated news and content recommendation feed within Perplexity and the Comet browser. The feed surfaces content based on user interests and search history, comparable in function to OpenAI’s Pulse.

Standalone launch date is not documented in public materials. The feature is bundled with Comet and the Perplexity app rather than launched as a separate product. Discover availability spans all tiers within the Comet browser and Perplexity app.



Real URLs, sometimes fabricated content.
The structural failure mode.

The Citation System is core to Perplexity’s product positioning and was built into the original product rather than launched separately. At generation time, Perplexity’s search index retrieves candidate documents. The LLM generates a response and attaches numbered inline citation markers ([1], [2], etc.) to claims in the response. The API response includes a citations array of URLs and a search_results array with title, URL, date, and snippet for each cited source.

Sonar Pro delivers approximately 2x more citations than standard Sonar. The search_context_size parameter (Low, Medium, High) controls how much retrieved web evidence is injected per turn, which affects citation density and accuracy.

CJR Audit Result

The Columbia Journalism Review’s Tow Center tested eight AI search platforms in 2025-03 on news article citation tasks. Perplexity Sonar Pro answered 37% of queries incorrectly, the lowest error rate among tested platforms. ChatGPT Search: 67%. Grok 3: 94%. The 37% rate is the best in the field but still substantial in absolute terms.

A separate measurement on the “Pro variant” specifically reported 45% error rate per the Suprmind AI Hallucination Rates and Benchmarks reference. The variant-level distinction matters because the Pro variant is the model most users on the Pro subscription receive in standard usage.

The Structural Failure Mode

Per Suprmind’s AI Hallucination Rates and Benchmarks reference (May 2026 update), Perplexity’s hallucination pattern is structurally distinct from non-citation hallucination. The model cites real URLs with content that may be fabricated. The URL is genuine. The claim attributed to it may be invented. This is harder to detect than non-citation hallucination because the URL creates an appearance of verifiability that the user does not have time to manually audit.

The system also does not distinguish between citations that came from parametric training knowledge versus claims grounded via live web retrieval within a single response. All citations are nominally from the search results array, but the relationship between the model’s knowledge sources and the final cited URLs is not exposed at the per-claim level.

Citation Source Distribution

Independent research suggests Perplexity prioritizes Reddit at 46.7% of citations in one study. The source distribution affects citation quality: Reddit content quality varies enormously, and the visual treatment of Reddit citations alongside primary sources or peer-reviewed material does not consistently differentiate source authority.



Broad format support.
Plan-based file size and retention.

Perplexity accepts a broad set of file formats across consumer and API surfaces. Files in standard threads are auto-purged after a tier-specific retention period. Files in Spaces persist until explicitly deleted.

Supported Formats

  • Documents: PDF, DOC, DOCX, TXT, RTF, ODT.
  • Spreadsheets: XLSX, CSV (25 MB recommended maximum for best parsing results).
  • Presentations: PPTX.
  • Text and code: Markdown, JSON, HTML.
  • Images: PNG, JPEG (vision analysis).
  • Audio: MP3, WAV, OGG, FLAC (up to 40 MB, transcribed to text).
  • Video: MP4, MOV, AVI, WEBM (up to 40 MB, transcription only).

The audio and video processing is transcription-based rather than full multimodal understanding. The model receives the transcript text rather than processing the raw audio or video stream natively.

Plan-Based File Limits

Plan
Max File Size
Files Per Upload
Retention

Free
40 MB
10
30 days

Pro
50 MB
10
90 days

Enterprise Pro
1 GB
20
1 year

API (Sonar)
50 MB (URL bypass)
1 per request
Developer-managed

Parser fidelity: XLSX and CSV parsing is recommended at 25 MB or less for best extraction accuracy. Larger files may produce degraded extraction. OCR for scanned documents was listed as “coming next” in September 2025 documentation, so scanned PDF fidelity may remain limited as of the research date. For workflows that depend on scanned PDF extraction, test empirically before relying on the parser output.



Two-layer personalization.
OpenAI-compatible developer endpoint.

Memory: Two-Layer Personalization

Memory (also called Personal Search) is a two-layer system. The first layer is Memories, which are explicit preferences, interests, and personal facts extracted from repeated usage patterns. Repetition is the primary signal. The second layer is Search History, which is past queries and responses available for context enrichment.

Sensitive categories (health conditions, financial details) are filtered out regardless of repetition frequency. Users can manage and delete memories in Settings. Memory persists cross-model. Whether the user is querying Sonar, GPT, Claude, or Gemini through Perplexity’s interface, the same memory store is referenced.

The recall rate is reported at 95% following a February 2026 improvement. Available all tiers. Pro and Max can opt out. Enterprise tiers: data is never used for training by default.

Sonar API: Developer Access

OpenAI-compatible at https://api.perplexity.ai/v1/sonar (updated) and https://api.perplexity.ai/chat/completions (legacy). Standard chat completions request structure works with Perplexity-specific fields: search_context_size and citations in the response.

Rate limits tiered by lifetime credit purchase: Tier 0 (no purchase), Tier 1 ($50), Tier 2 ($250), Tier 3 ($500), Tier 4 ($1,000), Tier 5 ($5,000). Higher tiers receive higher RPM ceilings, documented as 20 to 100 RPM across the ladder.

Two integration issues for developers: the 300-second timeout requirement for sonar-deep-research, and the JSON parsing <think> block failure on sonar-reasoning-pro requiring custom parsers.



Every feature, every tier,
at a glance.

Tier availability for several features is not enumerated in official Perplexity documentation. Treat tier-specific limits as Volatile and verify at perplexity.ai before relying on the cap for production planning.

Feature
Free
Pro
Max
Enterprise Pro
Enterprise Max

Sonar models
Auto-selected
Full
Full
Full
All + advanced

Third-party models
No
Yes
Yes
Yes
Yes + advanced

Deep Research
5/day
~500/day
Highest
50/month
500/month

Spaces
No
50 files
50 files
500 files
5,000 files

Pages
Basic
Full
Full
Full
Full

Model Council
No
No
Yes
No
Yes

Labs
No
Yes
Yes
Yes
500/month

Comet browser
Yes
Yes
Yes
Yes
Yes

Comet Assistant
No
No
Yes
80/month
800/month

Shopping (US)
Yes
Yes + Buy with Pro
Yes + Buy with Pro
Yes
Yes

Memory
Yes
Yes (opt-out)
Yes (opt-out)
No training
No training

File uploads
40 MB / 30d
50 MB / 90d
50 MB / 90d
1 GB / 1yr
1 GB / 1yr



Perplexity Features: Frequently Asked Questions

What is Perplexity Deep Research?

Deep Research is Perplexity’s agentic research feature that decomposes a query into sub-queries, performs dozens of web searches, reads hundreds of source documents, and synthesizes a multi-page cited report. Consumer queries take 2 to 4 minutes per execution. The API exposes the same capability through sonar-deep-research with a 300-second timeout requirement that developers must explicitly configure.

What are Perplexity Spaces?

Spaces are workspace containers for related threads, files, and custom AI instructions. Users create named workspaces, attach reference files, and configure custom instructions that apply to all threads within. Files persist until explicitly deleted (unlike standard threads which auto-purge after 7 days). Pro: 50 files per Space. Enterprise Max: 5,000 files per Space.

How does Perplexity Pages work?

Pages is Perplexity’s knowledge-creation feature. Users enter a topic and select an audience level (beginner or expert). Perplexity generates a multi-section article with inline citations and images sourced from current web data. The Page is published on perplexity.ai with a shareable URL. Available on Free (basic) and Pro (full). Documented limitation: Pages cannot be exported to PDF, WordPress, or external CMS as of early 2026.

What is Model Council?

Model Council is Perplexity’s multi-model orchestration feature, launched 2026-02-05 and available exclusively at the Max ($200/month) and Enterprise Max ($325/seat/month) tiers. The feature dispatches a single user query to Claude Opus 4.6, GPT-5.2, and Gemini 3 Pro simultaneously. A chair model produces a synthesis output surfacing agreement, disagreement, and unique insights. Architecturally, this is parallel dispatch with synthesis, distinct from shared-thread orchestration where models read each other’s responses.

What is Perplexity Labs?

Perplexity Labs is a self-supervised work cycle running approximately 10 minutes per task. Labs combines web browsing, code execution, chart and image generation, and asset creation to deliver structured outputs (reports, spreadsheets, dashboards, web apps). Available on Pro, Max, Enterprise Pro, and Enterprise Max. Free tier does not have access. Documented hard limit: 10-minute maximum self-supervised work cycle per task.

What is the Comet browser?

Comet is Perplexity’s AI-native desktop browser for Mac and Windows, built on Chromium with a sidecar AI assistant in every tab. The assistant can answer questions about the current page, summarize content, perform cross-tab tasks, and execute background agentic workflows. Comet launched 2025-07-09 as Max-only and was made free for all users worldwide on 2025-10-01. The Comet Plus add-on at $5/month bundles premium publisher content from CNN, Washington Post, Fortune, LA Times, and Condé Nast.

How accurate are Perplexity’s citations?

Per the Columbia Journalism Review’s 2025-03 audit, Perplexity Sonar Pro recorded 37% citation error rate, the lowest of eight platforms tested. The Pro variant specifically scored 45% per the Suprmind AI Hallucination Rates and Benchmarks reference. Both rates are best-in-class but still mean more than one in three citations may be fabricated or misdirected. The structural failure mode is real URLs with content that may be invented. The URL is genuine. The claim attributed to it may not be.

What file formats does Perplexity accept?

Perplexity accepts documents (PDF, DOC, DOCX, TXT, RTF, ODT), spreadsheets (XLSX, CSV with 25 MB recommended max for best parsing), presentations (PPTX), text and code (Markdown, JSON, HTML), images (PNG, JPEG with vision analysis), audio (MP3, WAV, OGG, FLAC up to 40 MB transcribed), and video (MP4, MOV, AVI, WEBM up to 40 MB transcription only). Audio and video are transcription-based rather than full multimodal understanding.

How does Perplexity Memory work?

Memory is a two-layer system. Memories are explicit preferences, interests, and personal facts extracted from repeated usage patterns (typically three or more mentions). Search History is past queries and responses for context enrichment. Sensitive categories (health, financial details) are filtered out regardless of repetition. The recall rate is reported at 95% following a February 2026 improvement. Memory persists cross-model. Available all tiers. Pro and Max can opt out. Enterprise tiers: data is never used for training by default.

What integration issues should developers know about?

Two documented integration issues. First, sonar-deep-research has a 300-second timeout requirement that developers must explicitly configure since default request timeouts in many client libraries are shorter and queries can fail silently. Second, sonar-reasoning-pro outputs reasoning tokens in a <think> block before the JSON response. The response_format parameter does not strip these tokens. Developers requesting structured JSON output must implement custom parsers that extract content after the <think> block.



Perplexity’s features are deep.
Suprmind orchestrates five model families.

Use Perplexity for citation-grounded research. Pair with Claude for calibration, Gemini for multimodal breadth, GPT for math reasoning, and Grok for contrarian signal. All in one shared conversation, with cross-model fact-checking before any answer reaches your decision.

7-day free trial. All five frontier models. No credit card required.



Disagreement is the feature.

Last verified May 10, 2026. Next refresh due August 10, 2026.