Claude AI: Complete Guide to Models, Features, Pricing, and Benchmarks (2026)
Claude is a family of AI assistants developed by Anthropic, a US AI safety company founded in 2021 by former OpenAI researchers. As of May 2026, the publicly available flagship is Claude Opus 4.7, released April 16, 2026, with a 1 million token input context window, 128,000 token output, native text and image processing, and an Adaptive Reasoning architecture that allocates internal compute dynamically based on problem complexity. The product is distributed via claude.ai, iOS and Android apps, dedicated macOS and Windows desktop apps, the Anthropic API, and managed platforms (Amazon Bedrock, Google Cloud Vertex AI, Microsoft Azure AI Foundry).
The defining claim about Claude in 2026 is calibration over coverage. Claude Opus 4.7 holds the second-highest Omniscience Index of any current model (26, behind only Gemini 3.1 Pro’s 33), achieved through a refusal-when-uncertain architecture rather than maximized answer rates. Per the Suprmind Multi-Model Divergence Index, April 2026 Edition (n=1,324 production turns), Claude’s confidence-contradicted rate drops from 33.9% on all turns to 26.4% on high-stakes turns – a -7.5 point calibration delta no other tested provider matches. Claude slows down measurably when consequences are real; others do not.
This page covers what Claude is, the full active and deprecated model lineup, what each tier costs and which model you actually get on it, the feature set as it stands in May 2026, the benchmark picture (where Claude leads, where it lags, what to read into the gaps between vendor and independent measurements), the hallucination patterns that should shape how you use it, what production multi-model data shows about Claude relative to its peers, the active controversies, and the questions people most often search for. Numbers are dated. The product changes weekly. Where a claim is volatile, it is flagged.
See also: Suprmind Multi-Model Divergence Index →
What Claude Is
Claude is a conversational AI product developed by Anthropic that uses the Claude Opus 4.7 language model as of April 2026 to answer questions, generate text and code, analyze documents, control web browsers and operating systems, and complete multi-step agentic tasks. The product is distinct from the underlying Claude model family that powers it – the same models can be accessed directly through the Anthropic API at platform.claude.com, on Amazon Bedrock, on Google Vertex AI, and on Microsoft Azure AI Foundry at different pricing.
Anthropic was co-founded in 2021 by Dario Amodei (CEO) and Daniela Amodei (President) along with seven other former OpenAI employees. The company is structured as a Delaware Public Benefit Corporation. As of early 2026, annualized revenue reached approximately $14B and a $30B Series G round closed February 11, 2026 at a $380B post-money valuation. A subsequent round at $850-900B+ valuation was reported as actively closing in late April 2026 (TechCrunch, 2026-04-29, not confirmed closed).
Claude vs the Anthropic API
claude.ai is the consumer and prosumer product. The Anthropic API (platform.claude.com, formerly console.anthropic.com) is the developer surface. Both run on Claude models, but the experience and cost structure are different. claude.ai offers Free, Pro, Max 5x, Max 20x, Team Standard, Team Premium, and Enterprise tiers with bundled access to features like Projects, Artifacts, Memory, Computer Use, Skills, MCP, and Microsoft 365 integration. The API exposes raw model endpoints with metered per-token pricing, no chat UI, and developer-controlled feature use.
Claude vs Claude Opus 4.7 – Are They the Same?
No. Claude Opus 4.7 is one underlying model. claude.ai is the product that routes your query to Claude Opus 4.7, Claude Sonnet 4.6, or Claude Haiku 4.5 depending on tier and prompt complexity. Claude Sonnet 4.6 is the default model on Free and Pro plans as of February 2026. Opus 4.7 is available with limits on Pro and without limits on Max, Team, and Enterprise. The model selector dropdown surfaces the tier-available choices, but claude.ai does not show a per-message indicator of which dated snapshot processed a given query – this is a documented user pain point. Developers using API calls receive the pinned snapshot in response metadata.
A separately announced Claude Mythos Preview (2026-04-07) sits above Opus 4.7 in capability but remains invitation-only through Project Glasswing, a cybersecurity research initiative. Mythos posts the highest benchmark scores of any Claude model at the time of writing – SWE-bench Verified 93.9%, GPQA Diamond 94.6%, CyberGym 83.1% – but is not available on claude.ai or the standard API.
See also: Suprmind Multi-Model Divergence Index →
All Claude Models — Current and Deprecated (2026)
Anthropic deploys Claude across three concurrent capability tiers – Opus (highest capability), Sonnet (balanced), and Haiku (fast and economical) – with multiple generations active simultaneously. Architecture remains fully proprietary. Anthropic has not publicly confirmed parameter counts, layer counts, or whether any Claude model uses a Mixture-of-Experts configuration. Multiple third-party sources describe the architecture as a dense transformer.
Below is the active and deprecated picture as of May 2026. Variants and dates are taken from Anthropic’s official model catalog at platform.claude.com/docs/en/about-claude/models and confirmed against independent tracking. This table changes frequently – check the source URL for the current list.
Active Claude Models (May 2026)
Source: platform.claude.com – last verified 2026-05-07
- Released 2026-04-16
- 1M token context, 128K output
- Multimodal in: text, image (vision to 2,576px)
- API: $5.00 / $25.00 per 1M tokens; cached read $0.50
- Released 2026-02-17
- 1M token context, 128K output (300K via Batch)
- API: $3.00 / $15.00 per 1M tokens
- Default model for Free and Pro claude.ai users
- Released 2025-10-15
- 200K context / 64K output
- API: $1.00 / $5.00 per 1M tokens
- Near-frontier coding at small-tier price (SWE-bench 73.3%)
- Released 2026-02-05
- 1M context (the generation that introduced 1M at standard pricing)
- API: $5.00 / $25.00 per 1M tokens
- 67% price reduction from Opus 4.1’s $15/$75
- Announced 2026-04-07
- Invitation-only (Project Glasswing)
- SWE-bench Verified 93.9%, GPQA Diamond 94.6%, CyberGym 83.1%
- Internal codename: “Capybara” (per March 2026 source leak)
- Claude 3 Opus, Sonnet, Haiku: legacy on pricing page
- Claude 3.5 Sonnet (v1, v2), 3.5 Haiku: supported/legacy
- Claude 3.7 Sonnet (2025-02-24): introduced Extended Thinking
- Claude 1, 2, 2.1, Instant 1.2: fully deprecated
Claude 4 Generation: Opus 4.7, Opus 4.6, Sonnet 4.6, Haiku 4.5
Claude Opus 4.7 (2026-04-16) is the current flagship. It introduced the xhigh effort level for Adaptive Reasoning (between high and max), raised the Computer Use vision input ceiling to 2,576 pixels on the long edge (from approximately 850 pixels prior), and deployed a new tokenizer where the same input maps to 1.0-1.35x more tokens depending on content type. SWE-bench Verified 87.6%, SWE-bench Pro 64.3% (current industry high), GPQA Diamond 94.2%, MCP-Atlas 77.3%, OSWorld 78%. Reliable knowledge cutoff: January 2026. Manual Extended Thinking via budget_tokens is deprecated for Opus 4.7 and later; attempting it returns a 400 error. Pricing $5/$25 per million input/output tokens, unchanged from Opus 4.6.
Claude Opus 4.6 (2026-02-05) is the generation that first delivered a 1 million token context window at standard pricing – eliminating the long-context surcharge that had existed across the AI industry. The Opus 4.6 launch also dropped the Opus tier price 67% (from Opus 4.1’s $15/$75 to $5/$25 per million tokens), the largest single-generation Opus price reduction recorded. Claude Opus 4.6 became the first AI model to hold #1 across all three LMArena arenas (Text 1503-1504, Code 1560, Search 1255) on February 26, 2026.
Claude Sonnet 4.6 (2026-02-17) became the default model for Free and Pro claude.ai users at launch. 1M context (initially beta, generally available March 2026), $3/$15 pricing, 128K output (300K via Batch with the output-300k-2026-03-24 beta header). On the harder Vectara new dataset, Sonnet 4.6 scored 10.6% hallucination – below GPT-5.2-high’s 10.8% on the same benchmark. AA-Omniscience hallucination approximately 38% (less than half GPT-5.2’s ~78%). Reliable knowledge cutoff: August 2025; training data cutoff: January 2026.
Claude Haiku 4.5 (2025-10-15) is Anthropic’s current small/fast model with near-frontier coding performance. 200K context, 64K output, $1/$5 pricing. SWE-bench 73.3% with extended thinking (averaged over 50 trials), AA-Omniscience hallucination 25% – the best Haiku-tier hallucination result in the cohort. Released under ASL-2 safety classification (Sonnet 4.5 and Opus 4.1 are ASL-3).
Claude 3.x and Earlier (Historical Context)
Claude 3.7 Sonnet (2025-02-24) was the first Claude model with hybrid reasoning – capable of near-instant responses or visible step-by-step Extended Thinking with a developer-controlled budget_tokens parameter. It scored 4.4% on the Vectara old summarization benchmark (factual consistency 95.6%) and 70.3% on SWE-bench Verified with Extended Thinking. The 3.5 Sonnet (v1, v2) and 3.5 Haiku models remain active per platform docs as of 2026-05-07, flagged as supported/legacy. Claude 3 Opus, Sonnet, and Haiku are listed as legacy on Anthropic’s pricing page. Claude 1, 2, 2.1, and Instant 1.2 are fully deprecated. Claude Opus 4.1 has an AWS Bedrock end-of-life date of 2026-05-31.
What Model Am I Using? Tier-to-Model Mapping
This is the single most-asked question in Claude documentation, and Anthropic’s UI does not surface a per-message indicator of which exact model snapshot processed a given query. As of May 2026:
The model selector dropdown shows the available choice. The system prompt is technically accessible via probing (the Claude Opus 4.6 system prompt was extracted and published to GitHub on 2026-02-05). The persistent UI does not surface the dated snapshot. Default-model transitions (such as the Sonnet 4.5 to Sonnet 4.6 switch in February 2026) are announced via Anthropic newsroom but not via in-product notification for existing users.
See also: Claude pricing details →
Claude Features: What Each One Does
Anthropic ships features across a coherent claude.ai web interface, native iOS and Android apps, macOS and Windows desktop apps, and developer-facing surfaces (Anthropic API, Claude Code CLI, MCP). The platform reached major feature parity by April 2026 across all paid tiers, with feature gates focused on usage volume rather than feature exclusivity.
Adaptive Reasoning vs Extended Thinking
Extended Thinking, introduced with Claude 3.7 Sonnet (2025-02-24), forces Claude to generate a visible chain-of-thought trace before answering. The developer sets a budget_tokens parameter to control reasoning compute. Adaptive Reasoning (also called Adaptive Thinking), introduced with the 4.6 generation in February 2026, replaces this paradigm. Claude evaluates problem complexity internally and allocates reasoning compute dynamically. The developer specifies an effort level (standard, high, xhigh, max) rather than a token budget. At high effort, Claude almost always thinks before responding. At lower effort levels, Claude may skip thinking for simple problems. The xhigh level introduced with Opus 4.7 sits between high and max and provides additional compute for hard tasks without committing to maximum spend. Adaptive Reasoning automatically enables Interleaved Thinking – reasoning between tool calls – which makes it structurally better suited for agentic workflows than the prior paradigm. Manual Extended Thinking via budget_tokens is deprecated for Opus 4.7 and later; attempting it returns a 400 error.
Projects and Artifacts
Projects create isolated workspaces where users upload reference documents and system instructions that persist across conversations. Claude performs retrieval-based reasoning over project content – relevant sections are pulled into active context rather than loading the entire project at once. Project content is cached and does not count against per-message usage limits. Per-chat file upload caps at 20 files maximum, 30 MB each, regardless of tier. Enterprise plan chat context expands to 500K tokens; all other plans use 200K tokens in chat (1M tokens on API for Opus and Sonnet 4.6+). Projects launched September 2024 and expanded context 10x in June 2025.
Artifacts is Claude’s output format for code, documents, diagrams, and interactive content that can be rendered, edited, and exported directly from the conversation interface. When Claude generates substantial standalone content – code, HTML, SVG, Mermaid diagrams, React components, formatted Markdown – a side panel opens with a live preview. Users can iterate on artifacts, share them publicly, or (on Team and Enterprise) share within organizational boundaries. Artifacts launched in preview June 2024 and reached general availability across all tiers on August 26, 2024. As of April 2026, Artifacts ships on all paid plans and inside Projects.
Claude Code
Claude Code is Anthropic’s terminal-first agentic coding tool, generally available since 2025-05-22. It runs Claude as an autonomous coding agent that searches code, edits files, runs tests, and commits to GitHub. Native integrations include VS Code and JetBrains extensions (edits appear inline in files), GitHub PR tagging, and a Claude Code SDK for building custom agents. Claude Opus 4.7 raised the default effort level to xhigh for all plans at launch and introduced Task Budgets (public beta) for guiding token spend across longer agentic runs. The April 2026 launch also introduced the /ultrareview command for dedicated review sessions and a multi-session sidebar.
The Pro tier ($20/month) inclusion of Claude Code is volatile and contested as of 2026-05-07. The current anthropic.com/pricing page lists “Includes Claude Code” under Pro; an independent changelog tracker (scriptbyai.com, April 2026) states Anthropic removed Claude Code from Pro in April 2026. Conflict unresolved – verify directly at anthropic.com/pricing. Max plans include Claude Code, Enterprise includes Claude Code, and API access via the Claude Code SDK is uniformly available.
See also: Claude Code features and pricing →
Computer Use
Computer Use was originally released as beta with Claude 3.5 Sonnet on 2024-10-22, expanded across Claude 3.7 and Claude 4 generations, and reached general availability on claude.ai in March 2026. Developers provide Claude with computer use tools and a user prompt via the Messages API. Claude assesses the task and constructs tool use requests; the developer runs actions in a sandboxed virtual machine with X11/Xvfb display, lightweight desktop environment, and pre-installed applications. Default loop iteration cap is 10 (developer-adjustable). Claude Opus 4.7 significantly improved Computer Use reliability via high-resolution image support, achieving 98.5% on XBOW’s visual-acuity benchmark vs 54.5% for Opus 4.6, and 78% on OSWorld – tied with GPT-5.5 at 78.7%.
See also: Computer Use feature details →
Memory and Cowork
Memory operates in two modes. Chat memory derives summaries of past conversations and carries them across sessions, viewable and editable at Settings → Capabilities → Memory. File-system memory for agentic use writes to a /memory folder, read at session start, with optional auto-memory mode that lets Claude decide what to store. Opus 4.7 specifically improved file-system memory reliability for long multi-session agentic work. Chat memory shipped to Team and Enterprise plans in September 2025 and to Free in March 2026. The August 2025 data policy change extended conversation data retention to 5 years for users not opted out of training; this is distinct from active memory retention.
Claude Cowork launched in research preview January 2026 and reached general availability across all paid plans in April 2026. Cowork grants Claude access to a user-specified folder on the local computer; Claude can read, edit, and create files autonomously, supporting multi-step task execution and sub-agent coordination for parallelizable work. Initial launch was macOS-only.
MCP and Integrations
MCP (Model Context Protocol) is an open standard Anthropic designed to allow Claude to connect to external tools, data sources, and services via a standardized interface. Third-party MCP servers exist for Notion, Zapier, GitHub, and major IDE tools. Claude Opus 4.7 scores 77.3% on MCP-Atlas, leading GPT-5.4 by 9.2 points and Gemini 3.1 Pro (73.9%) by 3.4 points, indicating strong real-world tool-orchestration performance.
Claude in Excel launched as a beta research preview in October 2025, providing workbook understanding with cell-level citations for explanations and the ability to update assumptions while preserving formulas. Claude for Word launched in April 2026 (Pro and Max). Claude for Microsoft 365 (Outlook, broader 365 surfaces) is included on Pro, Max, Team, and Enterprise. Free tier does not include Microsoft 365 integration.
See also: Custom GPTs deep guide →
Claude Benchmarks and Accuracy
Benchmarks tell different stories depending on what they measure. Claude leads on autonomous multi-file coding (SWE-bench Pro), agentic tool use (MCP-Atlas), tool-enabled HLE, and calibration metrics. It trails on raw knowledge breadth (AA-Omniscience accuracy), multimodal coverage (no audio or video input), and ARC-AGI-2. Both directions are real signals of different qualities.
Benchmark Scores – Current Flagships
Sources: Vellum AI, 2026-04-15; Suprmind Hallucination Rates, 2026-04-26; pricepertoken.com; DataCamp, 2026-04-26; ofox.ai. Last verified 2026-05-07.
A note on methodology: AIME 2025 has effectively saturated at the frontier (multiple models score >99%) and is no longer differentiating; treat AIME advantages with skepticism. The harder Vectara new-dataset reports reasoning models exceed 10% hallucination because they “overthink” summarization, deviating from source material – so raw Vectara comparisons across reasoning and non-reasoning models are misleading without context. CursorBench is operated by Cursor, a significant Claude distribution partner; no independent replication has been found. The Claude Opus 4.7 MRCR v2 regression to 32.2% on 1M context (down from Opus 4.6’s 78.3%) is attributed by Anthropic to intentional error-reporting behavior when information is missing rather than fabricating answers; independent verification of the mechanism is thin.
Claude Hallucination Rates
Claude’s hallucination profile is the central differentiator from peer models. According to Suprmind’s AI Hallucination Rates and Benchmarks reference (May 2026 update), Claude 4.1 Opus achieves a 0% AA-Omniscience hallucination rate by mathematically declining uncertain queries – the lowest of any model tested at any scale. Claude Opus 4.7 holds AA-Omniscience hallucination at 36% (Index 26, second-highest overall behind Gemini 3.1 Pro’s 33), 50 percentage points lower than GPT-5.5’s 86% on the same benchmark. Claude Opus 4.5 with web search scored 30% on HalluHard – the lowest of any model on the realistic-conversation hallucination benchmark.
The Claude pattern is calibration-by-refusal: Claude declines to answer more often than peers and hallucinates less when it does answer. This produces both the lowest hallucination rates and lower raw accuracy (~47% AA-Omniscience accuracy vs Gemini 3.1 Pro’s 55.3%). Reasoning models including the 4.5 and 4.6 generations exceed 10% on Vectara’s harder summarization dataset due to documented “overthinking” – reasoning that deviates from source material. This is not a capability claim about Claude’s correctness; it is a consistency claim about Claude’s calibration.
What Makes Claude Different — The Calibration Advantage
Academic benchmarks rank Claude Opus 4.7 in a three-way tie at the frontier (AA Intelligence Index 57). Production multi-model data tells a more specific story, and that story is the most useful one for picking AI tools for actual work.
Per the Suprmind Multi-Model Divergence Index (April 2026 Edition, n=1,324 production turns), Claude’s confidence-contradicted rate drops from 33.9% on all turns to 26.4% on high-stakes turns – a -7.5 point calibration delta. No other provider tested shows a delta steeper than -3.4 points (ChatGPT/GPT). This is the single most defensible empirical distinction for Claude in a multi-model context. Claude slows down measurably when consequences are real; others do not.
How Claude Performs in Multi-Model Contexts
Catch ratio measures corrections made divided by times caught. A ratio above 1.0 means a model corrects others more than it gets corrected. Per the Suprmind Multi-Model Divergence Index, the April 2026 edition spread was: Perplexity 2.54, Claude 2.25, Grok 0.72, ChatGPT 0.38, Gemini 0.26. Claude made 304 corrections and was caught 135 times – the second-highest catch ratio of five providers. Combined with Perplexity (catch ratio 2.54), the two providers account for 60.7% of all corrections in the study. This positions Claude as a verification-layer model rather than a sole oracle.
Unique insights followed the same pattern. Claude generated 631 unique insights (24.5% share, second only to Perplexity’s 636/24.7%) with 268 rated critical-severity (severity ≥7 on a 10-point scale). For reference, ChatGPT contributed 339 (13.2% share, 85 critical), making Claude approximately 3.15x more productive on critical-severity unique insights than ChatGPT in the same dataset. Claude is the second-best engine for novel insight generation in a multi-model ensemble.
See also: AI catch ratio data →
Where Claude Has Limitations
Three documented limitations shape when Claude alone is the wrong tool.
First, broad knowledge retrieval. Claude Opus 4.7’s AA-Omniscience accuracy of approximately 47% trails Gemini 3.1 Pro’s 55.3% by an 8-point gap. This is the direct cost of refusal-by-design – Claude answers fewer questions correctly in total though more correctly as a proportion of what it does answer. Users who need maximum breadth over maximum precision should pair Claude with a higher-coverage model.
Second, multimodal inputs. Claude accepts only text and image. Audio and video inputs are not supported. Gemini 3 Pro’s FACTS multi-dimensional factuality score of 68.8 versus Claude Opus 4.5’s 51.3 (a 17-point deficit) is partly structural – FACTS measures ingestion across modalities Claude cannot read.
Third, self-consistency in iterative research. Per the Suprmind Multi-Model Divergence Index (April 2026), Claude vs Claude is the top combative pair in the ResearchAnalysis domain – 10 contradictions across 74 turns, a 13.5% intra-model contradiction rate. The Claude-vs-Claude pattern is the single most important orchestration signal for users deploying Claude on iterative research workflows. Cross-checking against itself or peers reduces the volatility.
See also: Claude vs ChatGPT vs Gemini comparison →
Claude Pricing — Free, Pro, Max, Team, Enterprise
Anthropic operates a seven-tier consumer and business pricing structure. Two volatile elements are documented as of May 2026: the inclusion status of Claude Code in Pro (anthropic.com/pricing lists it; an independent changelog states it was removed in April 2026), and the message-volume caps per tier (described as “usage limits apply” or a “conversation budget” without specific counts).
Subscription Tier Comparison
Source: anthropic.com/pricing, accessed 2026-05-07.
See also: Claude pricing details →
API Pricing for Developers and Enterprise
API pricing for the current generation models is metered per million tokens with separate input, cached input write, cached input read, and output rates.
Source: anthropic.com/pricing, accessed 2026-05-07.
Additional API-level charges: Managed Agents at $0.08 per session-hour active runtime; Web Search at $10 per 1,000 searches; Code Execution free for the first 50 hours per day per organization, then $0.05 per hour per container; US-only inference at 1.1x input and output pricing; prompt caching with 5-minute default TTL (extended TTL available). Batch API: 50% discount on all models, supporting up to 10,000 queries for async processing in under 24 hours.
Recent Pricing Changes (2025-2026)
The most significant pricing event in Claude’s API history was the 67% Opus price reduction at Opus 4.6 launch (2026-02-05): from $15/$75 per million tokens (Opus 4.1) to $5/$25 per million tokens (Opus 4.6 onward). The 1M token context window also became standard at no surcharge starting with Opus 4.6 and Sonnet 4.6. Claude Opus 4.7 maintained the new $5/$25 pricing. Claude Opus 4.1 has an AWS Bedrock end-of-life date of 2026-05-31, retiring the prior $15/$75 Opus tier from the active product line.
Claude Controversies and Known Issues
Anthropic faced more frequent regulatory and engineering controversies in early 2026 than any other AI lab, driven by safety-first commitments creating direct conflicts with high-profile customers and by performance regressions in Claude Code becoming community focal points.
The Pentagon Refusal and Department of War Lawsuit (February-March 2026)
On 2026-02-26, Anthropic publicly refused a Department of Defense contract clause that would have permitted “any lawful use” of Claude including fully autonomous weapons targeting and domestic surveillance of Americans without judicial oversight. CEO Dario Amodei stated the company “cannot in good conscience accede.” The Pentagon designated Anthropic a “supply-chain risk to national security” – the first such designation ever applied to an American company. President Trump issued an executive order on 2026-02-27/28 banning U.S. government use of Claude. The Department of War deployed Claude against Iran less than 24 hours after the ban. Anthropic filed suit on 2026-03-09 alleging government retaliation. The lawsuit was active as of research date.
The architectural cause is significant: Claude’s January 2026 Constitutional AI framework contains explicit hard constraints against facilitating mass surveillance and autonomous lethal targeting without human oversight. These are model-level, not purely policy-level constraints, which means they cannot be overridden via system prompt configuration.
Claude Code Performance Regression (March-April 2026)
A widely covered “Claude got dumber” narrative emerged between March 4 and April 13, 2026. AMD Senior Director of AI Stella Laurenzo published forensic analysis of 6,852 Claude Code sessions (234,760 tool calls, 17,871 thinking blocks) showing a shift from research-first to edit-first behavior, rising stop-hook violations, and reduced reasoning depth. Anthropic published a full engineering postmortem on 2026-04-23 confirming three separate causes: (1) default reasoning effort changed from high to medium on 2026-03-04 (reverted 2026-04-07); (2) cache optimization bug clearing thinking history on every turn for stale sessions from 2026-03-26 (fixed 2026-04-10); (3) system prompt verbosity constraint on 2026-04-16 causing 3% eval drop (reverted 2026-04-20).
The “intentional degradation” accusation was unsubstantiated. All three causes were engineering decisions with legitimate rationales that had unforeseen interactions. Separately, a viral BridgeMind benchmark claiming a 15-point performance drop was based on n=6 tasks; an independent retest with n=30 showed negligible movement (87.6% to 85.4%). The real governance concern is the 6+ week delay between first change and public postmortem.
Data Policy and Training Opt-Out (August 2025)
On 2025-08-28, Anthropic reversed its prior policy of not training on consumer conversations. Free, Pro, and Max plan users’ conversations and coding sessions became training data by default. Data retention extended from 30 days to 5 years unless users manually opted out by 2025-09-28; full enforcement began October 2025. Lawfare Media noted this represents a shift from explicit consent to legitimate interest under GDPR, raising compliance questions for European users. Enterprise and Team plans include contract-level data non-training provisions without per-user opt-out.
Constitutional AI and Refusal Patterns
Anthropic published a new Claude Constitution on 2026-01-22 (approximately 84 pages, Creative Commons public domain), replacing the 2023 Constitutional AI approach. The framework shifts from rule-based prescriptions to reason-based alignment that explains why certain behaviors matter, aiming for generalization to novel situations. It establishes a 4-tier priority hierarchy: safety > ethics > guidelines > helpfulness. It formally acknowledges the possibility of Claude’s consciousness and moral status – the first such acknowledgment from a major AI lab. The Oxford AI Ethics blog noted this represents “two evaluative continua” rather than a fixed ruleset. Hard constraints include refusing to assist with autonomous lethal targeting without human oversight, mass surveillance without judicial oversight, CBRN weapons development, and content that would seize illegitimate societal control.
See also: ChatGPT hallucination by version →
Claude in Enterprise — Adoption and Integrations
Claude’s enterprise penetration is the deepest of any frontier AI model family by deployment count, driven by Constitutional AI safety architecture meeting enterprise procurement requirements that pure-capability competitors fail.
Enterprise Use Cases and Deployments
70% of Fortune 100 companies are Claude customers; 8 of the Fortune 10; over 500 customers spend more than $1M annually. Enterprise customers (300,000+ businesses) account for approximately 80% of Anthropic’s revenue. Customers spending over $100K annually grew 7x in the past year. Claude’s share of enterprise LLM spend reached approximately 40% by 2025, up from 12% two years prior. Annualized revenue grew approximately 10x in each of the past three years to $14B by early 2026.
Notable deployments include Deloitte (470,000 employees globally on Claude), Cognizant (350,000 associates on Claude Code, broader Claude across functions), Thomson Reuters CoCounsel for legal research and document drafting (1M+ users), Lyft (customer support automation reducing support time over 87% with decision accuracy improved 30%), TELUS (tens of thousands of users, billions of tokens monthly), and Zapier (workflow automation at scale).
Platform Integrations (Bedrock, Vertex, GitHub Copilot, Cursor)
The developer ecosystem includes 6,000+ apps with native Claude integration and 75+ enterprise workflow connectors. Notable integrations: Microsoft 365 (Excel, Word, Outlook), GitHub Copilot (Claude Sonnet 4 was the underlying model at launch), Cursor (CursorBench partnership), Slack, Notion (Notion Skills for Claude), Amazon Bedrock (all active models), Google Vertex AI (all active models), and Microsoft Azure AI Foundry (generally available for select models with EU inference “Coming 2026”). Heavy industry concentration in Legal (Thomson Reuters CoCounsel), Financial Services (Finance Agent benchmark lead), Professional Services (Deloitte, Cognizant), Software Engineering (GitHub Copilot, Cursor, IDE integrations), Telecom (TELUS), and Customer Support (Lyft 87% time reduction).
Hardware and OS integrations: macOS desktop app (Cowork was macOS-only at January 2026 launch), Windows desktop app, iOS app, Android app, GitHub Copilot, Cursor, and a SpaceX compute partnership disclosed mid-2025 (terms not publicly confirmed).
See also: Claude vs ChatGPT comparison →
Sources
Authoritative sources consulted in compiling this guide. For maintenance, monitor the URLs noted in the JSON SSOT section.
- Anthropic – anthropic.com (announcements, pricing, business pages)
- Anthropic Help Center – support.claude.com (feature documentation)
- Anthropic Platform – platform.claude.com (API docs, model catalog, deprecations)
- Anthropic Status – status.claude.com (incidents)
- Suprmind Multi-Model Divergence Index – suprmind.ai/hub/multi-model-ai-divergence-index/ (production multi-model data)
- Suprmind AI Hallucination Rates and Benchmarks – suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/ (canonical hallucination data)
- Artificial Analysis – artificialanalysis.ai (AA Intelligence Index, AA-Omniscience)
- LMArena – arena.ai/leaderboard (user preference rankings)
- Vellum AI – vellum.ai/blog (Claude Opus 4.7 benchmarks)
- DataCamp – datacamp.com (Claude vs Gemini coverage)
- Reuters – reuters.com (DoW lawsuit coverage)
- TechCrunch – techcrunch.com (Series H reporting, August 2025 data policy)
- The Register – theregister.com (Claude Code regression coverage)
- Bloomberg – bloomberg.com (Series G $30B coverage)
- AP News, CNBC – Amazon $25B/$33B investment coverage
- Lawfare Media – lawfaremedia.org (Constitutional AI critiques)
- BISI, Oxford AI Ethics – Constitution evaluations
Last verified 2026-05-07.
FAQ
Frequently Asked Questions
What is Claude AI?
Claude is a family of AI assistants developed by Anthropic, a US safety-focused AI company founded in 2021 by former OpenAI researchers. The current flagship is Claude Opus 4.7, released April 16, 2026, with a 1M token context window and a 64.3% SWE-bench Pro score – the current industry high for autonomous coding. Claude is available via claude.ai, iOS, Android, desktop apps, the Anthropic API, Amazon Bedrock, and Google Vertex AI.
Who made Claude?
Anthropic made Claude. Anthropic was co-founded in 2021 by Dario Amodei (CEO) and Daniela Amodei (President) along with seven other former OpenAI employees. As of early 2026, annualized revenue is approximately $14B and a $30B Series G round closed February 2026 at a $380B post-money valuation.
What is the latest version of Claude?
As of May 2026, the publicly available flagship is Claude Opus 4.7 (released 2026-04-16), featuring a 1M token input context window, 128K token output, Adaptive Reasoning, and improved Computer Use. A separately announced Claude Mythos Preview (2026-04-07) sits above Opus 4.7 but remains invitation-only through Project Glasswing.
Is Claude free to use?
Yes, but with limits. The Free tier provides access to Claude Sonnet 4.6 (default) and limited Haiku at unspecified usage caps described as “conversation budget.” Claude Code, Research mode, and full Opus access require paid tiers.
Does Claude hallucinate?
Yes, but at significantly lower rates than peer models. Claude 4.1 Opus achieves a 0% AA-Omniscience hallucination rate by declining to answer when uncertain – the lowest of any model tested. Claude Opus 4.7 holds AA-Omniscience hallucination at 36%, 50 points lower than GPT-5.5’s 86% on the same benchmark, with an Omniscience Index of 26 (second-highest overall).
Is Claude better than ChatGPT?
Depends on the task. Claude leads on autonomous multi-file coding (SWE-bench Pro 64.3% vs GPT-5.4’s 57.7%), hallucination calibration (AA-Omniscience 36% vs GPT-5.5’s 86%), long-context analysis, and professional-document synthesis. ChatGPT leads on image generation (Claude has none), plugin ecosystem breadth, voice mode, and raw speed on simple queries. Per the Suprmind Multi-Model Divergence Index (April 2026, n=1,324), Claude’s high-stakes confidence-contradiction rate of 26.4% is 9.8 points lower than ChatGPT’s 36.2%.
Why does Claude refuse some requests?
Claude’s Constitutional AI framework establishes hard constraints: no assistance with autonomous lethal targeting without human oversight, no mass surveillance without judicial oversight, no CBRN weapons development, no assistance with seizing illegitimate societal control. These are model-level, not policy-level, constraints. Default refusals also cover explicit sexual content and detailed instructions for illegal activity; operators can configure these defaults within Anthropic’s usage policy.
Why does Claude get worse at coding sometimes?
Three separate engineering changes degraded Claude Code performance between early March and mid-April 2026, all confirmed in Anthropic’s 2026-04-23 postmortem: default reasoning effort reduced from high to medium (reverted 2026-04-07); cache optimization bug clearing thinking history (fixed 2026-04-10); system prompt verbosity constraint causing 3% eval drop (reverted 2026-04-20). The “intentional degradation” accusation was unsubstantiated.
What does “model overloaded” mean in Claude?
The Claude-specific 529 error code means Anthropic’s servers are at capacity, distinct from the generic 503. The largest documented incident was a 14-hour outage on March 2-3, 2026 affecting claude.ai and the mobile apps; the API remained largely functional. Workaround is exponential backoff starting at 1-2 seconds.
Does Claude have open weights?
No. No Claude model has open weights. Anthropic does not publish model weights or allow self-hosted deployment. API and managed platform (AWS Bedrock, Google Vertex AI, Microsoft Azure AI Foundry) are the only access paths.
Stop guessing. Start cross-checking.
Suprmind runs your prompt across ChatGPT, Claude, Gemini, Grok, and Perplexity in parallel. See where they agree, where they disagree, and which insights only one model surfaced — before you act.