Anthropic
Anthropic

Claude Opus 4.7

2026-04-16

Claude Opus 4.7 is Anthropic's flagship frontier model, released April 16, 2026 as a successor to Opus 4.6. It is Anthropic's strongest coding and agentic model to date, purpose-built for long-horizon autonomous work with minimal supervision. Opus 4.7 plans deliberately, writes verification steps (tests, runs, fixes) before declaring tasks complete, and uses file-system memory to carry context across sessions. New controls include task budgets and an xhigh effort level between high and max. At launch it topped SWE-bench Verified at 87.6% and SWE-bench Pro at 64.3%, ahead of GPT-5.4 and Gemini 3.1 Pro, with 69.4% on Terminal-bench 2.0 and ~77% on ARC-AGI-2. Vision capabilities tripled, enabling 98.5% computer-use visual acuity. It features a 1M-token context window and 128K output capacity.

VisionReasoningWeb SearchFile|Proprietary Model
Knowledge Cutoff
2026-01-01
Input → Output Format
Context Memory
1MIN128KOUT
Cost/1M Words
$5IN$25OUT
Calculate Cost

AI Performance Evaluation

Arena Overall Score
1503
±8
As of 2026-04-23
Overall Rank
🥇 No.1
5,321 Votes
Arena by Ability
Hard Prompts
1527±10🥉 No.3
Expert Knowledge
1551±25🥇 No.1
Instruction Following
1502±14🥈 No.2
Conversation Memory
1527±21🥇 No.1
Creative
1499±20🥇 No.1
Coding
1572±17🥇 No.1
Math
1496±33No.5
Arena by Occupation
Creative Writing
1497±17🥇 No.1
Social Sciences
1521±19🥇 No.1
Media
1473±18No.5
Business
1489±18No.6
Healthcare
1517±32No.5
Legal
1502±29No.7
Software
1555±13🥇 No.1
Mathematics
1511±33No.4
Overall
AA Intelligence Index
57%↑19%
LiveBench
68%↑8%
Reasoning & Math
GPQA Diamond
91%↑10%
HLE
40%↑23%
LB Reasoning
75%↑15%
LB Math
76%↑3%
LB Data
76%↑26%
Coding
AA Coding Index
53%↑18%
LB Coding
76%↑3%
LB Agentic
60%↑17%
TAU2
89%↑15%
TerminalBench
52%↑20%
SciCode
55%↑14%
Language & Instructions
IFBench
59%↑2%
AA-LCR
70%↑9%
Hallucination (HHEM)
12%↑2%
Factual (HHEM)
88%↓2%
LB Language
75%↑3%
LB IF
46%↑0%
Output Speed
Standard Mode
56tok/s↓26
First Output 1.36s
Reasoning Mode
65tok/s↓23
First Output 21.50s