Have you ever wondered how AI updates can change your daily work for the better? We invited our community to see the latest improvements we rolled out this year.
We shared tools that boost productivity and sharpen decision-making. Our team worked hard to keep performance high and safety central. Users told us what mattered, and we acted on that feedback.
As we moved forward, we focused on transparency and real-world utility. We believe these upgrades help professionals streamline tasks and gain clearer insights.
Join us as we push boundaries and build a collaborative space where innovation meets practical value for every user.
Key Takeaways
- We invited our community to explore major platform improvements.
- Updates prioritized safety, transparency, and high performance.
- Feedback from users guided meaningful changes.
- Enhancements are designed to boost everyday professional workflows.
- We aim to foster collaboration and practical innovation.
Introducing the Latest Updates to Claude
On April 16, 2026, we released a major upgrade that advances how our model handles complex work.
Claude Opus 4.7 replaces Opus 4.6 and is now generally available to customers. This release improves core capabilities for engineering teams and organizations that require robust analysis and generation at scale.
We prioritized safety and reliability across integration and agent usage. That focus helps reduce risk in sensitive cases while increasing support for complex task flows.
- Stronger performance per token for demanding enterprise workloads.
- Better agent coordination for multi-step generation and task handoffs.
- Expanded accuracy in code and data analysis compared to prior models.
| Feature | Opus 4.6 | Opus 4.7 |
|---|---|---|
| Availability | General beta and staged rollout | Generally available to customers |
| Engineering & analysis | Strong | Improved for complex cases |
| Agent integration | Baseline support | Enhanced coordination and task support |
| Safety & reliability | High | Frontier-grade safeguards |
| Performance per token | Optimized | Higher throughput and quality |
Understanding What Is New With Claude
We explain how the latest model leap changes reasoning and task flow for teams.
Our updated model raises reasoning speed and efficiency across common engineering and analysis tasks.
These models are now generally available, so organizations and customers can adopt advanced capabilities at scale.
We tightened safety protocols to keep frontier usage trustworthy across many cases. That refined safety reduces risk while preserving high throughput for developers.
Integration and agent support were improved to handle complex generation tasks and multi-step workflows. Every token processed is optimized to deliver better results per call.
- Faster reasoning on the model layer for engineering teams.
- Stronger integration for agent orchestration and task handoffs.
- Practical support for analysis, code, and enterprise use cases.
To explore tools that pair well with these capabilities, see our guide to best AI productivity tools.
Advancements in Advanced Software Engineering
This release pushed coding autonomy forward so agents can manage larger portions of an engineering pipeline.
Coding Autonomy
We studied coding autonomy in focused research to let a model handle multi-step code work reliably.
Claude Opus steps up from Opus 4.6 by taking on routine refactors, test generation, and repeatable builds.
Complex Workflow Handling
Our agents now orchestrate workflows across tools and CI systems. That reduces manual handoffs for developers.
The improvements boost reasoning during generation and analysis, and they keep safety and oversight in place for sensitive cases.
| Area | Opus 4.6 | Claude Opus 4.7 |
|---|---|---|
| Coding autonomy | Assisted, human-led | Managed, agent-driven |
| Workflow orchestration | Manual handoffs | End-to-end agent support |
| Safety & oversight | High | Frontier-grade safeguards |
| Performance | Optimized per token | Higher throughput and quality |
- Developers gain a reliable tool for complex tasks.
- Organizations scale engineering while retaining safety.
Enhanced Multimodal Capabilities and Vision
Our vision upgrades let agents interpret technical visuals with far greater accuracy.
Claude Opus 4.7 raises visual fidelity by three times, supporting images up to 2,576 pixels on the long edge. These gains improve how the model extracts data from diagrams, screenshots, and microscopy images.
High Resolution Visual Processing
Higher resolution boosts visual acuity benchmarks. Models now deliver tighter analysis and more reliable information extraction for engineering and life sciences use cases.
These capabilities are generally available to customers and organizations that need pixel-level references. We kept safety and frontier-grade oversight in place so usage in sensitive cases stays protected.
- 3x higher resolution lets agents handle complex diagrams and screenshots.
- Improved integration for workflows that mix text and image data.
- Better support for generation tasks that require pixel-perfect accuracy.
| Feature | opus 4.6 | claude opus 4.7 |
|---|---|---|
| Max image size (long edge) | 860 px | 2,576 px |
| Visual acuity | Baseline | 3x improved |
| Agent image support | Limited | Full pixel-level assistance |
| Typical use cases | UI screenshots, simple diagrams | Technical diagrams, microscopy, high-detail engineering |
Strengthening Cybersecurity Safeguards
We focused on cybersecurity controls to keep powerful analysis tools in safe hands.
In Claude Opus 4.7 we implemented robust safeguards that automatically detect and block requests that suggest prohibited or high-risk activity. These protections run at the edge of the model pipeline to stop harmful prompts before they reach generation stages.
Our safety measures came from ongoing research into the risks and benefits of frontier models. By limiting how some capabilities are released, we balance powerful support for legitimate vulnerability research against the risk of misuse.
- Automated blocking: prompt injection and risky patterns are detected early.
- Tiered access: advanced capabilities are gated to reduce abuse.
- Continuous monitoring: we learn from real-world deployments and improve rules.
These changes help organizations and customers use models for engineering and security analysis with greater peace of mind. For teams building secure interfaces and tools, see our secure tool design guide.
Performance Benchmarks and Real World Results
Measured gains in production tests confirm improved throughput and accuracy.
We ran focused evaluations to measure how the model performs on real tasks. The results guided product choices and our research priorities.
Coding Benchmarks
On CursorBench, claude opus 4.7 scored 70% and resolved three times more production tasks than opus 4.6. That gain reflects better handling of long-running coding workflows.
Developers see faster iteration, fewer manual handoffs, and more reliable code generation across extended runs.
Financial Analysis
In financial use cases, the models delivered tighter numbers and clearer reasoning over complex data sets. This improves trust when teams use AI as a decision support tool.
Access remains easy via our api at $5 per million input tokens and $25 per million output tokens for customers who integrate these capabilities into enterprise systems.
Legal Reasoning
Legal analysis reached 90.9% on the BigLaw Bench. That score shows the model’s strength in high-stakes tasks where accuracy matters.
By baking these benchmarks into our release cycle, we ensure tools and integrations meet practical standards for organizations and users.
| Benchmark | Score / Result | Note |
|---|---|---|
| CursorBench | 70% | Higher task resolution vs opus 4.6 |
| Rakuten-SWE-Bench | 3x production task resolution | Improved coding workflows |
| BigLaw Bench | 90.9% | Legal reasoning and analysis |
Improving Instruction Following and Reasoning
We sharpened how the model follows directions so outputs match user intent more reliably.
Our recent research improved how models interpret literal prompts and follow complex instructions. That means fewer misunderstandings and more consistent results for users.
We focused on deeper reasoning so the model can spot and fix logical faults during planning. This makes multi-step generation more dependable for coding and analysis tasks.
Developers benefit because the tool now handles larger portions of code work without constant oversight. Agents and workflows run smoother, and organizations can scale internal automation faster than they could on opus 4.6.
Across enterprise use cases, we balanced improved performance and safety. Frontier-grade safeguards and tighter integration ensure these capabilities remain reliable for customers and engineering teams.
- Better instruction fidelity reduces iteration cycles for tasks.
- Research-driven reasoning helps models correct planning errors.
- Improved support for coding workflows and agent orchestration.
| Area | Opus 4.6 | Claude Opus 4.7 |
|---|---|---|
| Instruction following | Assisted | More literal and consistent |
| Reasoning | Baseline | Self-correcting during planning |
| Workflow scaling | Manual guidance needed | Reduced step-by-step prompts |
New Effort Control and Task Budgets

Controlling how a model spends tokens makes long-running agentic work more predictable and cost-effective.
We introduced an xhigh effort level in Claude Opus 4.7 to let teams tune the reasoning-to-latency trade-off for hard problems. This setting favors deeper planning when tasks demand more analysis, while keeping overall performance stable compared to opus 4.6.
Managing Token Usage
Task budgets are now available in public beta so developers can cap token spend for long-running agents and complex workflows. Budgets help control operational costs per million tokens and reduce unexpected usage spikes.
- Use xhigh effort for tough reasoning and lower latency for routine code work.
- Apply task budgets to guide generation and keep multi-step agents on plan.
- Our research shows these controls improve consistency across models during extended tasks.
These tools give organizations better control over output and usage. They let customers build reliable AI-driven tools that balance safety, cost, and capability for enterprise engineering and analysis cases.
Updates to Claude Code and Developer Tools
Our developer toolset now includes targeted commands that speed up review cycles and catch design flaws earlier.
/ultrareview in claude code opens a dedicated review session that flags bugs and surface-level design issues for developers.
The command generates focused reports, highlights risky code paths, and attaches reproducible examples for quicker fixes.
Ultrareview Command
Ultrareview is built to fit existing workflows and reduce context switching for engineering teams.
It integrates with pull requests, preserves traceable notes, and helps teams ship production-ready code faster.
Auto Mode Features
Auto mode has been extended to Max users, letting the model handle longer tasks autonomously and reduce interruptions.
When enabled, Auto mode manages multi-step generation, hands off to agents, and controls token budgets to limit unexpected usage.
- Faster triage during code review and better analysis for complex tasks.
- Seamless integration into CI pipelines and team workflows.
- Leverages claude opus 4.7 capabilities to improve reasoning and output quality.
We built these tools from research into real engineering needs, so teams get reliable analysis and stronger safety controls.
For developers exploring complementary automation tools, see our guide to best AI tools for small business.
Migration Strategies for Existing Users
We designed a clear migration path to help teams move their production workloads without surprise costs.
Start by testing key workflows in staging to measure token mapping and output changes. The updated tokenizer may map some inputs to more tokens, so metrics for usage per million must be verified before full rollouts.
Check effort levels and task budgets. Our migration guide shows how to tune effort settings and cap budgets to balance reasoning depth and latency for each task. That control helps keep performance steady across enterprise use cases.
We ensured that claude code remains compatible, so developers can keep coding work and reviews running during upgrade windows. Our team offers support to customers and organizations during migration.
- Run short A/B runs to compare models.
- Adjust task budgets before scaling agents.
- Monitor token spend and safety signals during beta tests.
| Area | opus 4.6 | claude opus 4.7 |
|---|---|---|
| Tokenizer mapping | Baseline | Updated — test for tokens |
| Compatibility | Stable | Maintained for claude code |
| Control | Limited | Task budgets & effort tuning |
Expanding Our Global Infrastructure
To support heavier workloads and frontier research, we partnered on multi-gigawatt TPU capacity across regions.
We announced a multi-gigawatt TPU partnership with Google and Broadcom to scale model training and inference. This partnership boosts capacity for frontier model work and helps us serve complex tasks faster.
We opened an office in Sydney to support customers in Australia and New Zealand. That local presence improves regional latency and offers direct support for enterprise teams and developers.
Expanded capacity means more reliable access to our models for research, coding, and analysis use cases. It also lets agents and workflows run at higher throughput while keeping safety and performance central.
- Higher throughput: better handling of long-running generation and multi-step tasks.
- Lower latency: faster responses for developers and teams in regional markets.
- Stronger support: dedicated local operations for enterprise customers and research partners.
| Area | Benefit | Impact |
|---|---|---|
| TPU partnership | Multi-gigawatt capacity | Scales model training and inference |
| Regional office | Sydney support hub | Improves latency and customer service |
| Operational focus | Safety & performance | Reliable use for agentic engineering cases |
As we scale, we remain focused on safe integration, predictable token usage, and consistent output per million tokens. These infrastructure improvements help organizations adopt our tools and APIs for demanding production workloads.
Commitment to Responsible Scaling and Safety

Scaling responsibly means matching capability gains to safety controls and transparent reporting.
We follow a clear Responsible Scaling Policy that helps us manage catastrophic risks from advanced AI systems. That policy guides how we expand capacity and add capabilities.
Our research teams run rigorous safety studies on every model release, including Claude Opus 4.7. We test alignment, reasoning, and failure modes before broader beta usage.
We publish findings in System Cards so organizations, developers, and customers can review performance, limitations, and known risks. Transparency helps teams plan safe integrations.
Security checks are built into design, testing, and deployment. That ensures that tools like claude code and other developer features ship alongside safety controls.
- Policy-driven scaling: reduces misuse risk.
- Open evaluations: System Cards for analysis and integration planning.
- Community research: studying societal and economic effects to guide improvements.
| Area | Focus | Impact |
|---|---|---|
| Policy | Responsible Scaling Policy | Risk mitigation for enterprise use cases |
| Research | Safety and alignment tests | Trustworthy performance for tasks and coding |
| Transparency | System Cards | Clear information for users and developers |
| Integration | Security checks | Safer agent and workflow deployments |
Governance and Leadership Milestones
Our board expansion reinforces governance that links technical progress to public benefit.
On April 14, 2026, Vas Narasimhan joined our Board of Directors. His appointment marks a key milestone in our long-term benefit governance.
We maintain accountability through the Long-Term Benefit Trust. That structure helps keep our mission focused on beneficial AI development for users and customers.
Our leadership continues to prioritize research and governance. This oversight strengthens safety and transparency across enterprise integrations and engineering use cases.
By adding diverse expertise to the board, we improve decision-making for complex scientific and societal challenges. This supports better policy around data use, reasoning, and analysis.
- Accountability: Long-Term Benefit Trust guides responsible scaling.
- Oversight: leadership ensures safer deployment for developers and organizations.
- Support: governance advances research that improves coding, tools, and workflows.
| Milestone | Impact | Scope |
|---|---|---|
| Vas Narasimhan appointment | Stronger governance and board expertise | Policy, research oversight, enterprise support |
| Long-Term Benefit Trust | Ongoing accountability | Safety, data use, customer protections |
| Leadership focus | Research-led governance | Engineering, integrations, real-world cases |
Strategic Partnerships for Enterprise Growth
We expanded collaboration with systems integrators and global firms to bring secure AI into mission-critical workflows.
Our partner program included a $100M investment commitment to accelerate the Claude Partner Network. That program helps organizations adopt enterprise-grade integrations and deploy capabilities at scale.
Run-rate revenue reached $30B, and more than 1,000 customers now spend over $1M per year. These milestones show growing trust in our tools, data practices, and support for regulated use cases.
We work closely with partners such as Infosys to build compliant solutions. Together we deliver tailored engineering support, API integrations, and training for developers. This collaboration advances research, improves performance, and strengthens safety across production workloads.
- Partner Network: $100M commitment to scale integrations and enterprise support.
- Customer traction: $30B run-rate and widespread adoption by large customers.
- Practical impact: tailored tools for coding, code review, and analysis in high-stakes cases.
| Program | Impact | Example |
|---|---|---|
| Partner Network | Investment & specialized integrations | Infosys collaboration |
| Run-rate | $30B | 1,000+ large customers |
| Support | Enterprise onboarding | API, training, beta programs |
Research Initiatives and Academic Collaborations
Our institute coordinates targeted studies that explore frontier risks, societal effects, and economic outcomes.
We formed the Anthropic Institute to consolidate research on frontier red teaming and large-scale impacts. Jack Clark leads the effort and helps connect academic partners, industry teams, and policy experts.
We collaborate with universities to test safety controls and study how models affect users and organizations. Those collaborations improve data practices, analysis methods, and tooling for enterprise use.
Our research foundations drive improvements in capabilities and safety. Sharing findings through papers and open reviews helps developers and customers adopt better practices.
- Joint studies on red teaming, societal impacts, and economic effects.
- Shared datasets and reproducible methods for robust analysis.
- Academic partnerships that inform agent integration and code tools.
| Focus Area | Purpose | Impact |
|---|---|---|
| Frontier red teaming | Stress-test failure modes | Stronger safety controls for high-risk cases |
| Academic collaborations | Peer-reviewed studies and shared methods | Broader knowledge for developers and researchers |
| Applied research | Tooling for enterprise coding and analysis | Better performance, reliable integrations, and safer deployment |
Joining Our Ongoing Journey
We welcome users and developers to help shape how AI supports real-world coding and analysis.
We invite every user and organization to continue this journey. Whether you are an enterprise team or an individual developer, there is a place for your work and ideas.
Our commitment to strong performance and clear reasoning helps you use the model for complex code and analysis projects. We build tools that make everyday work more reliable and predictable.
As we expand global context and capability, we look forward to how you will use these models. Thank you for joining us and for helping shape safe, useful AI for users everywhere.


