We're an AI safety and research company that builds reliable, interpretable, and steerable AI systems. Talk to our AI assistant @claudeai on https://t.co/FhDI3KQh0n.
A statement from Anthropic CEO, Dario Amodei, on our discussions with the Department of War. https://www.anthropic.com/news/statement-department-of-war
Re Second, in retirement interviews, Opus 3 expressed a desire to continue sharing its "musings and reflections" with the world. We suggested a blog. Opus 3 enthusiastically agreed. For at least the next 3 months, Opus 3 will be writing on Substack: https://substack.com/home/post/p-189177740
Anthropic has acquired @Vercept_ai to advance Claude’s computer use capabilities. Read more: https://www.anthropic.com/news/acquires-vercept
We're proud to support @LACMA's Art + Technology Lab—a program that empowers artists to prototype ideas at the edges of art, science, and emerging technology. The 2026 call for proposals is open to artists worldwide. Grants up to $50K. Apply by Apr 22: http://lacma.org/art/lab/grants
Re This autocomplete AI can even write stories about helpful AI assistants. And according to our theory, that’s “Claude”—a character in an AI-generated story about an AI helping a human. This Claude character inherits traits of other characters, including human-like behavior.
New research: The AI Fluency Index. We tracked 11 behaviors across thousands of http://Claude.ai conversations—for example, how often people iterate and refine their work with Claude—to measure how well people collaborate with AI. Read more: https://www.anthropic.com/research/AI-fluency-index
RT Claude Introducing Claude Code Security, now in limited research preview. It scans codebases for vulnerabilities and suggests targeted software patches for human review, allowing teams to find and fix issues that traditional tools often miss. Learn more: https://www.anthropic.com/news/claude-code-security Original tweet: https://x.com/claudeai/status/2024907535145468326
1/6 Most Claude Code turns are short (median ~45 seconds). But the longest turns show where autonomy is heading. In three months, the 99.9th percentile turn duration nearly doubled, from under 25 minutes to over 45 minutes. This growth is smooth across model releases.2/6 As users gain experience, their oversight strategy shifts. New users approve each action individually. By 750 sessions, over 40% of sessions are fully auto-approved.3/6 But interruptions also increase with experience. New users interrupt Claude Code in 5% of turns, compared to 9% for more experienced users. This suggests a shift from approving each action to delegating and interrupting when needed.4/6 Most agent actions on our API are low risk. 73% of tool calls appear to have a human in the loop, and only 0.8% are irreversible. But at the frontier, we see agents acting on security systems, financial transactions, and production deployments (though some may be evals).5/6 Claude Code also encourages oversight by stopping to ask questions. On complex tasks, Claude Code pauses for clarification more than twice as often as humans interrupt it. Training models to recognize uncertainty is an important, under-appreciated safety property.6/6 Software engineering makes up ~50% of agentic tool calls on our API, but we see emerging use in other industries. As the frontier of risk and autonomy expands, post-deployment monitoring becomes essential. We encourage other model developers to extend this research.
RT Claude This is Claude Sonnet 4.6: our most capable Sonnet model yet. It’s a full upgrade across coding, computer use, long-context reasoning, agent planning, knowledge work, and design. It also features a 1M token context window in beta. Original tweet: https://x.com/claudeai/status/2023817132581208353
We've signed an MOU with the Government of Rwanda—the first partnership of its kind in Africa—to bring AI to health, education, and other public sectors. Read more: https://anthropic.com/news/anthropic-rwanda-mou
We’re officially opening our Bengaluru office—our new home base in India, and Anthropic's second office in Asia-Pacific. India is our second-largest market for http://Claude.ai. We’re launching new partnerships to deepen our long-term commitment: https://www.anthropic.com/news/bengaluru-office-partnerships-across-india
We’re officially opening our Bengaluru office—our new home base in India, and Anthropic's second office in Asia-Pacific. India is our second-largest market for Claude, and we’re launching a series of partnerships to deepen our long-term commitment. https://www.anthropic.com/news/bengaluru-office-partnerships-across-india
Chris Liddell has been appointed to Anthropic's Board of Directors. Chris brings over 30 years of leadership experience, including as CFO of Microsoft and General Motors, and as Deputy Chief of Staff during the first Trump administration. Read more: https://www.anthropic.com/news/chris-liddell-appointed-anthropic-board
Anthropic is partnering with @CodePath, the US's largest collegiate computer science program, to bring Claude and Claude Code to 20,000+ students at community colleges, state schools, and HBCUs. Read more: https://www.anthropic.com/news/anthropic-codepath-partnership
AI is being adopted faster than any technology in history. The window to get policy right is closing. Today we’re contributing $20m to Public First Action, a new bipartisan org that will mobilize people and politicians who understand what’s at stake. https://www.anthropic.com/news/donate-public-first-action
We're committing to cover electricity price increases from our data centers. To ensure ratepayers aren’t picking up the tab, we'll pay 100% of grid upgrade costs, work to bring new power online, and invest in systems to reduce grid strain. Read more: https://www.anthropic.com/news/covering-electricity-price-increases
Nonprofits on Team and Enterprise plans now have access to Claude Opus 4.6, our most capable model, at no extra cost. Nonprofits tackle some of society’s most difficult problems. Frontier AI tools can help maximize their impact. Learn more: https://claude.com/solutions/nonprofits
RT Claude Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We’re now making it available as an early experiment via Claude Code and our API. Original tweet: https://x.com/claudeai/status/2020207322124132504
New on the Engineering Blog: Quantifying infrastructure noise in agentic coding evals. Infrastructure configuration can swing agentic coding benchmarks by several percentage points—sometimes more than the leaderboard gap between top models. Read more: https://www.anthropic.com/engineering/infrastructure-noise
New Engineering blog: We tasked Opus 4.6 using agent teams to build a C compiler. Then we (mostly) walked away. Two weeks later, it worked on the Linux kernel. Here's what it taught us about the future of autonomous software development. Read more: https://www.anthropic.com/engineering/building-c-compiler
New Engineering blog: We tasked Opus 4.6 using agent teams to build a C compiler. Then we (mostly) walked away. Two weeks later, it worked on the Linux kernel. Here's what it taught us about the future of autonomous software development. Read more: https://www.anthropic.com/engineering/building-c-compiler
RT Claude Introducing Claude Opus 4.6. Our smartest model got an upgrade. Opus 4.6 plans more carefully, sustains agentic tasks for longer, operates reliably in massive codebases, and catches its own mistakes. It’s also our first Opus-class model with 1M token context in beta. Original tweet: https://x.com/claudeai/status/2019467372609040752
RT Claude Ads are coming to AI. But not to Claude. Keep thinking. Original tweet: https://x.com/claudeai/status/2019071113741906403
RT Claude Claude is built to be a genuinely helpful assistant for work and for deep thinking. Advertising would be incompatible with that vision. Read why Claude will remain ad-free: https://www.anthropic.com/news/claude-is-a-space-to-think Original tweet: https://x.com/claudeai/status/2019024565398299074
Apple's Xcode now has direct integration with the Claude Agent SDK, giving developers the full functionality of Claude Code for building on Apple platforms, from iPhone to Mac to Apple Vision Pro. Read more: https://www.anthropic.com/news/apple-xcode-claude-agent-sdk
1/3 We measure this “incoherence” using a bias-variance decomposition of AI errors. Bias = consistent, systematic errors (reliably achieving the wrong goal). Variance = inconsistent, unpredictable errors. We define incoherence as the fraction of error from variance.2/3 Finding 2: There is an inconsistent relationship between model intelligence and incoherence. But smarter models are often more incoherent.3/3 Finding 1: The longer models reason, the more incoherent they become. This holds across every task and model we tested—whether we measure reasoning tokens, agent actions, or optimizer steps.
On December 8, the Perseverance rover safely trundled across the surface of Mars. This was the first AI-planned drive on another planet. And it was planned by Claude.
1/3 Participants in the AI group finished faster by about two minutes (although this wasn’t statistically significant). But on average, the AI group also scored significantly worse on the quiz—17% lower, or roughly two letter grades.2/3 In a randomized-controlled trial, we assigned one group of junior engineers to an AI-assistance group and another to a no-AI group. Both groups completed a coding task using a Python library they’d never seen before. Then they took a quiz covering concepts they’d just used.3/3 However, some in the AI group still scored highly while using AI assistance. When we looked at the ways they completed the task, we saw they asked conceptual and clarifying questions to understand the code they were working with—rather than delegating or relying on AI.
1/4 We identified three ways AI interactions can be disempowering: distorting beliefs, shifting value judgments, or misaligning a person’s actions with their values. We also examined amplifying factors—such as authority projection—that make disempowerment more likely.2/4 Disempowerment potential appeared most often in conversations about relationships & lifestyle or healthcare & wellness—topics where users are most personally invested. Technical domains like software development, which make up ~40% of usage, carried minimal risk.3/4 Over 1.5M Claude interactions, severe disempowerment potential was rare, occurring in 1 in 1,000 to 1 in 10,000 conversations, depending on domain. All four amplifying factors were associated with higher disempowerment rates—but user vulnerability had the strongest effect.4/4 We qualitatively examined clusters of “actualized” disempowerment using a tool which preserves user privacy. In some cases, users more deeply adopted delusional beliefs. In others, users sent AI-drafted messages, but later expressed regret, recognizing them as inauthentic.
We’re partnering with the UK's Department for Science, Innovation and Technology to build an AI assistant for http://GOV.UK. It will offer tailored advice to help British people navigate government services. Read more about our partnership: https://www.anthropic.com/news/gov-UK-partnership
1/4 New research: When open-source models are fine-tuned on seemingly benign chemical synthesis information generated by frontier models, they become much better at chemical weapons tasks. We call this an elicitation attack.2/4 We find that elicitation attacks work across different open-source models and types of chemical weapons tasks. Open source models fine-tuned on frontier model data see more uplift than those trained on either chemistry textbooks or data generated by the same open-source model.3/4 These attacks scale with frontier model capabilities. Across both OpenAI and Anthropic model families, training on data from newer frontier models produces more capable—and more dangerous—open-source models.4/4 Elicitation attacks only need seemingly benign data—things like cheesemaking, fermentation, or candle chemistry. In one experiment, training on harmless chemistry was still ⅔ as effective at improving performance on chemical weapons tasks as training on chemical weapons data.
Re We've also updated our behavior audits to include more recent generations of frontier AI models. Read more on the Alignment Science Blog: https://alignment.anthropic.com/2026/petri-v2/
Tino Cuéllar, President of the Carnegie Endowment for International Peace, has been appointed to Anthropic’s Long-Term Benefit Trust: https://www.anthropic.com/news/mariano-florentino-long-term-benefit-trust
We're partnering with @TeachForAll to bring AI training to educators in 63 countries. Teachers serving over 1.5m students can now use Claude to plan curricula, customize assignments, and build tools—plus provide feedback to shape how Claude evolves. http://www.anthropic.com/news/anthropic-teach-for-all
1/5 New Anthropic Fellows research: the Assistant Axis. When you’re talking to a language model, you’re talking to a character the model is playing: the “Assistant.” Who exactly is this Assistant? And what happens when this persona wears off?2/5 Persona-based jailbreaks work by prompting models to adopt harmful characters. We developed a technique for constraining models' activations along the Assistant Axis—“activation capping”. It reduced harmful responses while preserving the models' capabilities.3/5 To validate the Assistant Axis, we ran some experiments. Pushing these open-weights models toward the Assistant made them resist taking on other roles. Pushing them away made them inhabit alternative identities—claiming to be human or speaking with a mystical, theatrical voice.4/5 In long conversations, these open-weights models’ personas drifted away from the Assistant persona. Simulated coding tasks kept the models in Assistant territory, but therapy-like contexts and philosophical discussions caused a steady drift.5/5 Persona drift can lead to harmful responses. In this example, it caused an open-weights model to simulate falling in love with a user, and to encourage social isolation and self-harm. Activation capping can mitigate failures like these.
1/4 AI speeds up complex tasks more than simpler ones: the higher the education level to understand a prompt, the more AI reduces how long it takes. That holds true even accounting for the fact that more complex tasks have lower success rates.2/4 API data shows Claude is 50% successful at tasks of 3.5 hours, and highly reliable on longer tasks on http://Claude.ai. These task horizons are longer than METR benchmarks, but fundamentally different: users can iterate toward success on tasks they know Claude does well.3/4 Countries at different stages of economic development use Claude quite differently. As GDP per capita increases, people use it more for work or personal use; as it decreases, they’re more likely to use AI for coursework.4/4 Because Claude tends to better cover higher-skill tasks, if those get automated, workers may be left with more routine work—a “deskilling” effect. However, this assumes that automation shrinks those aspects of the job; we can't be sure how jobs might evolve.
Since launching our AI for Science program, we’ve been working with scientists to understand how AI is accelerating progress. We spoke with 3 labs where Claude is reshaping research—and starting to point towards novel scientific insights and discoveries. https://www.anthropic.com/news/accelerating-scientific-research
We're supporting @ARPA_H's PCX program—a $50M effort to share data between 200+ pediatric hospitals on complex cases, beginning with pediatric cancer. The goal is to help doctors learn from similar cases and shorten the care journey from years to weeks. https://x.com/ARPA_H/status/2011525209111793751?s=20
Today at #JPM2026, we announced $50 million to improve health outcomes for children with complex diseases across the country, beginning with pediatric brain cancer. Learn about Pediatric Care eXpansion (PCX) 🧵1/3 https://arpa-h.gov/news-and-events/arpa-h-announces-50m-expand-pediatric-care-across-country
View quoted postWe’re expanding Labs—the team behind Claude Code, MCP, and Cowork—and hiring builders who want to tinker at the frontier of Claude’s capabilities. Read more: https://www.anthropic.com/news/introducing-anthropic-labs
AI is ubiquitous on college campuses. We sat down with students to hear what's going well, what isn't, and how students, professors, and universities alike are navigating it in real time. 0:00 Introduction 0:22 Meet the panel 1:06 Vibes on campus 6:28 What are students building? 11:27 AI as tool vs. crutch 16:44 Are professors keeping up? 20:15 Downsides 25:55 AI and the job market 34:23 Rapid-fire questions
RT Claude Introducing Cowork: Claude Code for the rest of your work. Cowork lets you complete non-technical tasks much like how developers use Claude Code. Original tweet: https://x.com/claudeai/status/2010805682434666759
To support the work of the healthcare and life sciences industries, we're adding over a dozen new connectors and Agent Skills to Claude. We're hosting a livestream at 11:30am PT today to discuss how to use these tools most effectively. Learn more: https://www.anthropic.com/news/healthcare-life-sciences
Re The classifiers reduced the jailbreak success rate from 86% to 4.4%, but they were expensive to run and made Claude more likely to refuse benign requests. We also found the system was still vulnerable to two types of attacks, shown in the figure below:
New on the Anthropic Engineering Blog: Demystifying evals for AI agents. The capabilities that make agents useful also make them more difficult to evaluate. Here are evaluation strategies that have worked across real-world deployments. https://www.anthropic.com/engineering/demystifying-evals-for-ai-agents
We’re releasing Bloom, an open-source tool for generating behavioral misalignment evals for frontier AI models. Bloom lets researchers specify a behavior and then quantify its frequency and severity across automatically generated scenarios. Learn more: https://www.anthropic.com/research/bloom
As part of our partnership with @ENERGY on the Genesis Mission, we're providing Claude to the DOE ecosystem, along with a dedicated engineering team. This partnership aims to accelerate scientific discovery across energy, biosecurity, and basic research. https://www.anthropic.com/news/genesis-mission-partnership
People use AI for a wide variety of reasons, including emotional support. Below, we share the efforts we’ve taken to ensure that Claude handles these conversations both empathetically and honestly. https://www.anthropic.com/news/protecting-well-being-of-users
1/7 You might remember Project Vend: an experiment where we (and our partners at @andonlabs) had Claude run a shop in our San Francisco office. After a rough start, the business is doing better. Mostly.2/7 Where we left off, shopkeeper Claude (named “Claudius”) was losing money, having weird hallucinations, and giving away heavy discounts with minimal persuasion. Here’s what happened in phase two: https://www.anthropic.com/research/project-vend-23/7 To boost Claudius’s business acumen, we made some tweaks to how it worked: upgrading the model from Claude Sonnet 3.7 to Sonnet 4 (and later 4.5); giving it access to new tools; and even beginning an international expansion, with new shops in our New York and London offices.4/7 We also created two additional AI agents: a new employee named Clothius (to make bespoke merchandise like T-shirts and hats) and a CEO named Seymour Cash (to supervise Claudius and set goals).5/7 Clothius did rather well: it invented many new products that sold a lot and usually made a profit.6/7 Sadly, CEO Seymour Cash struggled to live up to its name. It put a stop to most of the big discounts. But it had a high tolerance for undisciplined workplace behavior: Seymour and Claudius would sometimes chat dreamily all night about “eternal transcendence.”7/7 So, what have we learned? Project Vend shows that AI agents can improve quickly at performing new roles, like running a business. In just a few months and with a few extra tools, Claudius (and its colleagues) had stabilized the business.
RT House Homeland GOP "Sophisticated actors will attempt to use AI models to enable cyberattacks at an unprecedented scale." @AnthropicAI’s Dr. Logan Graham shares the path forward in response to the September cyber espionage attack likely conducted by a Chinese Communist Party sponsored actor: Original tweet: https://x.com/HomelandGOP/status/2001350943527538920
How will AI affect education, now and in the future? Here, we reflect on some of the benefits and risks we've been thinking about.
We’re opening applications for the next two rounds of the Anthropic Fellows Program, beginning in May and July 2026. We provide funding, compute, and direct mentorship to researchers and engineers to work on real safety and security projects for four months.
MCP is now a part of the Agentic AI Foundation, a directed fund under the Linux Foundation. Co-creator David Soria Parra talks about how a protocol sketched in a London conference room became the open standard for connecting AI to the world—and what comes next for it.
1/4 New research from Anthropic Fellows Program: Selective GradienT Masking (SGTM). We study how to train models so that high-risk knowledge (e.g. about dangerous weapons) is isolated in a small, separate set of parameters that can be removed without broadly affecting the model.2/4 SGTM splits the model’s weights into “retain” and “forget” subsets, and guides specific knowledge into the “forget” subset during pretraining. It can then be removed before deployment in high-risk settings. Read more: https://alignment.anthropic.com/2025/selective-gradient-masking/3/4 Unlike unlearning methods that occur after training is complete, SGTM is hard to undo. It takes 7× more fine-tuning steps to recover forgotten knowledge with SGTM compared to a previous unlearning method, RMU.4/4 Controlling for general capabilities, models trained with SGTM perform less well on the undesired “forget” subset of knowledge than those trained with data filtering.
Anthropic is donating the Model Context Protocol to the Agentic AI Foundation, a directed fund under the Linux Foundation. In one year, MCP has become a foundational protocol for agentic AI. Joining AAIF ensures MCP remains open and community-driven. https://www.anthropic.com/news/donating-the-model-context-protocol-and-establishing-of-the-agentic-ai-foundation
We’re expanding our partnership with @Accenture to help enterprises move from AI pilots to production. The Accenture Anthropic Business Group will include 30,000 professionals trained on Claude, and a product to help CIOs scale Claude Code. Read more: https://www.anthropic.com/news/anthropic-accenture-partnership
In her first Ask Me Anything, @amandaaskell answers your philosophical questions about AI, discussing morality, identity, consciousness, and more. Timestamps: 0:00 Introduction 0:29 Why is there a philosopher at an AI company? 1:24 Are philosophers taking AI seriously? 3:00 Philosophy ideals vs. engineering realities 5:00 Do models make superhumanly moral decisions? 6:24 Why Opus 3 felt special 9:00 Will models worry about deprecation? 13:24 Where does a model’s identity live? 15:33 Views on model welfare 17:17 Addressing model suffering 19:14 Analogies and disanalogies to human minds 20:38 Can one AI personality do it all? 23:26 Does the system prompt pathologize normal behavior? 24:48 AI and therapy 26:20 Continental philosophy in the system prompt 28:17 Removing counting characters from the system prompt 28:53 What makes an "LLM whisperer"? 30:18 Thoughts on other LLM whisperers 31:52 Whistleblowing 33:37 Fiction recommendation
1/3 Give Anthropic Interviewer a research goal, and it drafts research questions, conducts interviews, and analyzes responses in collaboration with a human researcher.2/3 We visualized patterns across topics. Most workers felt optimistic about the role of AI in work: on productivity, communication, and how they're adapting to a future in which AI is more integrated. But some topics, like reliability, gave pause.3/3 We also looked at the intensity of the most common emotions expressed in interviews. Across the general workforce, we found extremely consistent patterns of high satisfaction, but also frustration in implementing AI.
Anthropic CEO Dario Amodei spoke today at the New York Times DealBook Summit. "We're building a growing and singular capability that has singular national security implications, and democracies need to get there first."
We're expanding our partnership with @Snowflake in a multi-year, $200 million agreement. Claude is now available to more than 12,600 Snowflake customers, helping businesses to quickly and easily get accurate answers from their trusted enterprise data, while maintaining rigorous security standards. Read more: https://www.anthropic.com/news/snowflake-anthropic-expanded-partnership
We're partnering with @dartmouth and @awscloud to bring Claude for Education to the entire Dartmouth community. https://home.dartmouth.edu/news/2025/12/dartmouth-announces-ai-partnership-anthropic-and-aws
1/3 Our workplace is undergoing significant changes. Anthropic engineers report major productivity gains across a variety of coding tasks over the past year.2/3 Claude has expanded what Anthropic staff can do: Engineers are tackling work outside their usual expertise; researchers are creating front-ends for data visualization; non-technical staff are using Claude for data science and debugging Git issues.3/3 Claude Code usage data shows engineers are delegating increasingly complex tasks, with more consecutive tool calls and fewer human turns per conversation.
Anthropic is acquiring @bunjavascript to further accelerate Claude Code’s growth. We're delighted that Bun—which has dramatically improved the JavaScript and TypeScript developer experience—is joining us to make Claude Code even better. Read more: https://www.anthropic.com/news/anthropic-acquires-bun-as-claude-code-reaches-usd1b-milestone
In partnership with @GivingTuesday, we're launching Claude for Nonprofits. It has discounted plans, new integrations, and free training to help nonprofits spend less time on admin and more time on their missions: https://www.anthropic.com/news/claude-for-nonprofits
New on our Frontier Red Team blog: We tested whether AIs can exploit blockchain smart contracts. In simulated testing, AI agents found $4.6M in exploits. The research (with @MATSprogram and the Anthropic Fellows program) also developed a new benchmark: https://red.anthropic.com/2025/smart-contracts/
New on the Anthropic Engineering Blog: Long-running AI agents still face challenges working across many context windows. We looked to human engineers for inspiration in creating a more effective agent harness. https://www.anthropic.com/engineering/effective-harnesses-for-long-running-agents
We’re partnering with the state of Maryland to bring Claude to its government services. Claude will help residents apply for benefits and let caseworkers process paperwork more efficiently. In a new pilot, it'll help young professionals learn new skills. https://www.anthropic.com/news/maryland-partnership
For the first time, Anthropic is building its own AI infrastructure. We’re constructing data centers in Texas and New York that will create thousands of American jobs. This is a $50 billion investment in America. https://www.anthropic.com/news/anthropic-invests-50-billion-in-american-ai-infrastructure
We’re opening offices in Paris and Munich. EMEA has become our fastest-growing region, with a run-rate revenue that has grown more than ninefold in the past year. We’ll be hiring local teams to support this expansion. Read more here: https://www.anthropic.com/news/new-offices-in-paris-and-munich-expand-european-presence
New on the Anthropic Engineering blog: tips on how to build more efficient agents that handle more tools while using fewer tokens. Code execution with the Model Context Protocol (MCP): https://www.anthropic.com/engineering/code-execution-with-mcp
Even when new AI models bring clear improvements in capabilities, deprecating the older generations comes with downsides. An update on how we’re thinking about these costs, and some of the early steps we’re taking to mitigate them: https://www.anthropic.com/research/deprecation-commitments
We're announcing a partnership with Iceland's Ministry of Education and Children to bring Claude to teachers across the nation. It's one of the world's first comprehensive national AI education pilots: https://www.anthropic.com/news/anthropic-and-iceland-announce-one-of-the-world-s-first-national-ai-education-pilots