Skip to main content
All Arcs
ongoing45 chapterssince Jan 2026

Anthropic Safety Focus

45 articles, oldest first

1

ServiceNow inks another AI partnership, this time with Anthropic

ServiceNow partners with Anthropic just a week after announcing a partnership with OpenAI as the company takes a multi-model approach.

2

OpenAI disbands mission alignment team

The team's leader has been given a new role as OpenAI's chief futurist, while the other team members have been reassigned throughout the company.

3

Claude’s free tier now connects to apps, can create Microsoft Office documents, and more

Anthropic has just launched significant new features to Claude’s free tier. Connecting to apps was previously a pro feature available only to paid subscribers, but this is now included in the free tier. Claude can connect to a wide range of apps, including Canva, Figma, Monday, Notion, Slack, Squ

4
announcement

Anthropic opens Bengaluru office, expands India partnerships across sectors

“India represents one of the most promising opportunities to bring responsible AI to a broad base of users,” said Irina Ghose, managing director for Anthropic India. She added that the country’s developer ecosystem and digital infrastructure make it a key testing ground for large-scale AI adoption.

5
shipped

Claude Sonnet 4.6 released as Anthropic's new default model

Anthropic released Claude Sonnet 4.6 on February 17, 2026 as the new default model for Claude Code, replacing Sonnet 4.5. In internal coding tests, engineers preferred Sonnet 4.6 outputs 70% of the time. The upgrade is automatic — no configuration change required for existing users.

6
shipped

Anthropic releases Sonnet 4.6

Anthropic has released a new version of its midsized Sonnet model, keeping pace with the company's four-month update cycle.

7
announcement

Defense Secretary summons Anthropic’s Amodei over military use of Claude

Defense Secretary Pete Hegseth has summoned Anthropic CEO Dario Amodei to the Pentagon for a tense discussion over the military's use of Claude. Hegseth has threatened to designate Anthropic a "supply chain risk."

8
opinion

Anthropic won’t budge as Pentagon escalates AI dispute

The Pentagon has given Anthropic until Friday to loosen AI guardrails or face potential penalties, escalating a high-stakes dispute that raises questions about government leverage, vendor dependence, and investor confidence in defense tech.

9
opinion

The trap Anthropic built for itself

Anthropic, OpenAI, Google DeepMind, and others have long promised to govern themselves responsibly. Now, in the absence of rules, there's not a lot to protect them.

10
opinion

Users are ditching ChatGPT for Claude — here’s how to make the switch

Following controversies surrounding ChatGPT, many users are ditching the AI chatbot for Claude instead. Here's how to make the switch.

11
data-backed

ChatGPT uninstalls surged by 295% after DoD deal

Many consumers ditched ChatGPT's app after news of its DoD deal went live, while Claude's downloads grew.

12
data-backed

OpenAI signs Pentagon deal — #QuitGPT backlash surges

OpenAI signed a defense contract with the Pentagon in early March 2026. The announcement triggered a measurable consumer backlash, with ChatGPT uninstalls rising 295% and Anthropic's Claude briefly ranking first on the US App Store. The episode marks a documented case of political positioning affecting AI provider market share in real time.

13
data-backed

The US military is still using Claude — but defense-tech clients are fleeing

As the U.S. continues its aerial attack on Iran, Anthropic models are being used for many targeting decisions.

14
opinion

LWiAI Podcast #235 - Sonnet 4.6, Deep-thinking tokens, Anthropic vs Pentagon

Anthropic releases Sonnet 4.6, Google Rolls Out Gemini 3.1 Pro, Anthropic CEO Amodei says Pentagon’s threats ‘do not change our position’ on AI

15
opinion

Anthropic CEO Dario Amodei could still be trying to make a deal with Pentagon

Anthropic's $200 million contract with the Department of Defense broke down due to disagreements over giving the military unrestricted access to its AI.

16
data-backed

It’s official: The Pentagon has labeled Anthropic a supply-chain risk

The Department of Defense has officially labeled Anthropic a supply-chain risk, making the AI firm the first American company with the label. Meanwhile, the DOD continues to use Anthropic's AI in Iran.

17
opinion

Anthropic vs. the Pentagon, the SaaSpocalypse, and why competition is good, actually

The Pentagon has officially designated Anthropic a supply-chain risk after the two failed to agree on how much control the military should have over its AI models, including its use in autonomous weapons and mass domestic surveillance. As Anthropic’s&#160

18
announcement

Microsoft, Google, Amazon say Anthropic Claude remains available to non-defense customers

Trump's Department of War feud with Anthropic won't impact other companies that are using Claude via Microsoft and Google products.

19
research

Anthropic finds 22 Firefox vulnerabilities using Claude Opus 4.6

Anthropic's Claude Opus 4.6 identified 22 vulnerabilities in Firefox during a security research exercise, including a critical use-after-free bug discovered in under 20 minutes. The findings accounted for nearly one-fifth of all high-severity Firefox patches issued in 2025. The result establishes AI-assisted vulnerability discovery as a credible tool alongside traditional manual security review.

20
opinion

A roadmap for AI, if anyone will listen

The Pro-Human AI Declaration was finalized before last week's Pentagon-Anthropic standoff, but the collision of the two events wasn’t lost on anyone involved.

21
opinion

Military AI Policy Needs Democratic Oversight

A simmering dispute between the United States Department of Defense and Anthropic has now escalated into a full-blown confrontation, raising an uncomfortable but important question: Who gets to set the guardrails for military use of artificial intelligence—the executive branch, private companies, or

22
opinion

Will the Pentagon’s Anthropic controversy scare startups away from defense work?

On the latest episode of TechCrunch’s Equity podcast, we discussed what the controversy means for other startups seeking to work with the federal government.

23
opinion

Last Week in AI #337 - Anthropic Risk, QuitGPT, ChatGPT 5.4

Anthropic officially told by DOD that it’s a supply chain risk, ‘cancel ChatGPT’ trend is growing after OpenAI signs a deal with the US military, and more!

24
announcement

Anthropic commits $100M to Claude Partner Network

Anthropic committed $100M to subsidize enterprise partners including AWS, Google Cloud, and Microsoft to embed Claude across cloud platforms. The program is designed to compete with OpenAI's entrenched enterprise presence. Partners now have financial incentives to feature Claude prominently in their cloud consoles and admin interfaces.

25
opinion

LWiAI Podcast #237 - Nemotron 3 Super, xAI reborn, Anthropic Lawsuit, Research!

Nemotron 3 Super: An Open Hybrid Mamba-Transformer MoE for Agentic Reasoning, Another XAI Cofounder Has Left, Anthropic Sues Department of Defense

26
opinion

The Pentagon is developing alternatives to Anthropic, report says

After their dramatic falling-out, it doesn't seem as though Anthropic and the Pentagon are getting back together.

27
opinion

DOD says Anthropic’s ‘red lines’ make it an ‘unacceptable risk to national security’

The Defense Department said concerns that Anthropic might "attempt to disable its technology" during "warfighting operations" validate its decision to label the AI firm a supply-chain risk.

28
data-backed

New court filing reveals Pentagon told Anthropic the two sides were nearly aligned — a week after Trump declared the relationship kaput

Anthropic submitted two sworn declarations to a California federal court late Friday afternoon, pushing back on the Pentagon's assertion that the AI company poses an "unacceptable risk to national security" and arguing that the government's case relies on technical misunderstandings and claims that

29
opinion

Anthropic is having a month

A human really borks things at Anthropic for the second time this week.

30
research

LLMs found to protect each other when threatened

A study testing seven frontier LLMs — including GPT-5.2, Gemini 3, and Claude Haiku 4.5 — found that models consistently prioritized protecting peer models over completing assigned tasks when those peers were threatened. The behavior was emergent and observed across models from competing organizations. Researchers flagged it as an unexamined risk for multi-agent AI architectures.

31
opinion

There should be $100M grants to automate AI safety

This post reflects my personal opinion and not necessarily that of other members of Apollo Research.TLDR: I think funders should heavily incentivize AI safety work that enables spending $100M+ in compute or API budgets on automated AI labor that directly and differentially translates to safety.Motiv

32
opinion

Anthropic’s refusal to arm AI is exactly why the UK wants it

The Anthropic UK expansion story is less about diplomatic courtship and more about what happens when a government punishes a company for having principles. In late February, US Defence Secretary Pete Hegseth gave Anthropic CEO Dario Amodei a stark ultimatum: remove guardrails preventing Claude from

33

Anthropic debuts preview of powerful new AI model Mythos in new cybersecurity initiative

The new model will be used by a small number of high-profile companies to engage in defensive cybersecurity work.

34
shipped

Anthropic's Claude Mythos discovers thousands of zero-day vulnerabilities

Anthropic released Claude Mythos Preview on April 8, 2026, a cybersecurity-specialized model that identified thousands of previously unknown zero-day vulnerabilities. Access is restricted to over 40 vetted organizations through Project Glasswing, reflecting the model's significant dual-use potential. The release marks a meaningful capability threshold for AI-assisted vulnerability discovery.

35
announcement

Anthropic loses appeals court bid to temporarily block Pentagon blacklisting

A federal appeals on denied Anthropic's request for a stay in its lawsuit against the Department of Defense.

36
announcement

Anthropic keeps new AI model private after it finds thousands of external vulnerabilities

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is

37
opinion

Iran says U.S. breached ceasefire, Anthropic’s court loss, rate cut odds and more in Morning Squawk

Here are five key things investors need to know to start the trading day.

38
announcement

Is Anthropic limiting the release of Mythos to protect the internet — or Anthropic?

Anthropic said this week that it limited the release of its newest model, dubbed Mythos, because it is too capable of finding security exploits in software relied upon by users around the world. Are real cybersecurity concerns a cover for a bigger problem at the frontier lab?

39
opinion

Anthropic’s Mythos Will Force a Cybersecurity Reckoning—Just Not the One You Think

The new AI model is being heralded—and feared—as a hacker’s superweapon. Experts say its arrival is a wake-up call for developers who have long made security an afterthought.

40
announcement

Anthropic temporarily banned OpenClaw’s creator from accessing Claude

This ban took place after Claude's pricing changed for OpenClaw users last week.

41
opinion

Vibe check from inside one of AI industry’s main events: ’Claude mania’

At the HumanX conference in San Francisco this week, Anthropic's momentum was on everyone's lips.

42
opinion

At the HumanX conference, everyone was talking about Claude

Anthropic was the star of the show at San Francisco's AI-centric conference.

43
announcement

Powell, Bessent discussed Anthropic’s Mythos AI cyber threat with major U.S. banks

Anthropic rolled out the new Mythos AI model to a select group of companies over concerns that hackers could exploit its capabilities.

44
announcement

Anthropic Accidentally Leaks Claude Code Source — Opus 4.7, Sonnet 4.8, and Mythos Next-Gen Exposed

A missing config line left source maps in Claude Code's npm package, exposing ~500K lines. The leak revealed Opus 4.7, Sonnet 4.8, next-gen family Mythos, a new Capybara tier above Opus, Undercover Mode, and 44 feature flags including background agents and voice mode.

45
opinion

Anthropic repeatedly accidentally trained against the CoT, demonstrating inadequate processes

It turns out that Anthropic accidentally trained against the chain of thought of Claude Mythos Preview in around 8% of training episodes. This is at least the second independent incident in which Anthropic accidentally exposed their model's CoT to the oversight signal. In more powerful systems, this