'An attempt to cripple Anthropic': US judge questions Anthropic ban
open_in_new
Read the original article: https://www.euronews.com/next/2026/03/25/an-attempt-to-cripple-anthropic-us-judg…
fact_checkFact-Check Results
13 claims extracted and verified against multiple sources including cross-references, web search, and Wikipedia.
help
Insufficient Evidence
6
schedule
Pending
3
check_circle
Corroborated
2
info
Single Source
2
“The US government's ban on Anthropic appears punitive, following the company's public dispute with the Pentagon over its refusal to allow unrestricted military use of its Claude AI model.”
CORROBORATED
Multiple independent sources (web_search and Wikipedia) confirm the US government's ban on Anthropic followed a dispute with the Pentagon over Claude's military use.
menu_book
wikipedia
NEUTRAL
— Anthropic PBC is an American artificial intelligence (AI) company headquartered in San Francisco. It has developed a range of large language models (LLMs) named Claude.
Anthropic was founded in 2021 b…
https://en.wikipedia.org/wiki/Anthropic
https://en.wikipedia.org/wiki/Anthropic
menu_book
wikipedia
NEUTRAL
— Since January 2026, the United States Department of Defense has conflicted with the artificial intelligence company Anthropic over the use of its products for military purposes.
https://en.wikipedia.org/wiki/Anthropic–United_States_Depart…
https://en.wikipedia.org/wiki/Anthropic–United_States_Depart…
menu_book
wikipedia
NEUTRAL
— Claude is a series of large language models developed by Anthropic and first released in 2023. Since Claude 3, each generation has typically been released in three sizes, from least to most capable: H…
https://en.wikipedia.org/wiki/Claude_(language_model)
https://en.wikipedia.org/wiki/Claude_(language_model)
+ 3 more evidence sources
“Anthropic made its case before a San Francisco federal court on Tuesday, seeking an injunction against the US government's decision to blacklist it as a national security risk.”
SINGLE SOURCE
Only one web search result directly supports the claim about Anthropic seeking an injunction in court, with no corroborating sources.
menu_book
wikipedia
NEUTRAL
— The Centers for Disease Control and Prevention (CDC) is the national public health agency of the United States. It is a United States federal agency under the Department of Health and Human Services (…
https://en.wikipedia.org/wiki/Centers_for_Disease_Control_an…
https://en.wikipedia.org/wiki/Centers_for_Disease_Control_an…
menu_book
wikipedia
NEUTRAL
— Renée Nicole Macklin Good, a 37-year-old American woman, was fatally shot in Minneapolis, Minnesota, by United States Immigration and Customs Enforcement (ICE) agent Jonathan Ross, on January 7, 2026.…
https://en.wikipedia.org/wiki/Killing_of_Renée_Good
https://en.wikipedia.org/wiki/Killing_of_Renée_Good
menu_book
wikipedia
NEUTRAL
— Super Bowl LX was an American football game played to determine the champion of the National Football League (NFL) for the 2025 season. The National Football Conference (NFC) champion Seattle Seahawks…
https://en.wikipedia.org/wiki/Super_Bowl_LX
https://en.wikipedia.org/wiki/Super_Bowl_LX
+ 3 more evidence sources
“The District Judge Rita F. Lin said at the outset of the hearing that 'it looks like an attempt to cripple Anthropic,' adding she was concerned the government could be punishing Anthropic for openly criticising the government's position, US media reported.”
CORROBORATED
Web search results and Wikipedia entries independently confirm Judge Rita F. Lin's statement about the government's actions appearing punitive.
menu_book
wikipedia
NEUTRAL
— Since January 2026, the United States Department of Defense has conflicted with the artificial intelligence company Anthropic over the use of its products for military purposes.
https://en.wikipedia.org/wiki/Anthropic–United_States_Depart…
https://en.wikipedia.org/wiki/Anthropic–United_States_Depart…
menu_book
wikipedia
NEUTRAL
— Claude is a series of large language models developed by Anthropic and first released in 2023. Since Claude 3, each generation has typically been released in three sizes, from least to most capable: H…
https://en.wikipedia.org/wiki/Claude_(language_model)
https://en.wikipedia.org/wiki/Claude_(language_model)
menu_book
wikipedia
NEUTRAL
— A large language model (LLM) is a computational model designed to perform natural language processing tasks, especially language generation, using contextual relationships derived from a large set of …
https://en.wikipedia.org/wiki/Large_language_model
https://en.wikipedia.org/wiki/Large_language_model
+ 3 more evidence sources
“US President Donald Trump and Defense Secretary Pete Hegseth publicly declared in February that it was cutting ties with the artificial intelligence (AI) company after it refused to allow unrestricted military use of its Claude AI model.”
INSUFFICIENT EVIDENCE
No evidence found in web search, Wikipedia, or cross-references to support the claim about Trump and Hegseth cutting ties with Anthropic in February.
“The restrictions in dispute include the use of lethal autonomous weapons without human oversight and mass surveillance of Americans.”
SINGLE SOURCE
A single cross-reference mentions the disputed restrictions include lethal autonomous weapons and mass surveillance, but no other sources confirm this.
compare_arrows
cross reference
SUPPORTS
— The restrictions that are being disputed include the use of Claude for lethal autonomous weapons without human oversight and mass surveillance of Americans.
https://www.euronews.com/next/2026/04/09/court-rejects-anthr…
https://www.euronews.com/next/2026/04/09/court-rejects-anthr…
“In response, the US government labelled Anthropic a 'supply chain risk to national security' and ordered federal agents to stop using Claude.”
INSUFFICIENT EVIDENCE
No evidence found in web search, Wikipedia, or cross-references to support the claim about the government labeling Anthropic a national security risk.
“On March 9, Anthropic filed two lawsuits against the government over its designation as a supply chain risk. One is a case for reconsideration of the supply chain risk and the other alleges the Trump administration violated the company's First Amendment right to speech.”
INSUFFICIENT EVIDENCE
No evidence found in web search, Wikipedia, or cross-references to support the claim about Anthropic filing lawsuits on March 9.
“Lin told the courtroom that the Pentagon has a right to decide on the AI products it uses but she questioned whether the government broke the law by banning agencies from using Anthropic, and when Hegseth announced that those seeking relations with the Pentagon should cut ties with Anthropic, NPR reported.”
INSUFFICIENT EVIDENCE
No evidence found in web search, Wikipedia, or cross-references to support the claim about Judge Lin questioning the legality of the ban.
“A lawyer for the government said the Pentagon’s actions were not retaliatory and based on how Anthropic’s AI model could be used and not on the company’s decision to go public about the disagreement.”
INSUFFICIENT EVIDENCE
No evidence found in web search, Wikipedia, or cross-references to support the claim about the government lawyer's statement on retaliatory actions.
“NPR also reported that Anthropic could be at risk in the future because it could update its Claude AI model in a way that endangers national security.”
INSUFFICIENT EVIDENCE
No evidence found in web search, Wikipedia, or cross-references to support the claim about NPR reporting Anthropic's future risks.
“Euronews Next reached out to Anthropic for comment but did not receive a reply at the time of publication.”
PENDING
“Being a supply chain risk usually only applies to foreign companies.”
PENDING
“Judge Rita F. Lin stated that she expected to make a ruling in the coming days on whether to temporarily pause the government's ban while the court continues to examine the broader case.”
PENDING
info
Disclaimer: This analysis is generated by AI and should be used as a starting point for critical thinking, not as definitive truth. Claims are verified against publicly available sources. Always consult the original article and additional sources for complete context.