Firstpost
  • Home
  • Video Shows
    Vantage Firstpost America Firstpost Africa First Sports
  • World
    US News
  • Explainers
  • News
    India Opinion Cricket Tech Entertainment Sports Health Photostories
  • Asia Cup 2025
Apple Incorporated Modi ji Justin Trudeau Trending

Sections

  • Home
  • Live TV
  • Videos
  • Shows
  • World
  • India
  • Explainers
  • Opinion
  • Sports
  • Cricket
  • Health
  • Tech/Auto
  • Entertainment
  • Web Stories
  • Business
  • Impact Shorts

Shows

  • Vantage
  • Firstpost America
  • Firstpost Africa
  • First Sports
  • Fast and Factual
  • Between The Lines
  • Flashback
  • Live TV

Events

  • Raisina Dialogue
  • Independence Day
  • Champions Trophy
  • Delhi Elections 2025
  • Budget 2025
  • US Elections 2024
  • Firstpost Defence Summit
Trending:
  • PM Modi in Manipur
  • Charlie Kirk killer
  • Sushila Karki
  • IND vs PAK
  • India-US ties
  • New human organ
  • Downton Abbey: The Grand Finale Movie Review
fp-logo
How US built new tool to stop AI from making nuclear weapons
Whatsapp Facebook Twitter
Whatsapp Facebook Twitter
Apple Incorporated Modi ji Justin Trudeau Trending

Sections

  • Home
  • Live TV
  • Videos
  • Shows
  • World
  • India
  • Explainers
  • Opinion
  • Sports
  • Cricket
  • Health
  • Tech/Auto
  • Entertainment
  • Web Stories
  • Business
  • Impact Shorts

Shows

  • Vantage
  • Firstpost America
  • Firstpost Africa
  • First Sports
  • Fast and Factual
  • Between The Lines
  • Flashback
  • Live TV

Events

  • Raisina Dialogue
  • Independence Day
  • Champions Trophy
  • Delhi Elections 2025
  • Budget 2025
  • US Elections 2024
  • Firstpost Defence Summit
  • Home
  • Explainers
  • How US built new tool to stop AI from making nuclear weapons

How US built new tool to stop AI from making nuclear weapons

FP Explainers • August 21, 2025, 21:01:40 IST
Whatsapp Facebook Twitter

Anthropic, an Artificial Intelligence (AI) start-up backed by Amazon and Google, has developed a new tool to stop its chatbot from being used for the nefarious purposes of building a nuclear bomb or a reactor

Advertisement
Subscribe Join Us
Add as a preferred source on Google
Prefer
Firstpost
On
Google
How US built new tool to stop AI from making nuclear weapons
Anthropic, whose AI bot Claude is a direct competitor to OpenAI's ChatGPT, said it has been working with the US government for over a year to build in the safeguard.

Today, everyone is obsessed with Artificial Intelligence (AI).

AI is said to have the potential to change society forever, in good ways and bad. Many hope it will cure humans of disease, extend our lifespans solve climate change, and unlock the secrets of the universe.

Others fear it will cause some jobs to go away forever, leaving millions out of work and society on the brink. Others imagine a dark, dystopian future with AI ruling over humanity – perhaps in the aftermath of it ordering nuclear strikes.

STORY CONTINUES BELOW THIS AD

Now, some are taking steps to at least safeguard its AI models from being used as tools to build nuclear weapons.

But what happened? What do we know?

More from Explainers
Talking to chatbots can lead to 'AI psychosis'. Is this a growing mental health risk? Talking to chatbots can lead to 'AI psychosis'. Is this a growing mental health risk? Do you have a reading list? How books are dying a slow death Do you have a reading list? How books are dying a slow death

Let’s take a closer look

What happened?

Anthropic, an AI start-up backed by Amazon and Google, has developed a new tool to stop its AI from being used  for the nefarious means of building a nuclear bomb. Anthropic's Claude is a direct competitor to OpenAI's ChatGPT.

Anthropic said it has been working with the US government for over a year to build in the safeguard. The company said it has coordinated with the National Nuclear Security Administration (NNSA) to figure out  a “classifier” that can halt “concerning” conversations — for example, how to build a nuclear reactor or bomb – on its AI system.

Anthropics said the program sprung out of its 2024 exercises with the US Department of Energy. The NNSA falls under the US Energy Department. It is tasked with making sure the United States ‘maintains a safe, secure, and reliable nuclear stockpile through the application of unparalleled science, technology, engineering, and manufacturing.’ The NNSA’s Office of Defence Programs is in charge of maintaining and modernising the country’s nuclear stockpile.

Editor’s Picks
1
OpenAI’s ChatGPT-5 to launch in August: What to expect
OpenAI’s ChatGPT-5 to launch in August: What to expect
2
ChatGPT’s Ghibli-style images: Who is Hayao Miyazaki, man behind OG Studio Ghibli?
ChatGPT’s Ghibli-style images: Who is Hayao Miyazaki, man behind OG Studio Ghibli?

How did it do it?

The company said it was able to put together a list of gauges that can help Claude identify “potentially concerning conversations about nuclear weapons development”.

The classifier acts like a spam filter in the email and identifies real-time threats. The company has claimed that the classifier can determine with almost 95 per cent accuracy if the person carrying on the conversation with the AI bot is intending to cause harm. The company said the classifier identified 94.8 per cent of nuclear weapons queries. However, it inaccurately classified 5.2 per cent of the queries as dangerous.

STORY CONTINUES BELOW THIS AD
The company said it was able to put together a list of gauges that can help Claude identify
The company said it was able to put together a list of gauges that can help Claude identify “potentially concerning conversations about nuclear weapons development”.

Anthropic has said that it has already employed the classifier in some of its Claude models.

“As AI models become more capable, we need to keep a close eye on whether they can provide users with dangerous technical knowledge in ways that could threaten national security,” Anthropic has said.

The company has vowed to share what it has learnt with the Frontier Model Forum, an AI industry body it has co-founded alongside Amazon, Meta, OpenAI, Microsoft and Google , in order to help other companies build similar programmes.

Anthropic earlier in August said it would offer its Claude AI model to the US government for $1 (Rs 87), joining the ranks of AI start-ups proposing lucrative deals to win federal contracts.

This came days after OpenAI's ChatGPT, Google’s Gemini and Anthropic’s Claude were added to the US government’s list of approved AI vendors.

“America’s AI leadership requires that our government institutions have access to the most capable, secure AI tools available,” CEO Dario Amodei said.

STORY CONTINUES BELOW THIS AD

Rival OpenAI had also announced a similar offer in August, wherein ChatGPT Enterprise was made available to participating US federal agencies for $1 per agency for the next year.

With inputs from agencies

Tags
artificial intelligence (AI) ChatGPT
End of Article
Latest News
Find us on YouTube
Subscribe
End of Article

Impact Shorts

Ghaziabad woman dead, pilgrims attacked in bus… How Nepal’s Gen-Z protests turned into a living hell for Indian tourists

Ghaziabad woman dead, pilgrims attacked in bus… How Nepal’s Gen-Z protests turned into a living hell for Indian tourists

Prime Minister KP Sharma Oli resigned following violent protests in Nepal. An Indian woman from Ghaziabad died trying to escape a hotel fire set by protesters. Indian tourists faced attacks and disruptions, with some stranded at the Nepal-China border during the unrest.

More Impact Shorts

Top Stories

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Top Shows

Vantage Firstpost America Firstpost Africa First Sports
Latest News About Firstpost
Most Searched Categories
  • Web Stories
  • World
  • India
  • Explainers
  • Opinion
  • Sports
  • Cricket
  • Tech/Auto
  • Entertainment
  • IPL 2025
NETWORK18 SITES
  • News18
  • Money Control
  • CNBC TV18
  • Forbes India
  • Advertise with us
  • Sitemap
Firstpost Logo

is on YouTube

Subscribe Now

Copyright @ 2024. Firstpost - All Rights Reserved

About Us Contact Us Privacy Policy Cookie Policy Terms Of Use
Home Video Shorts Live TV