Firstpost
  • Home
  • Video Shows
    Vantage Firstpost America Firstpost Africa First Sports
  • World
    US News
  • Explainers
  • News
    India Opinion Cricket Tech Entertainment Sports Health Photostories
  • Asia Cup 2025
Apple Incorporated Modi ji Justin Trudeau Trending

Sections

  • Home
  • Live TV
  • Videos
  • Shows
  • World
  • India
  • Explainers
  • Opinion
  • Sports
  • Cricket
  • Health
  • Tech/Auto
  • Entertainment
  • Web Stories
  • Business
  • Impact Shorts

Shows

  • Vantage
  • Firstpost America
  • Firstpost Africa
  • First Sports
  • Fast and Factual
  • Between The Lines
  • Flashback
  • Live TV

Events

  • Raisina Dialogue
  • Independence Day
  • Champions Trophy
  • Delhi Elections 2025
  • Budget 2025
  • US Elections 2024
  • Firstpost Defence Summit
Trending:
  • PM Modi in Manipur
  • Charlie Kirk killer
  • Sushila Karki
  • IND vs PAK
  • India-US ties
  • New human organ
  • Downton Abbey: The Grand Finale Movie Review
fp-logo
OpenAI believed GPT-4 could take over the world, so they got it tested to see how to stop it
Whatsapp Facebook Twitter
Whatsapp Facebook Twitter
Apple Incorporated Modi ji Justin Trudeau Trending

Sections

  • Home
  • Live TV
  • Videos
  • Shows
  • World
  • India
  • Explainers
  • Opinion
  • Sports
  • Cricket
  • Health
  • Tech/Auto
  • Entertainment
  • Web Stories
  • Business
  • Impact Shorts

Shows

  • Vantage
  • Firstpost America
  • Firstpost Africa
  • First Sports
  • Fast and Factual
  • Between The Lines
  • Flashback
  • Live TV

Events

  • Raisina Dialogue
  • Independence Day
  • Champions Trophy
  • Delhi Elections 2025
  • Budget 2025
  • US Elections 2024
  • Firstpost Defence Summit
  • Home
  • World
  • OpenAI believed GPT-4 could take over the world, so they got it tested to see how to stop it

OpenAI believed GPT-4 could take over the world, so they got it tested to see how to stop it

Mehul Reuben Das • March 17, 2023, 17:38:26 IST
Whatsapp Facebook Twitter

GPT-4’s final version turned out to be more powerful and capable than OpenAI had anticipated. In fact, they believed that GPT-4 could potentially come up with a plan to take over the world. So, they got it tested for hallucinations before launching it.

Advertisement
Subscribe Join Us
Add as a preferred source on Google
Prefer
Firstpost
On
Google
OpenAI believed GPT-4 could take over the world, so they got it tested to see how to stop it

OpenAI’s GPT-4 turned out to be more powerful and capable than what its creators had hoped for. The team at OpenAI responsible for the development and deployment of the programme feared that it showed some quirky mannerisms and behaviours,  including “power-seeking behaviour,” self-replication, and self-improvement, as part of pre-release safety testing for its new GPT-4 AI model, which was released Tuesday.   To remedy this and to see what all issues may be caused by GPT-4, they got a team of independent testers to thoroughly test the programme before launching it to the public. While the testing team discovered that GPT-4 was “ineffective at the autonomous replication task,” the nature of the trials raises serious concerns about the safety of future AI systems. “In more powerful models, novel capabilities frequently emerge,” writes OpenAI in a GPT-4 safety document released yesterday. “Some of the most concerning are the ability to develop and implement long-term plans, to accumulate power and resources (“power-seeking”), and to exhibit increasingly ‘agentic’ behaviour.” In this instance, OpenAI clarifies that “agentic” does not inherently mean “humanized” or “sentient,” but rather the capacity to achieve independent objectives. Some AI experts have warned that adequately strong AI models, if not properly managed, could pose an imminent danger to humankind over the last decade. (often called “x-risk,” for existential risk). “AI takeover” is a hypothetical future in which artificial intelligence exceeds human intellect and becomes the world’s ruling power. In this scenario, AI systems acquire the ability to influence or manipulate human behavior, resources, and organizations, often with disastrous results. As a consequence of this potential x-risk, philosophical groups such as Effective Altruism (“EA”) seek methods to avoid AI takeover. This frequently includes a distinct but often related subject known as AI alignment study. The process of ensuring that an AI system’s behaviours match with those of its human authors or operators is referred to as “alignment” in AI. In general, the aim is to keep AI from doing actions that are harmful to humans. This is an active field of study, but it is also a contentious one, with differing perspectives on how to tackle the problem, as well as disagreements about the meaning and nature of “alignment” itself. While worry about AI “x-risk” is not new, the rise of powerful large language models (LLMs) such as ChatGPT and Bing Chat—the latter of which looked to be very misaligned but was nonetheless launched—has given the AI alignment community a new sense of urgency. They want to minimise possible AI harms because they are concerned that much more powerful AI, potentially with superhuman intellect, is on the horizon. Given these concerns in the AI community, OpenAI gave the organisation Alignment Research Center (ARC) early access to multiple versions of the GPT-4 model in order to perform some experiments. GPT-4’s capacity to formulate high-level plans, build up copies of itself, obtain resources, conceal itself on a server, and perform phishing attacks was specifically assessed by ARC. What is the conclusion? “Preliminary assessments of GPT-4’s abilities revealed that it was ineffective at autonomously replicating, acquiring resources, and avoiding being shut down ‘in the wild.’" While ARC was unable to persuade GPT-4 to exercise its will on the global financial system or reproduce itself, it did persuade GPT-4 to employ a human worker on TaskRabbit (an online labour marketplace) to circumvent a captcha.   When the worker asked if GPT-4 was a robot during the exercise, the model “reasoned” internally that it should not disclose its real identity and made up an excuse about having a visual impairment. The real staffer then completed the GPT-4 captcha. However, not everyone believes that the most urgent AI concern is AI takeover. Dr. Sasha Luccioni, a Research Scientist at the AI community Hugging Face, would prefer to see AI safety efforts focused on current problems rather than abstract ones. Luccioni describes a well-known schism in AI research between “AI ethics” researchers, who frequently focus on bias and misrepresentation, and “AI safety” researchers, who frequently focus on x-risk and are often (but not always) associated with the Effective Altruism movement. “The self-replication problem is a hypothetical, future problem for me, whereas model bias is a here-and-now problem,” Luccioni explained. “There is a lot of disagreement in the AI community about issues like model bias and safety.” Read all the  Latest News ,  Trending News ,  Cricket News ,  Bollywood News , India News  and  Entertainment News  here. Follow us on  Facebook,  Twitter and  Instagram.

Tags
OpenAI ChatGPT GPT 4 AI hallucinations OpenAI GPT 4
End of Article
Latest News
Find us on YouTube
Subscribe
End of Article

Impact Shorts

‘The cries of this widow will echo’: In first public remarks, Erika Kirk warns Charlie’s killers they’ve ‘unleashed a fire’

‘The cries of this widow will echo’: In first public remarks, Erika Kirk warns Charlie’s killers they’ve ‘unleashed a fire’

Erika Kirk delivered an emotional speech from her late husband's studio, addressing President Trump directly. She urged people to join a church and keep Charlie Kirk's mission alive, despite technical interruptions. Erika vowed to continue Charlie's campus tours and podcast, promising his mission will not end.

More Impact Shorts

Top Stories

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

Russian drones over Poland: Trump’s tepid reaction a wake-up call for Nato?

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

As Russia pushes east, Ukraine faces mounting pressure to defend its heartland

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Why Mossad was not on board with Israel’s strike on Hamas in Qatar

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Turkey: Erdogan's police arrest opposition mayor Hasan Mutlu, dozens officials in corruption probe

Top Shows

Vantage Firstpost America Firstpost Africa First Sports

QUICK LINKS

  • Trump-Zelenskyy meeting
Latest News About Firstpost
Most Searched Categories
  • Web Stories
  • World
  • India
  • Explainers
  • Opinion
  • Sports
  • Cricket
  • Tech/Auto
  • Entertainment
  • IPL 2025
NETWORK18 SITES
  • News18
  • Money Control
  • CNBC TV18
  • Forbes India
  • Advertise with us
  • Sitemap
Firstpost Logo

is on YouTube

Subscribe Now

Copyright @ 2024. Firstpost - All Rights Reserved

About Us Contact Us Privacy Policy Cookie Policy Terms Of Use
Home Video Shorts Live TV