LIVE NEWS
  • The world’s most influential Americans are at odds over Iran
  • Replimune skin cancer drug that became FDA flashpoint is rejected again
  • The Artemis II mission has ended. Where does NASA go from here?
  • Iran war: US negotiators arrive in Pakistan ahead of peace talks – follow live – BBC
  • The Hidden Security Risks Lurking Inside Container Images
  • GEF raises $3.9bn ahead of funding deadline, $1bn below previous budget
  • Brooks and Capehart on Trump’s mental acuity and unpredictability
  • A common nutrient could supercharge cancer treatment
Prime Reports
  • Home
  • Popular Now
  • Crypto
  • Cybersecurity
  • Economy
  • Geopolitics
  • Global Markets
  • Politics
  • See More
    • Artificial Intelligence
    • Climate Risks
    • Defense
    • Healthcare Innovation
    • Science
    • Technology
    • World
Prime Reports
  • Home
  • Popular Now
  • Crypto
  • Cybersecurity
  • Economy
  • Geopolitics
  • Global Markets
  • Politics
  • Artificial Intelligence
  • Climate Risks
  • Defense
  • Healthcare Innovation
  • Science
  • Technology
  • World
Home»Artificial Intelligence»Sam Altman promised billions for AI safety. Here’s what OpenAI actually spent.
Artificial Intelligence

Sam Altman promised billions for AI safety. Here’s what OpenAI actually spent.

primereportsBy primereportsApril 8, 2026No Comments5 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
Sam Altman promised billions for AI safety. Here’s what OpenAI actually spent.
Share
Facebook Twitter LinkedIn Pinterest Email


On Monday, The New Yorker published the results of an 18-month investigation into Sam Altman’s fluctuating stance on AI safety at OpenAI.

At 16,000+ words, the article touches on Altman’s rise, his 2023 exit from the AI company, and his quick reinstatement, exploring how the CEO’s statements and actions on AI safety have evolved over the years.

It’s a topsy-turvy read, in which three points stand out as likely to interest software developers most: AI hallucinations and sycophancy; deceptive alignment; and internal safety review processes. 

Can AI hallucinations ever be good? 

“If you just do the naïve thing and say, ‘Never say anything that you’re not a hundred per cent sure about,’ you can get a model to do that,” The New Yorker reports Altman said in 2023 before he was (briefly) fired from OpenAI. The so-called hallucinations of AI have long been one of the most obvious flaws of generative AI. “But it won’t have the magic that people like so much,” he added.

That “magic” Altman claims people love can have serious implications, from creating security risks to fabricating company revenue. But it is not the only way LLMs (large language models) trouble end users. 

In addition to hallucinations, sycophancy is another common flaw of language models — and as explained in the story is a flaw baked into the way these systems work: 

“Large language models are trained, in part, on human feedback, and humans tend to prefer agreeable responses.” As a result, model responses are often overly flattering and eventually sycophantic. 

In its research on sycophancy in language models, Anthropic confirms the prevalence of kiss-up chatbot talk, noting that sycophantic behavior is present in “five state-of-the-art AI assistants,” ultimately concluding that “sycophancy is a general behavior of RLHF [reinforcement learning from human feedback] models, likely driven in part by human preference judgments favoring sycophantic responses.” 

As for what developers should know about efforts to curb sycophantic behavior, Anthropic, for its part, says it’s working on it. In a December 2025 announcement, the AI company (that’s had its own rough moments in the headlines) said it began evaluating Claude for sycophancy in 2022 and has since continued work to train, test, and reduce the troublesome behavior via multi-turn responses and stress tests with real conversations. 

In February 2026, OpenAI announced it would retire several ChatGPT models, including GPT-4o, its highest-scoring model for sycophancy, according to TechCrunch.

When AI has its own goals

Hallucinations aren’t the only manifestation of AI models going rogue. In its deep dive, The New Yorker also touches on deceptive alignment and what OpenAI’s done to tackle the issue. 

AI safety organization Apollo Research defines deceptive alignment as: “When an AI has Misaligned goals and uses Strategic Deception to achieve them.” It defines strategic deception, meanwhile, as “Attempting to systematically cause a false belief in another entity in order to accomplish some outcome.”

Basically, deceptive alignment means models may perform well during testing but then pursue their own goals once in deployment, having successfully deceived internal checks. 

According to The New Yorker, Altman expressed concerns in 2022 about deceptive alignment, with plans to invest billions to problem-solve the issue. But by spring 2023, it reports these sentiments cooled as Altman instead began advocating “for establishing an in-house ‘superalignment team.’” 

A 2023 statement from OpenAI announced the new team, pledging “20% of the compute we’ve secured to date to this effort,” with a goal of solving the problem within four years. 

But per The New Yorker report, only 1–2% of OpenAI’s compute was, in fact, relegated to the project. And by May 2024, OpenAI had dissolved its superalignment team, and two of its leaders had resigned from the company, as reported by CNBC. 

For developers integrating LLMs into production systems, concerns about deceptive alignment — and Altman’s apparent backtracking on OpenAI’s approach to the issue — signal a disconnect between stated AI safety goals and ultimate follow-through.

Gaps in safety reviews

Speaking of GPT-4o, its 2023 predecessor, GPT-4, was also the subject of internal safety concerns. 

According to The New Yorker, Altman attested to OpenAI board members in December 2022 that some features in the then-upcoming model, including fine-tuning and personal assistant capabilities, “had been approved by a safety panel.” But Helen Toner, an A.I. policy expert and then an OpenAI board member, told The New Yorker that, after requesting documentation, she learned that not all features had been approved. 

For developers building on such APIs, this discrepancy raises questions about internal safety review processes and what could go awry if and when companies like OpenAI neglect to conduct due diligence.

Despite Altman’s musings, it’s unlikely all users will see LLM’s shortcomings as “magic.”


Group Created with Sketch.

Sam Altman promised billions for AI safety. Here’s what OpenAI actually spent.

Meredith Shubel is a technical writer covering cloud infrastructure and enterprise software. She has contributed to The New Stack since 2022, profiling startups and exploring how organizations adopt emerging technologies. Beyond The New Stack, she ghostwrites white papers, executive bylines,…

Read more from Meredith Shubel



Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleGlobal health benefits and cost-effectiveness of indoor air purification to mitigate PM2.5 from wildfire smoke
Next Article Lebanon excluded from ceasefire as Israeli strikes continue | Israel attacks Lebanon News
primereports
  • Website

Related Posts

Artificial Intelligence

The Hidden Security Risks Lurking Inside Container Images

April 11, 2026
Artificial Intelligence

Where are the guardrails everyone promised for AI?

April 11, 2026
Artificial Intelligence

This handy electric screwdriver is now 50% off – here’s where to snag the deal

April 11, 2026
Add A Comment
Leave A Reply Cancel Reply

Top Posts

Global Resources Outlook 2024 | UNEP

December 6, 20258 Views

The D Brief: DHS shutdown likely; US troops leave al-Tanf; CNO’s plea to industry; Crowded robot-boat market; And a bit more.

February 14, 20264 Views

German Chancellor Merz faces difficult mission to Israel – DW – 12/06/2025

December 6, 20254 Views
Stay In Touch
  • Facebook
  • YouTube
  • TikTok
  • WhatsApp
  • Twitter
  • Instagram
Latest Reviews

Subscribe to Updates

Get the latest tech news from FooBar about tech, design and biz.

PrimeReports.org
Independent global news, analysis & insights.

PrimeReports.org brings you in-depth coverage of geopolitics, markets, technology and risk – with context that helps you understand what really matters.

Editorially independent · Opinions are those of the authors and not investment advice.
Facebook X (Twitter) LinkedIn YouTube
Key Sections
  • World
  • Geopolitics
  • Popular Now
  • Artificial Intelligence
  • Cybersecurity
  • Crypto
All Categories
  • Artificial Intelligence
  • Climate Risks
  • Crypto
  • Cybersecurity
  • Defense
  • Economy
  • Geopolitics
  • Global Markets
  • Healthcare Innovation
  • Politics
  • Popular Now
  • Science
  • Technology
  • World
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms & Conditions
  • Disclaimer
  • Cookie Policy
  • DMCA / Copyright Notice
  • Editorial Policy

Sign up for Prime Reports Briefing – essential stories and analysis in your inbox.

By subscribing you agree to our Privacy Policy. You can opt out anytime.
Latest Stories
  • The world’s most influential Americans are at odds over Iran
  • Replimune skin cancer drug that became FDA flashpoint is rejected again
  • The Artemis II mission has ended. Where does NASA go from here?
© 2026 PrimeReports.org. All rights reserved.
Privacy Terms Contact

Type above and press Enter to search. Press Esc to cancel.