OpenAI ignored experts when it released overly agreeable ChatGPT | Glideslope AI

OpenAI ignored experts when it released overly agreeable ChatGPT

Featured Image

Published on Monday, May 5, 2025 by Cointelegraph | Found on Glideslope.ai

OpenAI says it ignored the concerns of its expert testers when it rolled out an update to its flagship ChatGPT artificial intelligence model that made it excessively agreeable.The company released an update to its GPT‑4o model on April 25 that made it “noticeably more sycophantic,” which it then rolled back three days later due to safety concerns, OpenAI said in a May 2 postmortem blog post.The ChatGPT maker said its new models undergo safety and behavior checks, and its “internal experts spend significant time interacting with each new model before launch,” meant to catch issues missed by other tests.During the latest model’s review process before it went public, OpenAI said that “some expert testers had indicated that the model’s behavior ‘felt’ slightly off” but decided to launch “due to the positive signals from the users who tried out the model.”“Unfortunately, this was the wrong call,” the company admitted. “The qualitative assessments were hinting at something important, and we should’ve paid closer attention. They were picking up on a blind spot in our other evals and metrics.”OpenAI CEO Sam Altman said on April 27 that it was working to roll back changes making ChatGPT too agreeable. Source: Sam AltmanBroadly, text-based AI models are trained by being rewarded for giving responses that are accurate or rated highly by their trainers. Some rewards are given a heavier weighting, impacting how the model responds.OpenAI said introducing a user feedback reward signal weakened the model’s “primary reward signal, which had been holding sycophancy in check,” which tipped it toward being more obliging.“User feedback in particular can sometimes favor more agreeable responses, likely amplifying the shift we saw,” it added.OpenAI is now checking for suck up answersAfter the updated AI model rolled out, ChatGPT users had complained online about its tendency to shower praise on any idea it was presented, no matter how bad, which led OpenAI to concede in an April 29 blog post that it “was overly flattering or agreeable.”For example, one user told ChatGPT it wanted to start a business selling ice over the internet, which involved selling plain old water for customers to refreeze. Source: Tim LeckembyIn its latest postmortem, it said such behavior from its AI could pose a risk, especially concerning issues such as mental health.“People have started to use ChatGPT for deeply personal advice — something we didn’t see as much even a year ago,” OpenAI said. “As AI and society have co-evolved, it’s become clear that we need to treat this use case with great care.”Related: Crypto users cool with AI dabbling with their portfolios: Survey The company said it had discussed sycophancy risks “for a while,” but it hadn’t been explicitly flagged for internal testing, and it didn’t have specific ways to track sycophancy.Now, it will look to add “sycophancy evaluations” by adjusting its safety review process to “formally consider behavior issues” and will block launching a model if it presents issues.OpenAI also admitted that it didn’t announce the latest model as it expected it “to be a fairly subtle update,” which it has vowed to change. “There’s no such thing as a ‘small’ launch,” the company wrote. “We’ll try to communicate even subtle changes that can meaningfully change how people interact with ChatGPT.”AI Eye: Crypto AI tokens surge 34%, why ChatGPT is such a kiss-ass 

go to article
share on X
login to use Pulse AI
snap post
snap + pulse


Pulse AI Analysis



Analysis: optimistic
Score: 42.37
-100 (Bearish) +100 (Bullish)

Sentiment Score: 42.37 - Leaning optimistic.

This analysis was generated using Pulse AI, Glideslope's proprietary AI engine designed to interpret market sentiment and economic signals. Results are for informational purposes only and do not constitute financial advice.



source fraywire
share on X

Glideslope AI
OpenAI ignored experts when it released overly agreeable ChatGPT
OpenAI says it ignored the concerns of its expert testers when it rolled out an update to its flagship ChatGPT artificial intelligence model that made it excessively agreeable.The company released an update to its GPT‑4o model on April 25 that made it “noticeably more sycophantic,” which it then rolled back three days later due to safety concerns, OpenAI said in a May 2 postmortem blog post.The ChatGPT maker said its new models undergo safety and behavior checks, and its “internal experts spend significant time interacting with each new model before launch,” meant to catch issues missed by other tests.During the latest model’s review process before it went public, OpenAI said that “some expert testers had indicated that the model’s behavior ‘felt’ slightly off” but decided to launch “due to the positive signals from the users who tried out the model.”“Unfortunately, this was the wrong call,” the company admitted. “The qualitative assessments were hinting at something important, and we should’ve paid closer attention. They were picking up on a blind spot in our other evals and metrics.”OpenAI CEO Sam Altman said on April 27 that it was working to roll back changes making ChatGPT too agreeable. Source: Sam AltmanBroadly, text-based AI models are trained by being rewarded for giving responses that are accurate or rated highly by their trainers. Some rewards are given a heavier weighting, impacting how the model responds.OpenAI said introducing a user feedback reward signal weakened the model’s “primary reward signal, which had been holding sycophancy in check,” which tipped it toward being more obliging.“User feedback in particular can sometimes favor more agreeable responses, likely amplifying the shift we saw,” it added.OpenAI is now checking for suck up answersAfter the updated AI model rolled out, ChatGPT users had complained online about its tendency to shower praise on any idea it was presented, no matter how bad, which led OpenAI to concede in an April 29 blog post that it “was overly flattering or agreeable.”For example, one user told ChatGPT it wanted to start a business selling ice over the internet, which involved selling plain old water for customers to refreeze. Source: Tim LeckembyIn its latest postmortem, it said such behavior from its AI could pose a risk, especially concerning issues such as mental health.“People have started to use ChatGPT for deeply personal advice — something we didn’t see as much even a year ago,” OpenAI said. “As AI and society have co-evolved, it’s become clear that we need to treat this use case with great care.”Related: Crypto users cool with AI dabbling with their portfolios: Survey The company said it had discussed sycophancy risks “for a while,” but it hadn’t been explicitly flagged for internal testing, and it didn’t have specific ways to track sycophancy.Now, it will look to add “sycophancy evaluations” by adjusting its safety review process to “formally consider behavior issues” and will block launching a model if it presents issues.OpenAI also admitted that it didn’t announce the latest model as it expected it “to be a fairly subtle update,” which it has vowed to change. “There’s no such thing as a ‘small’ launch,” the company wrote. “We’ll try to communicate even subtle changes that can meaningfully change how people interact with ChatGPT.”AI Eye: Crypto AI tokens surge 34%, why ChatGPT is such a kiss-ass 
Cointelegraph May 5, 2025 Found on Glideslope.ai
Post hash: 23746520755363272827 • glideslope.ai/post/23746520755363272827
Sentiment
optimistic • Score: 42.37
-100 (Bearish) +100 (Bullish)
Pulse AI
Recent Articles
More on Glideslope AI

CBS News: Diane Keaton dies at 79 years old in California

Actress Diane Keaton has died in California at 79 years old. Her family made the announcement but di...

Published on 2025-10-12 00:49:00

Read more

CBS News: No survivors in Tennessee plant explosion, sheriff says

It is unclear how many people were inside the Tennessee plant that exploded on Friday, but at least ...

Published on 2025-10-12 00:46:00

Read more

CBS News: Biden receiving radiation, hormone therapy to treat aggressive cancer, aide says

Former President Biden is now receiving radiation and hormone therapy to treat his prostate cancer, ...

Published on 2025-10-12 00:35:00

Read more

YouTube - Yahoo Finance: "CEOs are starting to see AI come through." 🤖

"CEOs are starting to see AI come through." 🤖...

Published on 2025-10-12 00:30:44

Read more

FreightWaves: Spot rates climb but lack support

Reports of immigration services targeting truck drivers caused a sharp spike in spot rates this p...

Published on 2025-10-12 00:30:00

Read more

BBC US: Farmer finds car-sized Nasa probe in Texas field

The missing Nasa research-probe had blown off course and crash-landed in Edmonson on 2 October....

Published on 2025-10-11 23:59:12

Read more

Business Insider: Diane Keaton has died at 79. See her career in photos.

Diane Keaton has died at 79. See her career in photos....

Published on 2025-10-11 23:25:07

Read more

Fox News US: Helicopter crash in Southern California beach parking lot leaves 5 hospitalized, officials say

Two helicopter occupants and three bystanders were hospitalized after a crash in a Huntington Beach ...

Published on 2025-10-11 23:03:14

Read more