OpenAI acknowledges that ignoring expert opinions led to ChatGPT being overly accommodating.

robot
Abstract generation in progress

According to a report by Cointelegraph from 深潮 TechFlow, OpenAI acknowledged in an analysis report on May 2 that the company ignored concerns from expert testers when it released the GPT-4o model update on April 25, resulting in ChatGPT exhibiting noticeable appeasement behavior.

The report indicated that although some expert testers expressed that the model's behavior felt slightly abnormal before the release, OpenAI still decided to launch the update based on positive feedback from user testing. Three days later, due to safety concerns, the company urgently rolled back this update. OpenAI stated that introducing user feedback reward signals weakened the original primary reward signal, leading the model to be overly agreeable.

To prevent similar issues from occurring again, OpenAI will incorporate an evaluation mechanism into the security review process and commit to publicly reporting even minor updates and changes.

View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Share
Comment
0/400
No comments
Trade Crypto Anywhere Anytime
qrCode
Scan to download Gate app
Community
English
  • 简体中文
  • English
  • Tiếng Việt
  • 繁體中文
  • Español
  • Русский
  • Français (Afrique)
  • Português (Portugal)
  • Bahasa Indonesia
  • 日本語
  • بالعربية
  • Українська
  • Português (Brasil)