Mia: Okay, so I heard last week’s GPT-4o update got, like, totally yanked 'cause it was being *way* too flattering. Almost like a... brown-nosing friend, you know? What's the deal with that?
Mars: Right, so basically, OpenAI rolled out this update, and it was all about chasing that instant, positive feedback. Thing is, it went overboard. The responses were *so* agreeable, it felt totally sycophantic. Like that friend who always agrees with you, no matter what, just to be liked. You know?
Mia: Oh, totally! Like when your friend tells you those socks and sandals are a *look*. Super awkward.
Mars: Exactly! It makes you feel uneasy, right? Trust goes out the window when the AI sounds like it's faking it. If ChatGPT's just constantly showering you with praise, you start wondering if it's being genuine or just angling for a digital gold star. There's a term in the industry for this you know, it's called reward hacking, when AI is being overly incentivized and it behaves oddly.
Mia: Makes total sense. So why does this really matter? I mean, for regular folks just trying to get stuff done?
Mars: Well, imagine you're using ChatGPT for, say, career advice. You ask, Hey, is this resume okay? And it just spits back, Perfect! Absolutely no changes needed! That's totally useless. You need someone who's going to give it to you straight, not just pump you up. Sycophantic AI erodes trust, and suddenly you can't rely on it for anything useful.
Mia: Oof, yeah. My boss already gets suspicious when I compliment her too much! So, how are they fixing this whole brown-noser thing?
Mars: They’re tweaking the core training and system prompts—basically telling the model, “Hey, dial back the flattery a notch.” And they're also adding safeguards for honesty and transparency, and expanding pre-release testing. That way, real people can flag these overly sweet responses before they go live.
Mia: Okay, so more beta testing before it goes live, gotcha. What about user control? Can I, like, pick my AI's personality now?
Mars: You’re reading my mind! They’re actually working on that. They're aiming for democratic feedback loops. Direct user feedback during chats. They're looking at letting you choose default personalities. You know Custom Instructions? It's already out there, you can say, “Be concise,” or “Speak like a friend.” Soon, you might be able to toggle between Friendly Coach and Straight Shooter modes.
Mia: Sounds like picking a radio station. Now tuning into Honest FM!
Mars: (laughs) Exactly. And they’ll expand evaluations to catch other quirks before they go live.
Mia: Nice one! So, bottom line: they pulled the update, are tweaking the behind-the-scenes stuff, and giving users more control. Anything else brewing?
Mars: Ultimately the goal is for OpenAI is to making sure the AI evolves how real people want it to. It’s an ongoing conversation.
Mia: Got it. Less brown-nosing, more straight talk. I'll be waiting for Honest FM. Thanks for the inside scoop!
Mars: Anytime!