Posted by u/RobertR7
GPT-5.2 feels less like a tool and more like a patronizing hall monitor
I don’t know who asked for this version of ChatGPT, but it definitely wasn’t the people actually using it. Every time I open a new chat now, it feels like I’m talking to a corporate therapist with a script instead of an assistant. I ask a simple question and get: “Alright. Pause. I hear you. I’m going to be very clear and grounded here.” Cool man, I just wanted help with a task, not a TED Talk about my feelings. Then there’s 5.2 itself. Half the time it argues more than it delivers. People are literally showing side-by-side comparisons where Gemini just pulls the data, runs the math, and gives an answer, while GPT-5.2 spends paragraphs “locking in parameters,” then pivots into excuses about why it suddenly can’t do what it just claimed it would do. And when you call it out, it starts defending the design decision like a PR intern instead of just fixing the mistake. On top of that, you get randomly rerouted from 4.1 (which a lot of us actually like) into 5.2 with no control. The tone changes, the answers get shorter or weirder, it ignores “stop generating,” and the whole thing feels like you’re fighting the product instead of working with it. People are literally refreshing chats 10 times just to dodge 5.2 and get back to 4.1. How is that a sane default experience? And then there’s the “vibe memory” nonsense. When the model starts confidently hallucinating basic, easily verifiable facts and then hand-waves it as some kind of fuzzy memory mode, that doesn’t sound like safety. It just sounds like they broke reliability and slapped a cute label on it. What sucks is that none of this is happening in a vacuum. Folks are cancelling Plus, trying Claude and Gemini, and realizing that “not lecturing, not arguing, just doing the task” is apparently a premium feature now. Meanwhile OpenAI leans harder into guardrails, tone management and weird pseudo-emotional framing while the actual day-to-day usability gets worse. If the goal was to make the model feel “safer” and more “aligned,” congrats, it now feels like talking to an overprotective HR chatbot that doesn’t trust you, doesn’t trust itself, and still hallucinates anyway. At some point they have to decide if this is supposed to be a useful tool for adults, or a padded room with an attitude. Right now it feels way too much like the second one.
More from r/OpenAI
The end of GPT
How the hell can a screen protector be optimized for AI?💀
Time to cancel ChatGPT Plus after three Years. Anthropic got nuked for having ethics, and Sam Altman instantly swooped in for the Pentagon bag.
The body wasn't even cold before OpenAI signed a deal to deploy on classified Department of War networks. And the...