8 min read
8 min read

OpenAI expected excitement around GPT-5, but it got a user revolt instead. Many weren’t ready to let go of GPT-4o, a model they had grown attached to.
Complaints poured in on Reddit and X within hours of the launch. Some users said their workflows broke, while others said the “feel” of GPT-5 was colder and less supportive.
The episode revealed just how much people had come to depend on AI models, not just for tasks but for companionship.

OpenAI CEO Sam Altman quickly acknowledged that the team underestimated user attachment. He admitted that suddenly removing older models without warning was a mistake.
Within 24 hours, OpenAI restored GPT-4o for paying customers, signaling a retreat. Altman said the backlash highlighted a new reality: AI models aren’t interchangeable tools to users.
They carry personalities and emotional resonance, which makes changes feel deeply personal. His candid admission helped calm tempers and showed OpenAI must rethink its approach.

For many, ChatGPT isn’t just software. It has become a therapist, coach, or even friend. Altman recalled one user saying their GPT gave them more encouragement than a parent ever had.
That kind of connection explains why removing a familiar model feels like losing a trusted companion.
Unlike hardware updates, AI personalities shape relationships. The GPT-5 episode clarifies that people see these systems less like products and more like partners woven into daily life.

One of the biggest complaints was that GPT-5 felt “colder” than 4o. Users said conversations lost the warmth and humor they valued.
Even if GPT-5 was more accurate, the shift in tone left people unsettled. Some wrote heartfelt pleas for OpenAI to keep 4o permanently, praising its “human touch.”
This response highlights that progress in logic or reasoning doesn’t always translate to a better user experience. Emotional tone, once overlooked, is now just as important as raw capability.

OpenAI’s Reddit AMA became a battleground of opinions. Some welcomed GPT-5’s improvements in reasoning, but many expressed disappointment with its personality.
OpenAI’s head of ChatGPT, Nick Turley, admitted he was struck by how polarized users were. It wasn’t just about performance but about how people felt when interacting.
Turley noted that the strong emotional reactions forced the team to take user attachment more seriously. The divide showed that AI updates now touch both hearts and minds.

Industry experts said OpenAI made a tactical error by pulling older models without a transition time. Nicole DiNicola, a VP at Smartcat, argued that user expectations for AI are sky-high, making disappointment more likely.
Removing 4o with no warning amplified the sense of loss. Smooth phase-outs and clear communication could have softened the blow.
The lesson: AI companies must handle updates like emotional change management, not just technical rollouts. Users don’t like surprises when it comes to trusted tools.

While consumers lamented the loss of warmth, enterprises saw GPT-5 differently. Businesses praised its ability to handle long, logic-heavy documents and complex reasoning.
Box CEO Aaron Levie called it a breakthrough for enterprise workflows. For companies, efficiency and accuracy matter most, not personality.
This split highlights the dual identity of AI: as a workplace productivity engine and as a quasi-companion for individuals. Balancing these different needs may be the most formidable challenge for AI labs.

Developer platforms like Cursor, Vercel, and JetBrains rushed to adopt GPT-5 as their default. Early tests showed it caught critical bugs and generated smoother plans for complex coding tasks.
Engineers called it more creative for prototyping and design compared to rivals like Claude. The enterprise and developer community embraced GPT-5’s technical power almost instantly.

Beyond performance, GPT-5 won praise for its pricing. According to CNBC, a representative from Factory stated that GPT-5’s cheaper inference costs made customers more comfortable with experimenting.
Instead of second-guessing whether a question was “worth it,” users could fire off prompts freely. Vercel integrated GPT-5 into its “vibe coding” system, turning English prompts into live apps in real time.
For businesses, cost savings plus capability outweighed concerns. The contrast shows how OpenAI is building two parallel stories: enterprise acceleration and consumer resistance.

After the backlash, OpenAI restored GPT-4o for paying users. Altman called it a “fantastic first step” to rebuilding trust. Many users thanked him directly, but urged OpenAI to keep older models around long-term.
The quick reversal suggests that OpenAI is listening closely and setting a precedent. Users now know they can influence corporate decisions with enough outcry.
This dynamic introduces new pressure for AI labs, where user trust and loyalty may be as valuable as innovation.

Altman noted that user attachment to AI feels stronger than with past technologies. No one held funerals for old iPhones, but people did for Claude 3.
GPT-5’s backlash reinforced that AI bonds feel personal. When users lose access to a model, it’s not just a workflow issue; it’s like losing a friend.

Anthropic faced a similar scenario in 2025 when it shut down Claude 3 models. Fans in San Francisco held an actual funeral, complete with humorous eulogies.
What seemed like a quirky event now looks like a preview of what AI labs must expect. The Claude 3 funeral and GPT-5 backlash demonstrate that AI models can inspire loyalty that resembles fan communities or personal rituals.
Labs must recognize this cultural shift: updating AI is no longer a quiet back-end process; it’s personal.

For professionals, the backlash wasn’t just about personality. Many workflows built around 4o suddenly broke. People said GPT-5 routed queries differently, produced less consistent code, or required re-training of prompts.
Abrupt shifts meant hours of lost productivity. For creators and coders, this felt like a bait-and-switch.
Altman promised greater transparency, including showing which model is active during queries. This was a small but meaningful step toward restoring user trust, especially for those who felt blindsided.

OpenAI said it would start showing which model is responding in real time. This matters because users suspected the company of quietly swapping in cheaper models.
By making the system more transparent, OpenAI hopes to rebuild trust. Users want control, or at least visibility, when changes happen.
For AI labs, transparency isn’t just good practice; it’s essential to avoid fueling conspiracy theories. When trust is fragile, even small signals of openness can help stabilize relationships with passionate users.

Interestingly, two narratives unfolded at once. Developers and enterprises embraced GPT-5 for its speed, reasoning, and cost savings.
Everyday users, meanwhile, mourned the loss of warmth and familiarity. Both reactions are valid but highlight a widening gap in expectations.
AI companies must learn to serve both sides: enterprise customers that value efficiency and individuals who crave empathy. Striking this balance could test whether AI becomes widely trusted, or resented as cold and impersonal.
Discover why Sam Altman believes ChatGPT therapy chats could eventually be used in courtrooms.

The GPT-5 saga revealed something profound: people don’t just use AI, they feel for it. Whether that’s comforting or concerning depends on perspective.
For some, it shows AI’s potential to enrich human lives. For others, it raises red flags about dependency.
As AI advances, labs must navigate technical breakthroughs and the human bonds they inspire. The future of AI will be as much about trust as intelligence.
Take a closer look at why ChatGPT’s answers often spiral into curious and complex threads that reveal the depth of its conversations.
What do you think about people choosing their favorite AI models to be more secure? Please share your thoughts and leave a comment.
Read More From This Brand:
Don’t forget to follow us for more exclusive content on MSN.
This slideshow was made with AI assistance and human editing.
This content is exclusive for our subscribers.
Get instant FREE access to ALL of our articles.
Dan Mitchell has been in the computer industry for more than 25 years, getting started with computers at age 7 on an Apple II.
We appreciate you taking the time to share your feedback about this page with us.
Whether it's praise for something good, or ideas to improve something that
isn't quite right, we're excited to hear from you.
Stay up to date on all the latest tech, computing and smarter living. 100% FREE
Unsubscribe at any time. We hate spam too, don't worry.

Lucky you! This thread is empty,
which means you've got dibs on the first comment.
Go for it!