AussieBytes Blog ◖ᵔᴥᵔ◗

A reaction to the ChatGPT‑5 launch – what it means for people and organisations, and where we are on the technology S‑curve

It was hyped as a milestone weekend for AI. Instead, it turned into a case study of how people react to badly managed AI changes, and revealed a little bit more about where we are on the AI technology S-curve.

TL;DR


1. What ChatGPT launched

OpenAI’s GPT-5 landed with a few headline promises: higher reasoning ability, no model picker, multimodal input / output, and an expanded context window. In practice:

From my testing, the new models (chatgpt-5, chatgpt-5-thinking, chatgpt-5-pro) are the frontier models for most knowledge work right now, and will expedite a range of use cases, particularly in enterprise. Anthropic's models are very competitive (if not better) at specific things, but OpenAI's seem to get more right across more areas. On the coding side - it's v close between 4 Sonnet / Opus 4.1 and ChatGPT 5, and the choice will come down to specific workflows and individual preferences. Generally - I still find Opus / Sonnet better at traversing larger projects, and Claude Code (on Anthropic's Max plan) is an excellent deal and experience overall.


2. What went wrong

The launch weekend had its share of gremlins:

My experience was fairly similar in the first few hours. ChatGPT-5 in the app was colder, terser, and was failing to execute standard jobs in my workflow. However, testing it via the API was very impressive. Adjust some settings in ChatGPT app, and looking closely at the model / system cards and prompting guides revealed some interesting items. Not only does the GPT-5 family demonstrate noticeably higher reasoning ability, but according to OpenAI’s own model documentation¹ it exhibits a markedly lower hallucination rate than previous flagship models (notably GPT-4o, with reductions measured in OpenAI’s benchmark evaluations²) and delivers strong performance on instruction-following and complex, multi-step or “long-horizon” tasks³. The model required me to adjust my set-up and prompting approach, resulting in my work outputs becoming much better or delivered faster, versus the previous models. Creative writing was also excellent.


3. What the reaction really means

Beneath the noise, the weekend exposed several deeper shifts in the AI landscape:

badvibes2

Social Media early reaction - Reddit

AI Behaviour2

AI User Attachment Example

Screenshot 2025-08-11 at 11

Sam Altman Reaction on User Attachment Post Launch

chatgpt5 system prompt

*Leaked ChatGPT System Prompt:

chatgpt mdma hack ChatGPT-5 Prompt Injection Hack as posted on X here

For enterprises embedding AI into customer and employee experiences, the bar is no longer just “does it work?” — it’s “does it work flawlessly from day one across our workflows, and does it align with the emotional, ethical, and operational expectations our users now bring?” This means brand, user experience, and safety must be designed together: abrupt changes risk eroding trust; personality shifts can alter customer sentiment; and safety lapses can undermine both brand equity and compliance.


4. Where we are on the technology S‑curve

Did the good, but less than hyped, improvement by OpenAI with the launch of ChatGPT-5 herald that AI innovation is slowing down?  I think my previous post on where we are still holds up. Ultimately, we are in a period of rapid incremental change based on optimisations across compute, data, and algorithmic improvement.

We are still on the rise through the S‑curve: meaningful gains are likely over the next 6–9 months (driven by better routing, tool‑use, memory, and inference‑time optimisation), models are also set to get more capability driven by new larger data centres and model improvements.

METR Analysis

METR Analysis - Model performance on long horizon tasks

What I am keeping a close eye on is whether we see a step-change in model improvement, or alternatively a flattening of the innovation curve, as new data centers come online from XAI with their Colossus infrastructure, the Stargate program with OpenAI, and Meta's new clusters, which will all be underpinned by the latest chips.

Beyond that horizon, the trajectory is uncertain — contingent on breakthroughs in training data, model architecture, chips, cooling/power, and smarter scaling laws rather than more brute force.

Regardless of where we are on the curve, without any further improvements in generative AI, there are major benefits to stitching the AI capabilities we have today into enterprise systems and consumer experiences, with a long road of benefits realisation available already.


5. Closing thoughts..

In the end, the GPT-5 launch was less about the model itself and more about the messy, very human dynamics around it. We saw how hype can outrun reality, how small UX and safety choices ripple into public perception, and how trust is built—or lost—at the speed of a social feed.

There is also a very dark side to how people grow attached to some more sycophantic models which has ramifications for how we design AI experiences and protect the vulnerable.

Ultimately, this weekend's events were a intriguing snapshot of an industry still sprinting up the S-curve: brilliant, chaotic, and just beginning to show its true shape.


Footnotes

¹ OpenAI, “GPT‑5 Chat Latest Model Card,” https://platform.openai.com/docs/models/gpt-5-chat-latest
² OpenAI, “GPT‑5 Benchmark Results,” section on factual accuracy, https://platform.openai.com/docs/models/gpt-5-chat-latest#benchmarks
³ OpenAI, “GPT‑5 Performance Guide,” section on long‑form and multi‑step reasoning, https://platform.openai.com/docs/guides/reasoning This means that the user needs to use great prompt structures and be specific on more complex tasks.