15 Comments
User's avatar
Sean Kucer's avatar

Thank you for your write-up on this!

Joshua Blake's avatar

> As a data point, GPT-5 confirmed all the pre-existing trends, suggesting that AI progress is moving exactly as fast as I thought before (median AGI arrival date of 2033).

In general, being on trend shouldn't update your mean time to arrival but reduce the variance around it. Mean time to an event is almost always greater than median but reducing variance will push them closer together. Therefore, being on trend would normally increase your median arrival time but leave the mean unchanged. This update might be very small though.

Tim's avatar

But Claude are still the best coding models, and Claude Code is still the best coding/agentic tool. If you really believe in a self improving feedback loop, why would you care about short term consumer demand? Maybe I’m missing something, but it feels like OpenAI is not banking on AGI anymore.

Peter Wildeford's avatar

short term consumer demand pays the Stargate bills! Self improving feedback loops aren't cheap!

David Dabney's avatar

not to mention (theoretically) reducing compute costs to serve its current users! Check out them per token costs, wowee

Loic's avatar

This post gave me an interesting thought;

What if Open-AI had

* Postponed the December o3 announcement to February

* Named o3 GPT-5 instead of o3

* credibly announced the progress on ARC and Frontier Math

What would our impression of the rate of AI progress be now? It seems like the decision to create the o-series was hoping that pre-training would remain the most powerful lever, but the result has been people getting impatient and claiming AI progress has been underwhelming, maybe prematurely.

Peter Wildeford's avatar

Yeah I think this would've kept the perception of progress higher.

Fergus Argyll's avatar

At some point, the best benchmark is revenue / gdp growth etc.

It bothers me that we keep finding indirect ways to measure what we really want to measure; does this change the world?

kevin's avatar

yes, but it's a lagging indicator

Matt's avatar

Yes, but it's really hard to build a benchmark that's strongly correlated with the ability to provide economic value. The METR task horizon benchmarks may be the closest.

Kshitij Parikh's avatar

By annualized revenue, do you mean revenue earned in December 2025 multiplied by 12 or the revenue earned in whole of 2025? According to this https://www.wheresyoured.at/howmuchmoney/, OpenAI + Anthropic have made around $7.5B in the first 7 months of 2025 and the ARR based on July 2025 is $12B and $5B equating to $17B.

Peter Wildeford's avatar

Yes, I mean 2025 December * 12. I have added a footnote now to clarify this. Thanks!

kevin's avatar

Great point about revenue. But I am unsure how this release impacts the chances of this paradigm reaching real general intelligence. Isn't it marginally-moderately bearish for it scaling to AGI?

Peter Wildeford's avatar

For me, GPT-5 produced exactly the benchmarks I was expecting given my projections. So I think the pace towards AGI is still on track for sometime in the 2029-2040 range (median 2033). But if you were expecting something like 'AI 2027' I agree GPT-5 looks bearish for that.