← Back to feed
Audio mode - plays with screen off

Graphs AI Companies Would Prefer You To Misunderstand | Toby Ord, Oxford University

80,000 Hours
Open in YouTube
*To skip to the graphs go to: The 'Scaling Paradox' (01:25:12), Misleading charts from AI companies (01:34:28), Policy debates should dream much bigger — some radical suggestions (01:46:32).* The era of making AI smarter by just making it bigger is ending. But that doesn’t mean progress is slowing down — far from it. AI models continue to get much more powerful, just using very different methods. And those underlying technical changes force a big rethink of what coming years will look like. Toby Ord — Oxford philosopher and bestselling author of _The Precipice_ — has been tracking these shifts and mapping out the implications both for governments and our lives. As he explains, until recently anyone can access the best AI in the world “for less than the price of a can of Coke.” But unfortunately, that’s over. What changed? AI companies first made models smarter by throwing a million times as much computing power at them during training, to make them better at predicting the next word. But with high quality data drying up, that approach petered out in 2024. So they pivoted to something radically different: instead of training smarter models, they’re giving existing models dramatically more time to think — leading to the rise in “reasoning models” that are at the frontier today. The results are impressive but this extra computing time comes at a cost: OpenAI’s o3 reasoning model achieved stunning results on a famous AI test by writing an Encyclopedia Britannica’s worth of reasoning to solve individual problems — at a cost of over $1,000 per question. This isn’t just technical trivia: if this improvement method sticks, it will change much about how the AI revolution plays out — starting with the fact that we can expect the rich and powerful to get access to the best AI models well before the rest of us. Learn more and full transcript: https://80k.info/to25 _Recorded on May 23, 2025._ Chapters: • Cold open (00:00:00) • Toby Ord is back — for a 4th time! (00:01:20) • Everything has changed and changed again since 2020 (00:01:39) • Is x-risk up or down? (00:08:02) • The new scaling era: compute at inference... (00:09:32) • Means less concentration (00:32:40) • But rich people will get access first. And we may not even know. (00:36:34) • 'Compute governance' is now much harder (00:42:51) • 'IDA' might let AI blast past human level — or crash and burn (00:50:11) • Reinforcement learning brings back 'reward hacking' agents (01:07:32) • Will we get warning shots? (01:17:35) • The 'Scaling Paradox' (01:25:12) • Misleading charts from AI companies (01:34:28) • Policy debates should dream much bigger. Some radical suggestions. (01:46:32) • Moratoriums have worked before (01:59:55) • AI might 'go rogue' early on (02:17:38) • Lamps are regulated much more than AI (02:25:25) • Companies made a strategic error shooting down SB 1047 (02:34:35) • They should build in emergency brakes for AI (02:40:39) • Toby's bottom lines (02:49:37) *Tell us what you thought!* https://forms.gle/enUSk8HXiCrqSA9J8 _Video editing: Simon Monsour_ _Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, & Dominic Armstrong_ _Music: Ben Cordell_ _Camera operator: Jeremy Chevillotte_ _Transcriptions and web: Katy Moore_

📝 Transcript

Transcribing... This may take a few minutes.
Top Comments
@eightythousandhours
To skip to the graphs go to: The 'Scaling Paradox' (01:25:12), Misleading charts from AI companies (01:34:28), Policy debates should dream much bigger — some radical suggestions (01:46:32). All of the graphs and results we're talking about can be found on Toby's website: • The Scaling Paradox — https://www.tobyord.com/writing/the-scaling-paradox • The Precipice Revisited — http://www.tobyord.com/writing/the-precipice-revisited • Inference Scaling Reshapes AI Governance — https://www.tobyord.com/writing/inference-scaling-reshapes-ai-governance • Inference Scaling and the Log-x Chart — https://www.tobyord.com/writing/inference-scaling-and-the-log-x-chart • Is There a Half-Life for the Success Rates of AI Agents? — https://www.tobyord.com/writing/half-life There are many more links to learn more, plus a transcript and summary on our website: https://80000hours.org/podcast/episodes/toby-ord-inference-scaling-ai-governance/
40 likes
@0xCAFEF00D
This feels fresh to me. Most discussions about AI that I'm exposed to are the direct capabilities and shallow speculation on the future potential. I've seen the scaling factors in some cases but I don't think I've been given the perspective on just how prohibitive they are.
35 likes
@RedmotionGames
"Ye canne break the laws of physics, Captain"
4 likes
@penguinista
We are brute forcing performance with tons of compute at this point, but biology provides proof that there are much more efficient ways of getting cognition. At any time, someone could crack that problem and suddenly be able to make vastly better use of the ridiculous amounts of compute we have built up. That could result in a sudden step change to much more powerful models.
41 likes
@quest_onchannel54
Love how dynamic ai research and anlaysis is. I feel like im watching the shuttlecock in a game of badmitton between ai developers and critics. Whap asi in a 5 years.. whap its not gonna happen cuz of costs.., whap! the super rich get asi first.
20 likes
@Apjooz
Quantitative increases of compute brought qualitative shifts in the past but everyone assumes it can't happen again.
12 likes
@johanleion
Terrific grounded and insightful interview. Thank you all so much1
@artukikemty
Excellent comments from Toby Ord. The thing is for keeping hype and investments, companies need to exagerate the results and capabilities. Anyway if they were eventually true, this technology will need to be regulated in order keep social stability.
5 likes
@tomoconnor3580
Great, thoughtful conversation
1 likes
@carlcproductions
At this rate we’ll never invent the computer…
4 likes
@hopsgegangen2575
Excellent wide ranging discussion I think the only thing that’s missing is a discussion of data quality and application specific data sets such as medical journals
@JulianLuckeeSouth
finally someone addressing the massive elephant in the room. I can't fathom how often I've heard gearheads crowing about exponential scaling when it you just look at the graphs they're talking logarithmic. Literally the inverse of exponential.
24 likes
@Ryan.G.Spalding
This was an exceptional conversation. Well done. Subscribe
2 likes
@peters616
Great discussion about the reality (and limits) of scaling, and other types of improvements to model ability like RL and inference time compute. Some excellent takes on what's really going on those fronts. Regarding the latter discussion, I was a bit surprised that Toby is so strongly in favor of a moratorium on AI improvement especially after the first 2 hour discussion going in the other direction basically that in many ways we're reaching an asymptote on model intelligence gains and there is a decent amount of overhyping of the models abilities. I work all day with the latest models and can say there is really no chance that the current transformer based models pose any physical risk to the human race, they do not have any memory (they're stateless) and can only think for a very discrete time period when given an input and only have continuity until they generate an output. Any other trappings of memory or continuing agency is in a way an illusion, that is all done with re-prompting and a wrapper program. The models themselves can only generate output tokens in response to receiving input tokens they can't do anything else, and are completely reset back to a blank slate after generating the output.
3 likes
@eyuan2364
I love this episode. Great discussions all throughout