Discussion about this post

User's avatar
Neural Foundry's avatar

The Orion failure is probaly the most underrated story in AI right now. OpenAI betting $500M twice on brute-force scaling and getting no improvement in coding/math is basically the death of the old paradigm. I've been watching the shift to inference-time compute play out across our own systems, and the cost dynamics are genuinly different when a model can "think longer" instead of just being bigger. The TPU advantage for Google is huge tho, avoiding that Nvidia tax entirely changes the game when running agentic workflows that burn tokens continuously for hours.

Expand full comment
1 more comment...

No posts

Ready for more?