Content
summary Summary

As we enter 2025, the AI community remains divided on what really drives progress in artificial intelligence.

Ad

Gary Marcus kicked off the year with his characteristic skepticism. "OpenAI will continue to preview products months and perhaps years before they are solid and widely available at an accessible price," he writes, predicting modest returns from AI models and warning that the much-hyped AI "agents" won't live up to their promise in 2025 because they are "far from reliable, except possibly in very narrow use cases."

AI researcher François Chollet - who recently left Google after a decade to start his own company and whose ARC-AGI benchmark recently made waves after OpenAI's o3 model performed well in it - suggests that throwing more compute and data at the problem isn't the answer. In fact, he claims, anyone who truly understands intelligence could develop an AGI with just a million dollars in training costs.

Toby Pohlen, who left DeepMind to join Elon Musk's xAI as a founding member, pushed back against Chollet. According to Pohlen, DeepMind used to share Chollet's view that massive scale wasn't necessary - but the evidence kept pointing in the opposite direction. However, even xAI's massive cluster, planned to reach 200,000 Nvidia GPUs, hasn't yet delivered on Musk's promise of "world's most powerful AI by every metric" by the end of 2024.

Ad
Ad

Tim Dettmers from Allen AI recently mapped out three possible paths: traditional data center scaling (viable for about two more years), dynamic scaling (leading to either specialized or flexible models), and knowledge distillation (which might follow entirely different rules than other scaling approaches). But ultimately, Dettmers sees a "perfect storm" brewing - a combination of physical limitations and diminishing returns that could spell the end of the scaling era.

New frontiers, same debate

A new frontier in this debate opened up with test-time compute scaling, the approach behind OpenAI's o-models. Hugging Face researchers recently validated this strategy, which shifts massive computing power to the inference stage rather than pre-training. While promising, the high costs of running o3 on benchmarks like ARC suggest we're trading one scaling challenge for another.

This isn't just an academic debate. The soaring valuations of Big Tech and AI companies like OpenAI and Nvidia have been built on the promise that scaling equals progress. If the scaling skeptics are right, we're not just facing a technical dead end - we could be looking at a massive correction in tech valuations that would send shockwaves through financial markets.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • The AI community remains divided on what drives progress in artificial intelligence as we enter 2025. LLM Skeptics like Gary Marcus predict modest returns from AI models and warn that hyped AI "agents" won't live up to their promise due to reliability issues. 
  • AI researcher François Chollet suggests that throwing more compute and data at the problem isn't the answer, claiming that with a true understanding of intelligence, an AGI could be developed with as little as a million dollars including training costs. However, xAI's Toby Pohlen argues that the evidence points to the need for massive scale.
  • The debate extends to test-time compute scaling, which shifts massive computing power to the inference stage, but the astronomical costs suggest we're trading one scaling challenge for another.
Online journalist Matthias is the co-founder and publisher of THE DECODER. He believes that artificial intelligence will fundamentally change the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.