Content
summary Summary

Google says it now processes more than 1.3 quadrillion tokens every month with its AI models. But this headline number mostly reflects computing effort, not real usage or practical value, and it raises questions about Google's own environmental claims.

Ad

According to Google, it processes over 1.3 quadrillion tokens per month with its AI products and interfaces. This new brand was announced by Google CEO Sundar Pichai at a Google Cloud event.

Google announced the milestone during a recent Google Cloud event, with CEO Sundar Pichai highlighting the figure. Back in June, Google said it had reached 980 trillion tokens, more than double May's total. The latest jump adds about 320 trillion tokens since June, but growth has already slowed, a trend not reflected in Pichai's presentation.

Token consumption is growing faster than actual usage

Tokens are the smallest unit processed by large language models, similar to word fragments or syllables. A huge token count sounds like surging usage, but in reality, it's primarily a measure of rising computational complexity.

Ad
Ad

The main driver is likely Google's rollout of reasoning models like Gemini 2.5 Flash. These models perform far more internal calculations for every request. Even a simple greeting like "Hi" can trigger dozens of processing steps before a response appears in today's reasoning models.

A recent analysis showed that Gemini Flash 2.5 uses about 17 times more tokens per request than its previous version and is up to 150 times pricier for reasoning tasks. Moreover, complex features like video, image, and audio processing are likely factored into the total, but Google doesn't break those out.

So, the token number is mostly a measure of backend computing load and infrastructure scaling, not a direct indicator of user activity or actual benefit.

Google's token consumption vs. Google's environmental claims

Google's new token stats also reinforce a key criticism of Google's own environmental report: by measuring only the smallest unit of computation, the study ignores the true scale of AI operations and downplays the real environmental impact. The report claims a typical Gemini text prompt uses only 0.24 watt-hours of electricity, 0.03 grams of CO₂, and 0.26 milliliters of water—supposedly less than nine seconds of TV time.

Those figures assume a "typical" short text prompt in the Gemini app. Google doesn't say whether these are for lightweight language models (likely) or for the much more resource-hungry reasoning models (unlikely). The study also leaves out heavier use cases like document analysis, image or audio generation, multimodal prompts, or agent-driven web searches.

Recommendation

Viewed in this light, Google's 1.3 quadrillion tokens mainly highlight how rapidly its computing demands are accelerating. Yet this surge in system-wide usage doesn't appear in Google's official environmental assessment. It's a bit like an automaker touting low fuel consumption while idling, then calling the entire fleet "green" without accounting for real-world driving or manufacturing.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Google reports processing over 1.3 quadrillion tokens per month with its AI products, a sharp increase since May that is likely driven by the adoption of more complex reasoning models like Gemini 2.5 Flash.
  • The token count mainly measures the computing effort and scalability of Google's infrastructure, since reasoning models require significantly more internal calculations, but it does not directly reflect actual usage or practical benefits.
  • The rapid growth in tokens brings new scrutiny to Google’s own environmental analysis, which downplays the energy use of generative AI by overlooking the mounting computing requirements that the increase in tokens actually highlights.
Sources
Matthias is the co-founder and publisher of THE DECODER, exploring how AI is fundamentally changing the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.