Content
summary Summary

OpenAI has started using Google's Tensor Processing Units (TPUs) to run ChatGPT and other AI products, according to The Information. This marks the first time OpenAI is relying on chips beyond Nvidia's graphics processors at scale.

Ad

The TPUs are rented through Google Cloud and are aimed at lowering the costs of inference - the process of running trained models to generate answers for new prompts. Until now, OpenAI has been one of Nvidia's biggest customers, using its GPUs for both training and deploying large language models.

The partnership with Google has limits. According to The Information, Google is not giving OpenAI access to its most powerful TPU models. A Google Cloud employee confirmed this restriction.

A message to Microsoft

The move sends a clear signal to Microsoft, OpenAI's largest investor and the company that has been providing much of the infrastructure for OpenAI's products. By shifting some workloads onto Google's infrastructure, OpenAI is using its relationship with a key Microsoft competitor as strategic leverage.

Ad
Ad

OpenAI CEO Sam Altman and Microsoft CEO Satya Nadella are reportedly in ongoing talks about the companies' partnership. OpenAI has also expanded its compute capacity through a deal with Oracle.

Google Cloud competes directly with Microsoft Azure and is a major growth driver for Google. OpenAI's decision to use Google's infrastructure impacts not just the AI market, but also the cloud sector, which has been central to Microsoft's stock performance in recent years.

Google originally kept its TPUs for internal use, but is now opening them up to more outside partners. Besides OpenAI, customers include Apple and the startups Anthropic and Safe Superintelligence, both founded by former OpenAI executives.

According to research firm Epoch AI, Google's infrastructure gives it the world's largest AI computing capacity. The TPU cloud is a core part of Google's AI strategy and goes head to head with Nvidia's GPUs, especially for running large models

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • OpenAI is now using Google's TPU chips to run ChatGPT and other AI products, shifting from exclusive reliance on Nvidia GPUs and Microsoft's infrastructure, with the main goal of lowering inference costs.
  • Google has not provided OpenAI with its most powerful TPU models, according to a Google Cloud employee, so the cooperation is limited but still represents a significant step in the relationship between the companies.
  • The move serves as a signal to Microsoft, OpenAI's main investor and infrastructure partner, showing that OpenAI is prepared to use the infrastructure of a direct competitor, which could impact ongoing negotiations between the firms.
Sources
Matthias is the co-founder and publisher of THE DECODER, exploring how AI is fundamentally changing the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.