Content
summary Summary

Newly released data from Chinese AI provider Deepseek reveals that AI language models could, in theory, generate substantial profit margins—even at prices significantly lower than OpenAI’s.

Ad

Deepseek has offered a rare glimpse into the operating costs and potential profitability of its AI services. The numbers suggest the company could achieve a theoretical profit margin of 545 percent if it fully monetized its services, even while maintaining its open source strategy and charging less than competitors like OpenAI.

Smart resource management keeps costs down

During a 24-hour test period, Deepseek's models processed 608 billion input tokens and 168 billion output tokens. The company managed to serve more than half of the inputs (56.3 percent) from a cache, significantly reducing costs.

To maximize efficiency, Deepseek uses a dynamic resource allocation system. During peak daytime hours, all nodes handle inference requests. At night, when demand drops, the company redirects resources to research and training tasks.

Ad
Ad

The hardware infrastructure for this operation costs $87,072 per day, using an average of 226.75 server nodes. Each node contains eight Nvidia H800 GPUs, with calculations based on an estimated leasing cost of two dollars per GPU per hour.

A single H800 node processes about 73,700 input tokens per second during prefilling or 14,800 output tokens during decoding. The average output speed reaches 20 to 22 tokens per second.

Bar chart: 24-hour comparison of constant costs (orange) and fluctuating theoretical revenues (blue) based on R1's API prices.
Deepseek's financial data shows theoretical peak revenue could exceed operating costs by 6x when using optimal R1 model pricing. | Image: Deepseek

If Deepseek charged full price for every processed token using its premium R1 model rates ($0.14 per million input tokens for cache hits, $0.55 for cache misses, and $2.19 per million output tokens), daily revenue would reach $562,027.

But real-world revenues fall well below these theoretical figures, the company says. Deepseek's standard V3 model is priced below R1, most services are offered for free, and the company applies nightly discounts. For now, only API access is generating revenue.

The commoditization of AI services

Deepseek's unusual transparency reflects an industry dynamic: while AI language models can theoretically generate substantial profit margins, capturing that value proves difficult. Between market competition, tiered pricing structures, and the need to provide free services, actual profits tend to shrink dramatically.

Recommendation

In this context, OpenAI's recent pricing strategy is particularly noteworthy. The company's latest GPT-4.5 commands premium prices far above both its predecessors and competitors such as Deepseek, despite offering only modest performance improvements.

Deepseek's data suggests that language models are evolving into commodity services where premium pricing no longer reflects actual performance advantages. This creates additional pressure on Western AI companies like OpenAI, which are losing billions and facing significant operating costs while market pressures drive down prices.

This could be one of the reasons that OpenAI's GTM Manager Adam Goldberg recently emphasized that success in AI requires controlling the entire value chain - from infrastructure and data to models and applications. As language models become commoditized, the competitive advantage may lie less in the models themselves and more in a company's ability to integrate and optimize across the complete technology stack.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Deepseek has shared details about the operational expenses and potential earnings of its AI offerings, revealing that with full monetization, the company could achieve a theoretical profit margin of 545%, even while charging much lower prices compared to OpenAI.
  • By employing smart resource management techniques, including the use of caches and dynamically allocating resources for inference, research, and training, Deepseek substantially reduces its costs.
  • The high theoretical margins stand in stark contrast to OpenAI's most recent GPT-4.5 model, which is significantly more expensive than its predecessors, despite only marginal improvements in performance.
Sources
Matthias is the co-founder and publisher of THE DECODER, exploring how AI is fundamentally changing the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.