Content
summary Summary

Nvidia's Avatar Cloud Engine began as an AI service for the service industry. Now, Nvidia is bringing generative AI tools to video games.

Since the breakthrough of large language models, there have been attempts to use such AI systems in video games to enable dynamic dialog with non-player characters (NPCs). At Computex, Nvidia announced that it will open the Avatar Cloud Engine to video games.

Nvidia's ACE consists of three components: NeMo, Riva, and Omniverse Audio2Face. NeMo provides large language models that developers can customize with backstory and dialogue data. Riva can recognize speech and convert text to audio, enabling live conversations with NeMo. Audio2Face converts Riva's audio output into facial animations that can be used, for example, to control MetaHuman characters in Unreal Engine 5 via Omniverse links.

In collaboration with AI startup Convai, Nvidia shows what an ACE-powered MetaHuman could look like in a video game.

Ad
Ad

 

According to Nvidia, NPC Jin answers the player's questions in natural language, realistically and in line with a background story. The low latency in the demo is striking, details about the language model and the hardware in the background are not available.

Different models should enable low latencies

"With NVIDIA ACE for Games, Convai’s tools can achieve the latency and quality needed to make AI non-playable characters available to nearly every developer in a cost efficient way," said Purnendu Mukherjee, founder and CEO of Convai.

According to Nvidia, the AI models in ACE vary in size, performance, and quality. "The ACE for Games foundry service will help developers fine-tune models for their games, then deploy via Nvidia DGX Cloud, GeForce RTX PCs or on premises for real-time inferencing." The models are optimized for latency, a "critical requirement for immersive, responsive interactions in games," the company said.

GSC Game World is one of the first development studios to use at least one of ACE's tools: Audio2Face will be used in S.T.A.L.K.E.R. 2 Heart of Chernobyl. However, AI-generated stories will not be told by NPCs around the campfire for the time being.

Recommendation

Nvidia's DGX GH200 is like a giant GPU

In addition to ACE for Games, Nvidia announced a number of other collaborations and products, including the DGX GH200 AI supercomputer with 256 Grace Hopper chips totaling 144 terabytes of memory and exaflop AI performance. According to Nvidia, the DGX GH200 GPT-3 trains 2.2 times faster than a DGX H100 cluster.

According to Nvidia CEO Jensen Huang, Google, Meta, and Microsoft are among the first companies to gain access to the DGX GH200 and evaluate the platform's capabilities.

Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Nvidia is opening its Avatar Cloud Engine (ACE) to video games, enabling large language models, speech recognition, text-to-speech conversion, and facial animation.
  • Working with AI startup Convai, Nvidia is demonstrating what an ACE-powered NPC could look like in a video game.
  • The company also announced the DGX GH200 AI supercomputer, which stacks 256 "Grace Hopper" chips together.
Max is managing editor at THE DECODER. As a trained philosopher, he deals with consciousness, AI, and the question of whether machines can really think or just pretend to.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.