Content
summary Summary

Tencent has released Hunyuan World Model 1.0, an open-source generative AI model that creates 3D virtual scenes from text or image prompts.

Ad

The company says it's the first open-source model designed for standard graphics pipelines, making it compatible with game engines, VR platforms, and simulation tools. The goal is to help creators move quickly from concept to 3D content without running into proprietary barriers.

Three-part panorama: loose brushstrokes of mountains in moonlight, a coral-covered underwater city, and a floating castle in a colorful sea of clouds.
Three panoramas generated with Hunyuan World Model 1.0 show how the model automatically creates immersive 3D scenes from text prompts that can be used in game engines. | Image: Tencent (Screenshot)

A key feature is the model's ability to separate objects within a scene, letting users move or edit elements like cars, trees, or furniture individually. The sky is also isolated and can be used as a dynamic lighting source to help with realistic rendering and interactive experiences.

Hunyuan World Model 1.0 lets users edit individual scene elements after generation. | Image: Tencent

360-degree panoramas with limited exploration

Hunyuan World Model 1.0 combines panoramic image generation with hierarchical 3D reconstruction. It supports two main input types: text-to-world and image-to-world. The generated scenes can be exported as mesh files and, according to Tencent, integrate smoothly into standard 3D workflows.

Ad
Ad

In practice, the model doesn't produce fully explorable 3D worlds like those in modern video games. Instead, users get interactive 360-degree panoramas. They can look around and navigate to a limited extent, but free movement is restricted. For more advanced camera movement or longer, consistent 3D video sequences, the Voyager add-on is required, as detailed in a recent research paper.

Video: Tencent

Tencent sees these visualizations as a starting point for VR, but the model is also suitable for a wide range of interactive and creative applications. Its text interpretation is designed to map complex scene descriptions into virtual spaces accurately, and it supports various compression and acceleration techniques for web and VR environments. The architecture uses a generative, semantically layered approach, producing scenes in a range of styles for creative and design uses.

Hunyuan World Model 1.0 is available as open source on GitHub and Hugging Face. An interactive demo is also available at sceneTo3D, but access requires a China-compatible login.

The release is part of Tencent's broader open-source push in AI. Alongside Hunyuan World Model 1.0, the company has released Hunyuan3D 2.0 for textured 3D model generation, HunyuanVideo for AI-powered video, and the Hunyuan-A13B language model with dynamic reasoning.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Recommendation
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Tencent has released Hunyuan World Model 1.0, an open-source AI model that generates 3D scenes based on text or image specifications.
  • The model combines panoramic image synthesis with hierarchical 3D reconstruction and exports the results as mesh files for various workflows in common graphics pipelines and game engines.
  • The generated environments are currently limited to interactive 360-degree panoramas and do not allow full free movement yet.
Matthias is the co-founder and publisher of THE DECODER, exploring how AI is fundamentally changing the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.