Yu Yang1,2, Alan Liang2, Jianbiao Mei1, Yukai Ma1, Yong Liu1, Gim Hee Lee2
1 Zhejiang University 2 National University of Singapore
-
[2025-09-19]Our π³-Scene is accepted by NeurIPS 2025! -
[2025-06-18]We released our project website here. -
[2025-06-16]The paper can be accessed at arxiv.
Overview of π³-Scene. a unified world generator that supports multi-granular controllability through high-level text-to-layout generation and low-level BEV layout conditioning. It performs joint occupancy, image, and video generation for 3D scene synthesis and reconstruction with high fidelity.
Please refer to the following documents to set up the environment and run π³-Scene:
- π οΈ Installation Guide
- π Dataset Preparation
- β‘ Train and Evaluation
- Paper & Project Page
- Release the Training Code
- Release the Inference Code
- Release the Processed Data
We are grateful for the following open-source projects that inspired or assisted the development of π³-Scene:
| Occupancy Generation | Video & Driving Synthesis |
|---|---|
| SemCity | MagicDrive |
| DynamicCity | DriveArena |
| OccSora | LiDARCrafter |
| UniScene | X-Drive |
Special thanks to these communities for their incredible contributions to the field!
@article{yang2025xscene,
title={X-Scene: Large-Scale Driving Scene Generation with High Fidelity and Flexible Controllability},
author={Yang, Yu and Liang, Alan and Mei, Jianbiao and Ma, Yukai and Liu, Yong and Lee, Gim Hee},
journal={arXiv preprint arXiv:2506.13558},
year={2025}
}


