[ICLR26] Learning physics-grounded 4D dynamics with neural gaussian force fields

Abstract

Predicting physical dynamics from raw visual data remains a major challenge in AI. While recent video generation models have achieved impressive visual quality, they still cannot consistently generate physically plausible videos due to a lack of modeling of physical laws. Recent approaches combining 3D Gaussian splatting and physics engines can produce physically plausible videos, but are hindered by high computational costs in both reconstruction and simulation, and often lack robustness in complex real-world scenarios. To address these issues, we introduce Neural Gaussian Force Field (NGFF), an end-to-end neural framework that integrates 3D Gaussian perception with physics-based dynamic modeling to generate interactive, physically realistic 4D videos from multi-view RGB inputs, achieving two orders of magnitude faster than prior Gaussian simulators. To support training, we also present GSCollision, a 4D Gaussian dataset featuring diverse materials, multi-object interactions, and complex scenes, totaling over 640k rendered physical videos (∼4 TB). Evaluations on synthetic and real 3D scenarios show NGFF’s strong generalization and robustness in physical reasoning, advancing video prediction towards physics-grounded world models.

Publication
In International Conference on Learning Representations
Shiqian Li
Shiqian Li
Ph.D. '22
Ruihong Shen
Ruihong Shen
Zhi Class '23

My research interests include intuitive physics, few-shot learning, etc.

Chang Pan
Chang Pan
Tong Class '23

My research interests include cognitive reasoning, intuitive physics, etc.

Chi Zhang
Chi Zhang
Research Scientist
Yixin Zhu
Yixin Zhu
Assistant Professor

I build humanlike AI.

Related