AIResearch AIResearch
Back to articles
Science

AI System Reconstructs Dynamic Scenes With Unprecedented Accuracy

New LiDAR-visual method achieves 28.73 PSNR score, enabling robots to navigate complex environments while filtering out moving objects

AI Research
November 14, 2025
2 min read
AI System Reconstructs Dynamic Scenes With Unprecedented Accuracy

A new artificial intelligence system can now reconstruct dynamic three-dimensional environments with remarkable precision, potentially transforming how autonomous vehicles and robots perceive and navigate the real world. The technology addresses a critical limitation in current 3D mapping systems: their inability to accurately handle moving objects and changing scenes.

Researchers have developed LVD-GS, a LiDAR-visual system that achieves state-of-the-art performance in reconstructing dynamic environments. Unlike previous methods that relied on single representation schemes, this approach uses hierarchical explicit-implicit representations to handle moving objects while maintaining scene consistency. The system combines geometric, semantic, and visual features to create high-fidelity reconstructions that remain accurate even when objects are in motion.

The methodology integrates two key modules: a Hierarchical Representation module and an Explicit-Implicit Joint Modeling module. The system leverages foundation models including Grounded SAM and DINO features to extract semantic information, while using LiDAR data for geometric precision. Through a chain-of-thought inspired process, the system selectively focuses on relevant scene elements, similar to how humans process visual information. The explicit-implicit collaboration generates motion masks that filter out transient objects without losing important scene features.

Experimental results demonstrate significant improvements over existing methods. On the nuScenes dataset, LVD-GS achieved a PSNR (peak signal-to-noise ratio) of 28.73 dB compared to 24.25 dB for the next best method—a 4.48 dB improvement. The system also reduced tracking errors substantially, with ATE-RMSE (absolute trajectory error root mean square) dropping to 1.27 meters compared to 2.97 meters for competing approaches. The method showed consistent performance across urban, highway, and campus environments, generating photorealistic reconstructions with enhanced vehicle contours, architectural structures, and road surface details.

This advancement matters because accurate dynamic scene reconstruction is essential for real-world applications like autonomous driving and robotic navigation. Current systems often fail when objects move or scenes change rapidly, leading to trajectory errors that can compromise safety. By filtering out moving objects while maintaining scene consistency, LVD-GS enables more reliable operation in complex environments. The technology could improve how self-driving cars perceive other vehicles and pedestrians, or help robots navigate crowded spaces without losing track of their surroundings.

The system does have limitations. Researchers noted that some competing methods showed large errors in specific environments, limiting their applicability to dynamic scenes. While LVD-GS performs well across multiple scenarios, its evaluation was limited to the first 350 frames of sequences, and future work will need to address longer-term performance and instance-level cognitive mapping capabilities.

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn