The field of 3D scene understanding and reconstruction is rapidly advancing, with a focus on developing more efficient and accurate methods for processing and analyzing 3D data. Recent research has explored the use of novel frameworks and techniques, such as Gaussian splatting, point cloud processing, and semantic scene graph generation, to improve the accuracy and speed of 3D scene reconstruction. Additionally, there is a growing interest in developing methods for open-vocabulary 3D instance segmentation, which enables the recognition and segmentation of objects in 3D scenes without requiring prior knowledge of the object categories. Noteworthy papers in this area include ScenePainter, which proposes a new framework for semantically consistent 3D scene generation, and VisHall3D, which introduces a novel two-stage framework for monocular semantic scene completion. Other notable papers include FROSS, which presents an innovative approach for online and faster-than-real-time 3D semantic scene graph generation, and NeuroVoxel-LM, which proposes a novel framework for language-aligned 3D perception via dynamic voxelization and meta-embedding.
Advancements in 3D Scene Understanding and Reconstruction
Sources
3DGauCIM: Accelerating Static/Dynamic 3D Gaussian Splatting via Digital CIM for High Frame Rate Real-Time Edge Rendering
VisHall3D: Monocular Semantic Scene Completion from Reconstructing the Visible Regions to Hallucinating the Invisible Regions
GS-Occ3D: Scaling Vision-only Occupancy Reconstruction for Autonomous Driving with Gaussian Splatting
Co-Win: Joint Object Detection and Instance Segmentation in LiDAR Point Clouds via Collaborative Window Processing
SLTarch: Towards Scalable Point-Based Neural Rendering by Taming Workload Imbalance and Memory Irregularity