


This inspired the investigation of camera-based SSC solutions, which were initially put forth in the ground-breaking work of MonoScene. Yet, cameras are more affordable and offer better visual indications of the driving environment, but LiDAR sensors are more costly and less portable. LiDAR is regarded as a main modality by most current SSC systems to provide precise 3D geometric measurements. Nevertheless, modern SSC techniques still lag below human perception in driving scenarios in terms of performance.

Humans readily reason about scene geometry and semantics based on imperfect observations, which supports this endeavor. Scene reconstruction for viewable areas and scene hallucination for obstructed sections are two subtasks an SSC solution must handle concurrently. Semantic scene completion (SSC), a method for jointly inferring the whole scene geometry and semantics from sparse observations, was offered to solve the problems. The lack of sensor resolution and the partial observation caused by the small field of vision and occlusions make it challenging to get precise and comprehensive 3D information about the actual environment. It directly influences later activities like planning and map creation. Understanding a holistic 3D picture is a significant challenge for autonomous vehicles (AV) to perceive.
