The field of autonomous driving perception is rapidly advancing, with a focus on improving the robustness and accuracy of perception systems in various environmental conditions. Recent research has emphasized the importance of addressing the challenges posed by dynamic weather conditions, varying camera heights, and sensor setups.
Notable advancements include the development of novel domain adaptation pipelines, which enable the transformation of clear-weather images into adverse weather conditions, and the proposal of dual-illumination adaptive enhancement networks to handle diverse lighting conditions.
Furthermore, researchers have explored the use of cross-view transformers for bird's-eye view generation, and test-time adaptation methods to mitigate the effects of domain shifts. The introduction of new datasets, such as those focusing on weather-extended salient object detection and illumination-sensor mapping, has also facilitated the development of more robust and generalizable perception models.
Some noteworthy papers in this area include: CHARM3R, which proposes a camera height robust monocular 3D detector that averages both depth estimates within the model to improve generalization to unseen camera heights. WXSOD, which introduces a novel dataset for salient object detection in adverse weather conditions and proposes a weather-aware feature aggregation network to achieve superior performance on this dataset.