Dual-Modal Gated Fusion-Driven BEV 3D Object Detection: Enhancing Sustainable Intelligent Transportation in Nighttime Autonomous Driving

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Autonomous driving technology is a core enabler for new energy vehicle industrial upgrading and a critical pillar for achieving sustainable development goals (SDGs), especially sustainable urban mobility, low-carbon transportation, and efficient intelligent transportation systems (ITS). However, unstable nighttime low-light perception severely restricts autonomous driving deployment, hindering sustainable transportation development—rooted in visual feature degradation and cross-modal imbalance that impair 3D object detection (autonomous driving’s core perception technology). To address this and advance sustainable autonomous driving, this paper proposes a Bird’s-Eye View (BEV)-based multi-modal 3D object detection approach tailored for nighttime scenarios, integrating low-light adaptive components while preserving the original BEV pipeline. Without modifying core inference, it enhances low-light robustness and cross-modal fusion stability, ensuring reliable perception for sustainable autonomous driving operation. Extensive experiments on the nuScenes nighttime subset quantify performance via rigorous metrics (NDS, mAP, mATE). Results show the method outperforms BEVFusion with negligible parameter/inference overhead, achieving 1.13% NDS improvement. This validates its effectiveness and provides a sustainable technical tool for autonomous driving perception, promoting new energy vehicle popularization, optimizing urban ITS efficiency, reducing perception-related accidents and carbon emissions, and directly contributing to transportation and socio-economic sustainability.

Article activity feed