Abstract
The efficacy of autonomous driving systems is fundamentally linked to their ability to accurately perceive and interpret their surroundings. The investigation addresses the challenges inherent in the sparse and unordered 3D point cloud data produced by LiDAR sensors by introducing a refined PointPillars network architecture. This innovative framework, which seamlessly integrates Transformer modules and ECA-PP (Efficient Channel Attention PointPillars) modules, substantially elevates the precision and efficiency of 3D object detection tasks. The Transformer module adeptly harnesses self-attention mechanisms to extract and process global contextual information from point cloud data, capturing not only local features but also comprehending the overall structure and layout of the scene. Meanwhile, the ECA-PP module enhances feature discrimination by focusing on channel attention, effectively distinguishing and emphasizing salient features crucial for object detection. Comprehensive experimental evaluations, conducted using the KITTI dataset and nuScenes dataset, have yielded compelling evidence that the proposed algorithm significantly outperforms current technologies in terms of detection accuracy for various categories, including cars, pedestrians, and cyclists.
Get full access to this article
View all access options for this article.
