面向目标检测的图特征增强点云采样方法

Graph feature-enhanced point cloud sampling for object detection

  • 摘要: 激光雷达采集的点云中,前景目标点所占比例较小,传统无监督采样方法难以选择性地保留足够多的前景点,不可避免会丢失部分前景目标信息,影响基于点云的目标检测网络性能。本文提出了一种图特征增强的并行点云采样方法,利用前/背景分类标签进行监督,大幅提高了采样点中前景点的比例,相较于使用点特征进行监督的方法,所提出的基于图特征的方法能获取更多的点云局部特征,适用于目标检测网络浅层的采样过程。在KITTI自动驾驶数据集上进行测试,实验结果表明:本文方法采样的前景点比例高达99%,能够有效提取受遮挡目标和远处目标等点云稀疏区域的特征信息,从而提高目标检测网络的性能。引入该方法后,对困难情况下的车辆、行人和两轮车的检测平均精度分别提升了8.58%、2.27%和3.12%。此外,该方法设计灵活,易于集成到依赖点云采样过程的各种3D点云任务中。

     

    Abstract: The point cloud data acquired through LiDAR is extensive, and its feature extraction demands significant computational resources, making efficient sampling crucial for enhancing processing speed. Within point clouds, most points represent the background, with their density typically higher near the sensor and decreasing with distance. Over-concentration of sample points or excessive inclusion of background points can lead to the loss of critical foreground information, negatively impacting object detection performance. Traditional sampling techniques, such as farthest point sampling and random sampling, operate in an unsupervised manner, failing to harness the rich feature information embedded within the point cloud. Although farthest point sampling has been widely adopted in numerous object detection approaches with commendable outcomes, its inherently sequential nature, where each sampling step depends on the preceding one, can compromise overall detection efficiency. To address these limitations, we propose a novel supervised point cloud sampling method grounded in graph features. This innovative approach enables parallel sampling and utilizes foreground and background classification as supervisory signals, significantly boosting the proportion of foreground points in the sampled set. Compared with methods that directly use point features for supervision, the graph-feature-based method captures a greater number of local point cloud features, making it particularly well-suited for initial-stage sampling. Experimental results on the KITTI autonomous driving dataset show that the proposed method achieves a remarkable 99% ratio of foreground points in the sampled set. This effectively extracts feature information from sparse point cloud regions, such as occluded and distant targets, thereby enhancing the performance of the object detection network. After incorporating this method, the mean average precision for detecting car, pedestrians, and cyclist under hard conditions was improved by 8.58%, 2.27%, and 3.12%, respectively. Moreover, the proposed method emphasizes flexibility and ease of integration into a wide range of 3D point cloud applications that depend on effective sampling.

     

/

返回文章
返回