顾及小目标特征的无人机影像人员检测方法

Person Detection Method for UAV Imagery Considering Small Object Features

  • 摘要: 利用无人机影像进行人员检测在应急救援与社会安全中具有重要意义,然而,无人机影像中的人员目标较小且背景复杂导致人员检测精度较低。为解决这一问题,提出了一种顾及小目标特征的无人机影像人员检测网络,首先,在特征提取阶段使用空间深度转换卷积代替下采样层,减少特征提取过程中的小目标特征丢失;其次,设计选择性特征融合模块(Selective Feature Fusion,SFF),以深层特征权重引导获取浅层特征图中的重点区域,减轻低层特征中的背景噪声干扰;最后,建立上下文感知模块(Context-aware Module,CAM),综合多类型上下文信息,提高网络对于小目标人员的判别能力。为验证方法的有效性,在公开数据集上开展了案例实验分析,该方法平均准确率为68.9%,准确率为75.5%,召回率为67.7%,相较于以往经典的目标检测算法,分别提升了3.1%-29.5%、0.9%-8.6%、1.0%-57.9%,结果表明该方法具有较高的精度,此外,通过对不同场景、不同天气下的无人机影像进行测试,进一步证明了该方法的泛化性与适用性。

     

    Abstract: Objectives: The rapid development of UAV remote sensing technology is leading to an increased role for these devices in emergency search and rescue operations and law enforcement tracking. This is due to the unique advantages offered by UAVs, including rapid response, wide-area coverage, and a panoramic view. The utilisation of UAV imagery for person detection is of paramount importance in the context of emergency rescue and social security. However, due to the small and complex background of the person objects in the UAV imagery, the high-precision person detection task still presents a significant challenge. Methods: To address the aforementioned issue, a UAV image person detection network is proposed. The network employs a spatial depth-transformed convolution in lieu of a downsampling layer in the feature extraction stage. This approach ensures the retention of all information in the channel dimensions, preventing information loss and facilitating the preservation of small object features throughout the feature extraction process. Secondly, a selective feature fusion module is designed to address the issue of noise interference. The module employs highlevel features as weights to guide the crucial information in low-level features, thereby mitigating the impact of background noise on the latter. Furthermore, it facilitates a more comprehensive integration of the semantic information of high-level features and the detailed information of low-level features. This integration significantly enhances the network's capacity to distinguish between the background and the foreground information present in UAV images. Finally, a context-aware module is designed to address the issue of limited feature information associated with small objects. This module integrates environmental information surrounding the object, including local features, contextual information, and global context. The module integrates environmental information surrounding the object, including local features, surrounding context, and global context, thus enhancing the contextual features of small objects and improving the final object detection accuracy. Results: To verify the effectiveness of the proposed method, case experimental analyses are conducted on public datasets. The method achieved an mean Average Precision of 68.9%, with precision at 75.5% and recall at 67.7%. Compared to conventional object detection algorithms, the method showed improvements ranging from 3.1% to 29.5% in mean Average Precision, 0.9% to 8.6% in precision, and 1.0% to 57.9% in recall. These results indicate that the method exhibits a high level of accuracy. Additionally, the generalization and applicability of the method are confirmed through testing on UAV images in various scenarios and under different weather conditions. Conclusions: The method effectively solves the challenge of detecting small persons in UAV images and significantly improves the detection accuracy. It has a wide range of application prospects in the fields of emergency rescue and social security, and possesses good generalisation ability for UAV image detection tasks in different environments.

     

/

返回文章
返回