Remote Sensing, Vol. 17, Pages 2708: HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images


Remote Sensing, Vol. 17, Pages 2708: HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images

Remote Sensing doi: 10.3390/rs17152708

Authors:
Pengfei Zhang
Jian Liu
Jianqiang Zhang
Yiping Liu
Jiahao Shi

The growing use of remote-sensing technologies has placed greater demands on object-detection algorithms, which still face challenges. This study proposes a hierarchical adaptive feature aggregation network (HAF-YOLO) to improve detection precision in remote-sensing images. It addresses issues such as small object size, complex backgrounds, scale variation, and dense object distributions by incorporating three core modules: dynamic-cooperative multimodal fusion architecture (DyCoMF-Arch), multiscale wavelet-enhanced aggregation network (MWA-Net), and spatial-deformable dynamic enhancement module (SDDE-Module). DyCoMF-Arch builds a hierarchical feature pyramid using multistage spatial compression and expansion, with dynamic weight allocation to extract salient features. MWA-Net applies wavelet-transform-based convolution to decompose features, preserving high-frequency detail and enhancing representation of small-scale objects. SDDE-Module integrates spatial coordinate encoding and multidirectional convolution to reduce localization interference and overcome fixed sampling limitations for geometric deformations. Experiments on the NWPU VHR-10 and DIOR datasets show that HAF-YOLO achieved mAP50 scores of 85.0% and 78.1%, improving on YOLOv8 by 4.8% and 3.1%, respectively. HAF-YOLO also maintained a low computational cost of 11.8 GFLOPs, outperforming other YOLO models. Ablation studies validated the effectiveness of each module and their combined optimization. This study presents a novel approach for remote-sensing object detection, with theoretical and practical value.



Source link

Pengfei Zhang www.mdpi.com