Opto-Electronic Engineering, Volume. 51, Issue 11, 240212-1(2024)

DES-YOLO: a more accurate object detection method

Huawei Zheng... Fei Wang* and Jianbang Gao |Show fewer author(s)
Author Affiliations
  • School of Electronic Engineering, Xi’an Shiyou University, Xi’an, Shaanxi 710065, China
  • show less

    To address the challenges of complex backgrounds, small targets, and dense distributions in images, an improved method called DES-YOLO is proposed. By introducing the deformable attention module (DAM), the network can dynamically focus on key regions, improving object recognition and localization accuracy. The efficient intersection over union (EIoU) loss function is employed to reduce the impact of low-quality samples, enhancing the model's generalization ability and detection accuracy. A shallow feature map layer of 160 pixel×160 pixel is added to the network head to strengthen small target feature extraction. A stepwise training strategy is also adopted to further improve model performance. Experimental results show that the mAP@50 of the model increased by 1.4% on the remote sensing dataset and by 1.7% on the textile dataset, demonstrating the broad applicability and effectiveness of DES-YOLO.

    Keywords
    Tools

    Get Citation

    Copy Citation Text

    Huawei Zheng, Fei Wang, Jianbang Gao. DES-YOLO: a more accurate object detection method[J]. Opto-Electronic Engineering, 2024, 51(11): 240212-1

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Article

    Received: Sep. 9, 2024

    Accepted: Oct. 11, 2024

    Published Online: Jan. 24, 2025

    The Author Email: Wang Fei (王飞)

    DOI:10.12086/oee.2024.240212

    Topics