Laser & Optoelectronics Progress, Volume. 59, Issue 18, 1815004(2022)

Convolutional Neural Network Method for Crowd Counting Improved using Involution Operator

Zhaoxin Li, Shuhua Lu*, Lingqiang Lan, and Qiyuan Liu
Author Affiliations
  • College of Information and Cyber Security, People’s Public Security University of China, Beijing 102600, China
  • show less

    Most existing crowd counting methods use convolution operations to extract features. However, extracting and transmitting spatial diversity feature information are difficult. In this paper, we propose an Involution-improved single-column deep crowd-counting network to mitigate these problems. Using VGG-16 as the backbone, the proposed network uses an Involution operator combined with residual connection to replace the convolution operation, thereby enhancing the perception and transmission for spatial feature information. The dilated convolution was adopted to expand the receptive field while maintaining resolution to enrich deep semantic features. Additionally, we used the joint loss function to supervise the network training, improving counting accuracy and global information correlation. Compared with the baseline model, the performance of the proposed method across the ShangHaiTech, UCF-QNRF, and UCF_CC_50 datasets considerably is improved, demonstrating that our approach outperforms many current advanced algorithms. Furthermore, results show that the proposed crowd counting method has higher accuracy and better robustness than other methods.

    Tools

    Get Citation

    Copy Citation Text

    Zhaoxin Li, Shuhua Lu, Lingqiang Lan, Qiyuan Liu. Convolutional Neural Network Method for Crowd Counting Improved using Involution Operator[J]. Laser & Optoelectronics Progress, 2022, 59(18): 1815004

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Machine Vision

    Received: Jun. 21, 2021

    Accepted: Jul. 20, 2021

    Published Online: Aug. 29, 2022

    The Author Email: Lu Shuhua (lushuhua@ppsuc.edu.cn)

    DOI:10.3788/LOP202259.1815004

    Topics