Chinese Journal of Liquid Crystals and Displays, Volume. 37, Issue 12, 1614(2022)

Action recognition algorithm based on multi-scale and multi-branch features

Lei ZHANG1,2 and Guang-liang HAN1、*
Author Affiliations
  • 1Changchun Institute of Optics,Fine Mechanics and Physics,Chinese Acadamy of Sciences,Changchun 130033,China
  • 2University of Chinese Acadamy of Sciences,Beijing 100049,China
  • show less

    Aiming at the problems of insufficient feature extraction, incompleteness and low recognition accuracy in action recognition based on human skeleton sequence, a action recognition model based on multi-branch feature and multi-scale spatio-temporal feature is proposed in this paper. Firstly, the original data are enhanced by the combination of various algorithms. Secondly, the multi-branch feature input form is improved to multi-branch fusion feature information, which is input into the network, respectively. After a certain depth of network modules, it is fused together. Then, a multi-scale spatio-temporal convolution module is constructed as the basic module of the network to extract multi-scale spatio-temporal features. Finally, the overall network model is constructed to output action categories. The experimental results show that the recognition accuracy on Cross-subject and Cross-view of NTU RGB-D 60 data set is 89.6% and 95.1%, and the recognition accuracy on Cross-subject and Cross-setup of NTU RGB-D 120 data set is 84.1% and 86.0%, respectively. Compared with other algorithms,the more diversified and multi-scale action features are extracted, and the recognition accuracy of action categories is improved to a certain extent.

    Tools

    Get Citation

    Copy Citation Text

    Lei ZHANG, Guang-liang HAN. Action recognition algorithm based on multi-scale and multi-branch features[J]. Chinese Journal of Liquid Crystals and Displays, 2022, 37(12): 1614

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Research Articles

    Received: May. 25, 2022

    Accepted: --

    Published Online: Nov. 30, 2022

    The Author Email: Guang-liang HAN (hangl@ciomp.ac.cn)

    DOI:10.37188/CJLCD.2022-0176

    Topics