Chinese Optics, Volume. 17, Issue 1, 118(2024)

A hybrid network based on light self-limited attention for structured light phase and depth estimation

Xin-jun ZHU*, Hao-miao ZHAO, Hong-yi WANG, Li-mei SONG, and Rui-qun SUN
Author Affiliations
  • School of Artificial Intelligence, Tiangong University, Tianjin 300387, China
  • show less

    Phase retrieval and depth estimation are vital to three-dimensional measurement using structured light. Currently, conventional methods for structured light phase retrieval and depth estimation have limited efficiency and are lack of robustness in their results and so on. To improve the reconstruction effect of structured light by deep learning, we propose a hybrid network for structured light phase and depth estimation based on Light Self-Limited Attention (LSLA). Specifically, a CNN-Transformer hybrid module is constructed and integrated into a U-shaped structure to realize the advantages complementary of CNN and Transformer. The proposed network is experimentally compared with other networks in structured light phase estimation and structured light depth estimation. The experimental results indicate that the proposed network achieves finer detail processing in phase and depth estimation compared to other networks. Specifically, for structured light phase and depth estimation, its accuracy improves by 31% and 26%, respectively. Therefore, the proposed network improves the accuracy of deep neural networks in the structured light phase and depth estimation areas.

    Keywords
    Tools

    Get Citation

    Copy Citation Text

    Xin-jun ZHU, Hao-miao ZHAO, Hong-yi WANG, Li-mei SONG, Rui-qun SUN. A hybrid network based on light self-limited attention for structured light phase and depth estimation[J]. Chinese Optics, 2024, 17(1): 118

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category:

    Received: Apr. 14, 2023

    Accepted: --

    Published Online: Mar. 28, 2024

    The Author Email:

    DOI:10.37188/CO.2023-0066

    Topics