Advanced Photonics Nexus, Volume. 4, Issue 2, 026010(2025)

Large-scale single-pixel imaging and sensing

Lintao Peng1, Siyu Xie1, Hui Lu1, and Liheng Bian1,2,3、*
Author Affiliations
  • 1Beijing Institute of Technology, MIIT Key Laboratory of Complex-Field Intelligent Sensing, Beijing, China
  • 2Beijing Institute of Technology, Guangdong Province Key Laboratory of Intelligent Detection in Complex Environment of Aerospace, Land and Sea, Zhuhai, China
  • 3Yangtze Delta Region Academy of Beijing Institute of Technology, Jiaxing, China
  • show less
    Figures & Tables(11)
    Overview of the SPIS technique. (a) The optical setup of the SPIS technique. The structured illumination was generated using a DMD and a white-light source. A single-pixel detector was used to collect the light reflected from the target scene. The collected 1D measurements were digitized and then reshaped into 2D measurements. (b) In SPIS, we scan and sample the scene using small-sized optimized patterns, which achieve higher sampling performance with an order of magnitude fewer pattern parameters. The 2D measurements were then fed into the encoder to extract high-dimensional semantic features, and the features were sent to the task-specific plug-and-play decoder to complete large-scale SPI or image-free sensing. (c) The transformer-based encoder and UDL function can guide SPIS to pay more attention to the target area with more details in the scene, thus extracting high-dimensional semantic features that are effective for imaging and sensing. (d) The existing state-of-the-art SPI method ReconNet15" target="_self" style="display: inline;">15 cannot reconstruct clear images at a sampling rate of 3% and a resolution of 1024×1024, but our SPIS can still reconstruct high-quality images in this case.
    SPIS network structure. (a) Overview of the SPIS technique. (b) The transformer-based encoder. (c) The decoder for large-scale imaging and image-free segmentation. The decoder consists of multiple upsampling convolution blocks, each of which consists of multiple convolutional layers and one upsampling layer. (d) The decoder for image-free object detection.
    Two-step training strategy of the SPIS network. (a) The overview of the two-step training strategy for imaging and image-free segmentation. In the first stage, the network estimates both the output result and the uncertainty values. In the second stage, the uncertainty values are used to generate a spatially adaptive weight to guide the network to prioritize the pixels in the texture-rich regions and edge regions. (b) The overview of the two-step training method for image-free single-pixel object detection. (c) Ablation study of the UDL loss function. “Step 1” represents the results output by SPIS that has not been trained by UDL, and “step 2” represents the results output by SPIS that has been trained by UDL.
    SPI experiment. (a) The statistical results of pattern comparison, noise robustness, and ablation study for uncertainty loss functions. (b) The proof-of-concept setup for large-scale SPI and image-free single-pixel sensing. (c) The visualization results of large-scale SPI on 3D scenes at a sampling rate of 3% and a resolution of 1024×1024. (d) The visualization results of large-scale SPI on a resolution board at a sampling rate of 3% and a resolution of 1024×1024. Step 1 marks the results output by SPIS after the first training stage, which has not been trained by the UDL. Step 2 denotes the reconstruction results of SPIS with UDL.
    Image-free single-pixel segmentation experiment. (a) Experimental results of segmentation performance of the five methods involved in the comparison and our proposed image-free single-pixel image segmentation method at different sampling rates. (b) Visualization results of image-free single-pixel image segmentation performance comparison experiments at different sampling rates (SRs).
    Image-free single-pixel object detection experiment. (a) Statistic results of pattern sampling performance comparison and noise interference experiments. (b) Visualization results of image-free single-pixel object detection. The “min” and “max” represent the relative coordinates of the upper left corner and lower right corner of the target bounding box, respectively. To better demonstrate the detection results, we visualized the output of SPIS on the input scene.
    Sampling process of small-sized patterns. We take a 2×2 small-sized pattern to scan and sample an 8×8 scene as an example to show the sampling process of our small-sized optimized patterns. We embed the nonoverlapping 2×2 small-sized patterns in multiple zero-initialized 8×8 full-sized patterns and quickly switch between full-sized patterns.
    • Table 1. Quantitative results of large-scale SPI comparisona

      View table
      View in Article

      Table 1. Quantitative results of large-scale SPI comparisona

      SR = 3%SR = 5%SR = 7%SR = 10%SR = 15%
      PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
      DCAN2218.170.5118.860.5419.630.5920.030.6220.630.63
      GAN2318.820.5819.350.6120.190.6520.610.6821.440.72
      ReconNet1519.890.6220.010.6720.900.7221.120.7522.410.75
      Ours24.130.8324.640.8625.330.8625.690.8826.170.89
    • Table 2. Quantitative performance comparison of different image-free segmentation methods at different SRs. The methods for comparison include the image-free segmentation methods using the Hadamard pattern,22 random pattern,20 and optimized full-sized pattern (“Optimized”),24 the imaging-first15-sensing-later40 method (“Imaging”), and the UNet-based image-free segmentation method (SPS).25 In the “Imaging” method, we used the method in Ref. 15 for imaging and UNet40 for semantic segmentation.

      View table
      View in Article

      Table 2. Quantitative performance comparison of different image-free segmentation methods at different SRs. The methods for comparison include the image-free segmentation methods using the Hadamard pattern,22 random pattern,20 and optimized full-sized pattern (“Optimized”),24 the imaging-first15-sensing-later40 method (“Imaging”), and the UNet-based image-free segmentation method (SPS).25 In the “Imaging” method, we used the method in Ref. 15 for imaging and UNet40 for semantic segmentation.

      SR = 3%SR = 5%SR = 7%SR = 10%SR = 15%
      DicemIoUDicemIoUDicemIoUDicemIoUDicemIoU
      Hadamard220.1480.0710.2180.0990.1510.0720.2240.0950.2370.096
      Random200.2070.0920.1210.0610.1230.0620.1620.0770.1730.074
      Optimized240.7160.6340.7210.6520.7390.6490.7630.6840.7930.658
      Imaging15,400.6380.4080.6900.4660.7150.5240.7130.5460.6740.522
      SPS250.7050.6280.7170.6590.7290.6730.7430.6910.7850.648
      Ours0.8210.7010.8230.7040.8840.7930.8890.8030.9210.854
    • Table 3. Comparison of data throughput and running speed between SPIS and the other existing object detection methods.

      View table
      View in Article

      Table 3. Comparison of data throughput and running speed between SPIS and the other existing object detection methods.

      MethodData throughput (Mbps)Accuracy (mAP)Time (ms)Speed (fps)
      R-CNN411.28058.51%2005.0
      Faster R-CNN372.58673.22%9910.1
      YOLO276.73278.64%3826.3
      SSD424.91579.63%5219.2
      DETReg433.09881.16%8312.1
      SPIS (ours)0.39682.41%1662.5
    • Table 4. Experimental detection results of SPIS at 5% sampling rate on the testing set of Pascal VOC2012.

      View table
      View in Article

      Table 4. Experimental detection results of SPIS at 5% sampling rate on the testing set of Pascal VOC2012.

      Object class“horse”“cat”“bicycle”“sofa”“boat”
      Accuracy92%89%78%86%84%
      Object class“cow”“sheep”“person”“bird”“table”
      Accuracy81%76%76%86%74%
    Tools

    Get Citation

    Copy Citation Text

    Lintao Peng, Siyu Xie, Hui Lu, Liheng Bian, "Large-scale single-pixel imaging and sensing," Adv. Photon. Nexus 4, 026010 (2025)

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Research Articles

    Received: Aug. 28, 2024

    Accepted: Jan. 15, 2025

    Published Online: Feb. 27, 2025

    The Author Email: Liheng Bian (bian@bit.edu.cn)

    DOI:10.1117/1.APN.4.2.026010

    CSTR:32397.14.1.APN.4.2.026010

    Topics