Acta Optica Sinica (Online), Volume. 2, Issue 16, 1611001(2025)

Advances in Wide-Field High-Resolution Imaging Systems (Invited)

Lianjie Li, Jianqi Zi, and Liheng Bian*
Author Affiliations
  • State Key Laboratory of Communication, Navigation and Surveillance (CNS)/Air TrafficManagement (ATM), Beijing Institute of Technology, Beijing 100081, China
  • show less
    Figures & Tables(38)
    Principle of wide-field high-resolution imaging systems. (a) Single-device scanning systems, including rotational and translational scanning systems; (b) multi-chip mosaic system; (c) multi-camera array systems, including planar, curved, and unstructured configurations; (d) multi-scale imaging system
    GigaPan panoramic imaging system and stitching panoramagram[18]
    Single-device scanning imaging system[19]. (a) PixOrb mount; (b) Meade LX200 telescope mount; (c) panoramic image of Seattle skyline
    Jilin-1 platform 02A02 satellite and captured image (Doha, Qatar) [4]
    Large-format scanning camera developed by Microsoft Research Asia[20]. (a) Structural configuration of the camera assembly; (b) application in cultural heritage preservation (a high-fidelity imaging system for digitizing historical artifacts)
    MOA-cam3 system architecture and synthesized images[21]. (a) Photo of MOA-cam3; (b) photo of CCD array mounted on the AIN plate; (c) synthesized 10-chip images of the Large Magellanic Cloud taken by MOA-cam3
    Kepler space telescope CCD sensor array and observation field[22]. (a) Kepler's image sensor array; (b) a photo taken by Kepler telescope with two regions of interest outlined, and celestial north is towards the lower left corner
    Focal plane of the LSST[25]
    JWST NIRCam instrumentation [28]. (a) NIRCam near-infrared detector; (b) NIRCam mid-infrared detector; (c)(d) near-infrared detector array configuration
    ARGUS-IS system and its imaging principle[30]. (a) Schematic diagram of the wide field-of-view imaging principle of ARGUS-IS system; (b) ARGUS-IS sensor system; (c) the system consists of four co-aligned imaging modules sharing a common primary lens, each module contains 92 sensor chips arranged in a checkerboard pattern to minimize overlap of inactive regions between adjacent pixels, and this complementary arrangement between active and inactive regions enables seamless full coverage of the target area through image stitching
    Schematics of multi-aperture imaging technology[31]. (a) Conventional single-aperture imaging system; (b) 3×3 multi-aperture imaging system; (c) 5×5 multi-aperture multi-scale imaging architecture
    Planar camera array system. (a) Stanford's compact planar configuration employing telephoto lenses for high-resolution acquisition[32-33]; (b) Tokyo university's TransCAIP[34] (translational camera array imaging platform)
    Planar multi-device imaging system[35]
    Wide-angle imaging configuration of bionic dual-mode compound vision system[36-37]
    Mechanical structure of Panoptic and omnidirectional image[39]. (a) Panoptic system with 7 layers and 13 cm diameter hemisphere structure with 30 embedded cameras; (b) Panoptic system's omnidirectional view represented on a sphere
    OMNI-R and GigaEye-1 systems and corresponding imaging results[41-42]. (a) OMNI-R system; (b) omni-directional single frame obtained by OMNI-R system at 2.16×107 pixel resolution and 30 frame/s; (c) GigaEye-1 system; (d) static scene imaging performance of GigaEye-1; (e) dynamic scene imaging performance of GigaEye-1
    GigaEye-2 system and imaging effect[43]. (a) 3D model of the GigaEye-2 system; (b) physical diagram of the GigaEye-2 system; (c) a high-resolution panoramic frame with three selected regions of interest shown in full resolution
    Artificial compound eye system diagram and imaging effect[45]. (a) Artificial compound eye system; (b) system imaging effect
    Evryscope system and imaging effect[46-47]. (a) Physical system; (b) cutaway view of the Evryscope showing the internal structure of the telescope; (c) a full Evryscope image; (d) partial detail covering approximately 1% of the Evryscope's FOV
    Artificial compound eye system and imaging effect[49]
    Geometrically curved artificial compound eye imaging system[48,50]. (a) Prototype implementation; (b) primary image acquired by the system; (c) computationally stitched output with restoration enhancement
    Structure of bio-inspired compound eye optical systems[53-54]. (a) First-generation system; (b) second-generation compact system
    Prototype implementations and imaging characterization[55-56]. (a) Composite biomimetic imaging prototype; (b) multi-scale artificial compound eye system; (c) image stitching results [dashed regions correspond to C12, C21, C23, C32 coverage in Fig. 23(a)]; (d) central FOV super-resolution analysis comparing four super-resolution images with original image; (e) mosaiced multi-resolution images from far-field experiments
    Imaging system and effect with geometrically-curved, non-uniformly distributed camera array[57]. (a) Computational model; (b) engineered prototype; (c) acquired images demonstrating 150°×40° coverage, featuring zoomed regions at different resolution levels (LR: low resolution; HR: high resolution; SR: super-resolved reconstruction)
    The UnstructuredCam[5]. (a) UnsturctureCam module, consisting of a global camera and multiple local cameras; (b) schematic of system consisting of multiple subarrays; (c) ​gigapixel-scale video frame captured by the array, the red and blue frames on the top left represent the distributions of the global and local cameras
    Anktech's computational imaging products and representative applications. (a) Mantis series; (b) Bumblebee series; (c) exemplary deployment cases
    Gigapixel system and imaging performance[2]. (a) System photo; (b) a 1.6 gigapixel image whose resolution is 65000 pixel×25000 pixel, and the scene occupies a 104°×40° FOV
    AWARE-2 system architecture and imaging effect[59]. (a) An exploded view of the AWARE-2 prototype, showcasing the primary components [the objective lens (lower left), the dome (gray in the center), and microcamera array (green barrels and purple lenses)]; (b) microcamera optics in the mounting dome; (c) primary optical assembly of the system; (d) AWARE-2 system during assembly; (e) sample imaging results captured by the AWARE-2 system
    AWARE-10 system architecture and imaging effect[60]. (a) Three-dimensional rendering of the AWARE-10 system, where the purple ball in the front is the objective, and the green cylinders are microcameras; (b) internal assembly of the AWARE-10 system, including the G1 and G2 modules with microcameras, where the G2 module and mounting brackets are integrated with a recirculating water-cooling loop for thermal management; (c) magnified view of acquired imagery demonstrating spatial resolution; (d) field deployment configuration during outdoor testing; (e)‒(g) maritime targets imaged during open-water experiments, including a duck boat (4 m characteristic broadside dimension), a fishing boat (7 m), and a crab boat (10 m)
    AWARE-40 wide-area imaging system and acquired results[61]. (a) Internal configuration of the AWARE-40 prototype; (b) field deployment photograph of the AWARE-40 system; (c) panoramic image captured by the AWARE-40 system; (d)‒(i) magnified region-of-interest (ROI) demonstrating spatial resolution
    Three-dimensional optical architecture of a concentric multiscale aerial photogrammetric system[65]
    Distributed zoom-enabled concentric multiscale optical system and imaging effect[67]. (a) Optical path design of the system; (b) prototype model schematic; (c) photograph of the physical system; (d) independent sensor imaging effect; (e) heterogeneous-scale image synthesis results
    Integrated optical system physical assembly and tiled imaging effect[69]. (a) Integrated optical system architecture; (b) 3×3 sub-aperture stitching outcomes
    Heterogeneous-scale wide-FOV computational imaging system structure and imaging effect[71]. (a)(b) Objective lens group; (c) curved mounting substrate; (d) camera array housing; (e) unit camera deployment scheme; (f) multi-resolution output imagery
    Snapshot hyperspectral volumetric microscopy (SHVM) and real-time ultra-large-scale high-resolution microscopic imaging (RUSH) systems[7,72]. (a) SHVM system architecture; (b) RUSH system with in vivo brain-wide imaging of a mouse
    Bio-inspired hybrid imaging system combining fisheye and compound-eye architectures[74]. (a) Optomechanical design; (b) system setup with an array of 1×9 microcameras captured at distance of 7.4 km; (c) system setup with an array of 3×3 microcameras captured at distance of 7.4 km; (d) imaging effect with an array of 1×9 microcameras; (e) panoramic image from a 3×3 array system
    Schematic diagrams of cascade optical systems and corresponding light paths[76-77]. (a) Cascaded optical imaging system; (b) multi-channel imaging optical path of the cascaded system; (c) optical path with concentric spherical lenses; (d) folded objective subsystem; (e) single-channel imaging optical path of the folded cascaded system; (f) multi-channel folded cascaded optical architecture
    • Table 1. Comparative analysis of wide-field high-resolution imaging systems

      View table
      View in Article

      Table 1. Comparative analysis of wide-field high-resolution imaging systems

      Imaging techniquePrincipleFundamental limitation (inherent to imaging mechanism/physical law)Application
      Single-device scanningTemporal-to-spatial tradeoff: sequential acquisition via rotational/translational scanning

      Temporal resolution limit: instantaneous acquisition restricted to a sub-region, with acquisition time constrained by mechanical motion and shutter speed.

      Dynamic artifact: scene dynamics induce spatial misregistration and ghosting, precluding strict spatiotemporal consistency.

      SBP underutilization: inactive detector state during scanning; reduced system efficiency.

      Mechanical instability: micro-vibrations and thermal drift amplify to pixel-level misregistration; inherently structural

      Static scene imaging, cultural heritage preservation
      Multi-chip mosaicEffective aperture expansion: increased imaging area via chip mosaicking

      Pixel discontinuity & fill factor: inter-chip gaps/dead zones persist despite optical compensation.

      Beam-splitting loss: beam-splitters/secondary optics reduce optical throughput, introduce aberrations, and decrease SNR.

      Focal plane curvature: planar arrays exacerbate off-axis aberrations in large-format designs with inherent curved focal planes.

      Thermal management & synchronization: nanosecond-scale sync and thermal stability required for multi-sensor arrays, causing scaling-dependent noise and clock drift.

      Astronomy, space observation
      Multi-camera arrayParallel aperture acquisition: simultaneous multi-perspective capture

      Array baseline geometry error: minute mounting errors induce parallax distortion; necessitating high-dimensional calibration and real-time correction; susceptibility to thermo-mechanical drift.

      Inter-camera dispersion & PSF variance: spectral response and PSF variations between lenses; incomplete compensation causes resolution fluctuations at stitching seams.

      Communication & timing bottleneck: sub-microsecond synchronization for numerous cameras; photon arrival jitter limits high-bandwidth links over long distances.

      Algorithmic complexity scaling: computational cost of stitching algorithms scales as ON²) or higher with camera count N; real-time performance constrained by computational resources and power

      Surveillance, autonomous systems
      Multi-scale imagingHybrid optics: coaxial/off-axial primary and relay optics; multi-resolution allocation

      Primary optics coherence: primary lens aberrations superimpose non-uniformly across sub-fields; requiring global coaxial alignment.

      Relay optics cropping: insufficient FOV overlap necessitates sacrificing effective aperture or F#; reduced sub-camera SNR and resolution.

      Multi-layer aberration coupling: interdependence of sub-lens aberrations, primary lens off-axis aberrations, and stitching errors; NP-hard system optimization.

      SBP limitation: system SBP bounded by primary lens aperture and entrance pupil, regardless of sub-camera count; inherent physical saturation

      Remote sensing, biomedical imaging
    Tools

    Get Citation

    Copy Citation Text

    Lianjie Li, Jianqi Zi, Liheng Bian. Advances in Wide-Field High-Resolution Imaging Systems (Invited)[J]. Acta Optica Sinica (Online), 2025, 2(16): 1611001

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Computational Optics

    Received: Apr. 24, 2025

    Accepted: Jun. 16, 2025

    Published Online: Aug. 7, 2025

    The Author Email: Liheng Bian (bian@bit.edu.cn)

    DOI:10.3788/AOSOL250453

    CSTR:32394.14.AOSOL250453

    Topics