Laser & Optoelectronics Progress, Volume. 62, Issue 4, 0400004(2025)

Review of Event Camera-Based Target Detection and Tracking Algorithms

Jiayu Qiu*, Yasheng Zhang, Yuqiang Fang, Pengju Li, and Kaiyuan Zheng
Author Affiliations
  • Space Engineering University, Beijing 101416, China
  • show less
    Figures & Tables(16)
    Sampling principle of dynamic vision sensor
    Circuit structure of ATIS pixel[9]. (a) Part A; (b) part B
    Circuit structure of DAVIS pixel[13]
    Corner detectors of event-based adaptive event threshold[22]
    Processing steps of graph neural network. (a) Event stream; (b) subsampling; (c) graph generation; (d) GNN; (e) prediction
    DMANet target detection framework[28]
    Framework of mixed frame-/event-driven pedestrian detection[36]
    Multistage recurrent attention model[38]
    Algorithmic framework for real-time optical flow estimation based on pulsed neural network[50]
    Framework of mixed frame-/event-driven pedestrian detection[55]. (a) Tracker framework; (b) frame attention module
    Framework for event-based object motion estimation for target tracking methods[60]
    A framework for target tracking algorithms with multimodal fusion[63]. (a) Overview of cross-domain feature integrator; (b) edge-attention block; (c) cross-domain modulation and selection block; (d) Bbox regressor; (e) classifier
    • Table 1. Comparison of the characteristics and uses of the three sensors

      View table

      Table 1. Comparison of the characteristics and uses of the three sensors

      TypeAdvantageDisadvantageApplication scenario
      DVSIt only outputs event information and is able to maintain the advantages of low latency, low power consumption and high dynamic rangeUnable to output any grayscale information, poor visibilityIt is suitable for situations with low requirements for visualization and high requirements for low latency, such as tracking, counting or motion monitoring of high-speed moving objects, real-time positioning
      ATISIt is capable of outputting event information and gray scale information of pixels where light intensity changes occur, with some visibilityIt is prone to exposure anomalies and information loss when the ambient brightness changes slowlyIt is suitable for frequent changes in environmental brightness and the existence of high-speed movement occasions, mainly used for real-time monitoring of products in industrial manufacturing
      DAVISIt can combine event information and grayscale information with good visibility, high temporal resolution, and has the ability to obtain detailed information about movementsIt can be affected by defects such as redundancy of information in APS cameras, low temporal resolution, and low dynamic rangeIt is suitable for occasions with high visualization requirements and small dynamic range, such as target recognition, target detection, tracking and localization of targets, especially in the field of robotics and unmanned vehicles
    • Table 2. Object detection algorithms based on event camera

      View table

      Table 2. Object detection algorithms based on event camera

      TypeCategoryReferenceYear/sourceContribution
      Pure stream of eventsAngle detection182015/Neural NetworksTransforming the traditional problem of corner point detection into a local velocity estimation problem
      192015/IEEEConsideration of spatial correlation between data
      202017/IEEERealization of the estimation of the velocity of a corner point
      212023/ICVRDesign of an Enhanced Representation of Time Surface Events
      222023/IEEEProposing a novel asynchronous corner point detector Arc*
      Linear Detection232016/BMVCDevelopment of an event-based Hough transform method
      242018/IEEEProposing weighted least squares fitting method based on iterative events
      Pure stream of eventsDeep learning252021/IEEEProposing fast graph construction method based on radius search method
      262022/ IEEEDesign of an asynchronous event-based graph neural networks (AEGNNs)
      272021/Neural NetworksProposing fully Convolutional Recurrent Neural Network Architecture
      282023/arXivProposing a dual-memory aggregation network (DMANet)
      292023/Peking UniversityProposing pulsed streaming target detection method based on asynchronous spatio-temporal memory network (ASTMNet)
      302023/ICRAProposing a lightweight spiking neural network that can separate events based on the speed of the corresponding objects
      Combination of event streams and traditional framesDeep learning312016/ IEEECombining CNN and Particle Filtering
      322018/ IEEETraining CNNs on pseudo-labels
      332018/ IEEEApplying existing deep learning methods to the iCub robotics platform
      34-352019/Frontiers in NeuroroboticsUsing three different event stream encoding methods and two fusion methods
      362019/IEEEProposing a confidence fusion method based on CNN detection results
      372022/IEEEApplying Asynet to maritime incident datasets for target recognition
      382023/IEEEExploring a multistage design approach and proposing Recurrent Vision Transformers for object detection
      392023/Chinese Academy of SciencesProposing an adaptive temporal resolution method for visualizing event information
    • Table 3. Event camera-based target tracking algorithm

      View table

      Table 3. Event camera-based target tracking algorithm

      TypeCategoryReferenceYear/sourceContribution
      Pure stream of eventsAngle detection402018/IEEEEstablishing correspondence between asynchronous detections in the event stream in real time
      412018/IEEEEstablishing local correspondences between corner events
      Pure stream of events422021/Zhejiang University of TechnologyAdding speed constraints to the ACE tracker
      Clustering432012/IEEEGaussian mixture modeling to form event clusters for object characterization
      442017/IEEECombination of stereo matching and object tracking
      452018/IEEEmean-drift-clustering-based approach for event labeling
      462021/IEEEApplying SVM to event cameras,combining local and global sliding window search
      Optical flow estimation472012/Neural NetworksReplacing the gray value gradient by comparing the instantaneous activity values of neighboring active pixels
      482020/IEEEProposing an unsupervised learning algorithm for optical flow estimation of sparse event data
      492023/IEEEProposing incremental full flow estimation from sparse normal flow based measurements
      502023/IEEEProposing a real-time estimation algorithm for neuromorphic optical flow based on impulse neural networks
      Deep learning512019/Frontiers in NeuroroboticsExtraction of features after coding the event stream based on rate coding
      522022/Advances in Neural Information Processing SystemsDownsampling strategy to mine key events and embed the irregular spatio-temporal information of key events into a high-dimensional feature space
      Combination of event streams and traditional framesFeature tracking532016/IEEEProposing a feature trajectory tracking algorithm that combines event streams and grayscale maps
      542017/IEEEAn expectation maximization (EM) algorithm is used to quantify the probabilistic correlation between feature point sets and event data
      552023/IEEEA novel frame-attention module is introduced
      Correlation filtering422021/Zhejiang University of TechnologyKernel-correlation filtering is introduced
      Particle filtering562017/IEEEMulti-hypothesis filtering technology is introduced
      Clustering572018/CHREOCClustering based on spatial relationships
      582019/ICCSHierarchical clustering
      592021/ICCCAlignment and foreground enhancement models are introduced
      Deep learning602020/AAAIDesigning a temporal surface representation of TSLTD
      612021/Dalian University of TechnologyChannel Attention Mechanism added to Feature Fusion Module is introduced
      622022/IEEESTNET is proposed to dynamically extract and fuse information from the temporal and spatial domains
      632021/IEEEDiscretizing asynchronous events into time slices accumulated in conventional frames and designing a cross-domain feature integrator
      642023/IEEEDesign of an event-guided cross-modal alignment and crossover based module
      652023/arXivCross-modal hierarchical knowledge distillation scheme is introduced
      662024/IEEEDevelopment of a cross-modal converter for bimodal information interaction
      672022/SensorsAdaptive strategy to adjust the spatial and temporal domain of event data for event frame reconstruction
    • Table 4. Comparison of event camera-based visual tracking datasets

      View table

      Table 4. Comparison of event camera-based visual tracking datasets

      DatasetYearVideosFramesClassAttributesResolutionAimPublic
      VOT-DVS201660240×180Eval
      TD-DVS201677240×180Eval
      Ulster201619000240×180Eval×
      EED20187234240×180Eval
      FE1082021108208672214346×260Train&Eval
      VisEvent202182037112717346×260Train&Eval
    Tools

    Get Citation

    Copy Citation Text

    Jiayu Qiu, Yasheng Zhang, Yuqiang Fang, Pengju Li, Kaiyuan Zheng. Review of Event Camera-Based Target Detection and Tracking Algorithms[J]. Laser & Optoelectronics Progress, 2025, 62(4): 0400004

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Reviews

    Received: Apr. 10, 2024

    Accepted: Jun. 17, 2024

    Published Online: Feb. 10, 2025

    The Author Email:

    DOI:10.3788/LOP241073

    CSTR:32186.14.LOP241073

    Topics