Laser & Optoelectronics Progress, Volume. 56, Issue 14, 141008(2019)
Convolutional Neural Network-Based Dimensionality Reduction Method for Image Feature Descriptors Extracted Using Scale-Invariant Feature Transform
Since local feature descriptors extracted from an image using the traditional scale-invariant feature transform (SIFT) method are 128-dimensional vectors, the matching time is too long, which limits their applicability in some cases such as feature point matching based on the three-dimensional reconstruction. To tackle this problem, a SIFT feature descriptor dimensionality reduction method based on a convolutional neural network is proposed. The powerful learning ability of the convolutional neural network is used to realize the dimensionality reduction of SIFT feature descriptors while maintaining their good affine transformation invariance. The experimental results demonstrate that the new feature descriptors obtained using the proposed method generalize well against affine transformations, such as rotation, scale, viewpoint, and illumination, after reducing their dimensionality to 32. Furthermore, the matching speed of the feature descriptors obtained using the proposed method is nearly five times faster than that of the SIFT feature descriptors.
Get Citation
Copy Citation Text
Honghao Zhou, Weining Yi, Lili Du, Yanli Qiao. Convolutional Neural Network-Based Dimensionality Reduction Method for Image Feature Descriptors Extracted Using Scale-Invariant Feature Transform[J]. Laser & Optoelectronics Progress, 2019, 56(14): 141008
Category: Image Processing
Received: Jan. 14, 2019
Accepted: Feb. 21, 2019
Published Online: Jul. 12, 2019
The Author Email: Du Lili (ylqiao@aiofm.ac.cn), Qiao Yanli (lilydu@aiofm.ac.cn)