scholarly journals A Spherical Model Based Keypoint Descriptor and Matching Algorithm for Omnidirectional Images

2014 ◽  
Vol 6 ◽  
pp. 154376 ◽  
Author(s):  
Guofeng Tong ◽  
Xue Chen ◽  
Ning Ye

Omnidirectional images generally have nonlinear distortion in radial direction. Unfortunately, traditional algorithms such as scale-invariant feature transform (SIFT) and Descriptor-Nets (D-Nets) do not work well in matching omnidirectional images just because they are incapable of dealing with the distortion. In order to solve this problem, a new voting algorithm is proposed based on the spherical model and the D-Nets algorithm. Because the spherical-based keypoint descriptor contains the distortion information of omnidirectional images, the proposed matching algorithm is invariant to distortion. Keypoint matching experiments are performed on three pairs of omnidirectional images, and comparison is made among the proposed algorithm, the SIFT and the D-Nets. The result shows that the proposed algorithm is more robust and more precise than the SIFT, and the D-Nets in matching omnidirectional images. Comparing with the SIFT and the D-Nets, the proposed algorithm has two main advantages: (a) there are more real matching keypoints; (b) the coverage range of the matching keypoints is wider, including the seriously distorted areas.

2020 ◽  
Vol 12 (15) ◽  
pp. 2390 ◽  
Author(s):  
Fan Shi ◽  
Fang Qiu ◽  
Xiao Li ◽  
Yunwei Tang ◽  
Ruofei Zhong ◽  
...  

In recent years, satellites capable of capturing videos have been developed and launched to provide high definition satellite videos that enable applications far beyond the capabilities of remotely sensed imagery. Moving object detection and moving object tracking are among the most essential and challenging tasks, but existing studies have mainly focused on vehicles. To accurately detect and then track more complex moving objects, specifically airplanes, we need to address the challenges posed by the new data. First, slow-moving airplanes may cause foreground aperture problem during detection. Second, various disturbances, especially parallax motion, may cause false detection. Third, airplanes may perform complex motions, which requires a rotation-invariant and scale-invariant tracking algorithm. To tackle these difficulties, we first develop an Improved Gaussian-based Background Subtractor (IPGBBS) algorithm for moving airplane detection. This algorithm adopts a novel strategy for background and foreground adaptation, which can effectively deal with the foreground aperture problem. Then, the detected moving airplanes are tracked by a Primary Scale Invariant Feature Transform (P-SIFT) keypoint matching algorithm. The P-SIFT keypoint of an airplane exhibits high distinctiveness and repeatability. More importantly, it provides a highly rotation-invariant and scale-invariant feature vector that can be used in the matching process to determine the new locations of the airplane in the frame sequence. The method was tested on a satellite video with eight moving airplanes. Compared with state-of-the-art algorithms, our IPGBBS algorithm achieved the best detection accuracy with the highest F1 score of 0.94 and also demonstrated its superiority on parallax motion suppression. The P-SIFT keypoint matching algorithm could successfully track seven out of the eight airplanes. Based on the tracking results, movement trajectories of the airplanes and their dynamic properties were also estimated.


2012 ◽  
Vol 239-240 ◽  
pp. 1232-1237 ◽  
Author(s):  
Can Ding ◽  
Chang Wen Qu ◽  
Feng Su

The high dimension and complexity of feature descriptor of Scale Invariant Feature Transform (SIFT), not only occupy the memory spaces, but also influence the speed of feature matching. We adopt the statistic feature point’s neighbor gradient method, the local statistic area is constructed by 8 concentric square ring feature of points-centered, compute gradient of these pixels, and statistic gradient accumulated value of 8 directions, and then descending sort them, at last normalize them. The new feature descriptor descend dimension of feature from 128 to 64, the proposed method can improve matching speed and keep matching precision at the same time.


2020 ◽  
Vol 9 (1) ◽  
pp. 2711-2713

Image identification and matching is one of the very difficult assignment in different areas of mainframe vie w. Scale-Invariant Feature Transform is an algorithm to perceive and represent specific features in portryals to further use them as an image matching criteria. In this paper, the extracted SIFT matching features are against various image distortions such as rotation, scaling, fisheye and motion distortion are evaluated and false and true positive rates for a large number of image pairs are calculated and presented.


2018 ◽  
Vol 7 (2.8) ◽  
pp. 42
Author(s):  
D Rajasekhar ◽  
T Jayachandra Prasad ◽  
K Soundararajan

Feature detection and image matching constitutes two primary tasks in photogrammetric and have multiple applications in a number of fields. One such application is face recognition. The critical nature of this application demands that image matching algorithm used in recognition of features in facial recognition to be robust and fast. The proposed method uses affine transforms to recognize the descriptors and classified by means of Bayes theorem. This paper demonstrates the suitability of the proposed image matching algorithm for use in face recognition appli-cations. Yale facial data set is used in the validation and the results are compared with SIFT (Scale Invariant Feature Transform) based face recognition approach.


Author(s):  
M. B.Daneshvar

This paper presents an enhanced method for extracting invariant features from images based on Scale Invariant Feature Transform (SIFT). Although SIFT features are invariant to image scale and rotation, additive noise, and changes in illumination but we think this algorithm suffers from excess keypoints. Besides, by adding the hue feature, which is extracted from combination of hue and illumination values in HSI colour space version of the target image, the proposed algorithm can speed up the matching phase. Therefore, we proposed the Scale Invariant Feature Transform plus Hue (SIFTH) that can remove the excess keypoints based on their Euclidean distances and adding hue to feature vector to speed up the matching process which is the aim of feature extraction. In this paper we use the difference of hue features and the Mean Square Error (MSE) of orientation histograms to find the most similar keypoint to the under processing keypoint. The keypoint matching method can identify correct keypoint among clutter and occlusion robustly while achieving real-time performance and it will result a similarity factor of two keypoints. Moreover removing excess keypoint by SIFTH algorithm helps the matching algorithm to achieve this goal.


2016 ◽  
Vol 14 (1) ◽  
pp. 172988141668270
Author(s):  
XiangShao Liu ◽  
Shangbo Zhou ◽  
Hua Li ◽  
Kun Li

In this article, a bidirectional feature matching algorithm and two extended algorithms based on the priority k-d tree search are presented for the image registration using scale-invariant feature transform features. When matching precision of image registration is below 50%, the discarding wrong match performance of many robust fitting methods like Random Sample Consensus (RANSAC) is poor. Therefore, improving matching precision is a significant work. Generally, a feature matching algorithm is used once in the image registration system. We propose a bidirectional algorithm that utilizes the priority k-d tree search twice to improve matching precision. There are two key steps in the bidirectional algorithm. According to the case of adopting the ratio restriction of distances in the two key steps, we further propose two extended bidirectional algorithms. Experiments demonstrate that there are some special properties of these three bidirectional algorithms, and the two extended algorithms can achieve higher precisions than previous feature matching algorithms.


Sign in / Sign up

Export Citation Format

Share Document