Abstract

본 논문에서는 FAST(Features from Accelerated Segment Test) 특징점 검출기와 SIFT(Scale Invariant Feature Transform) 특징점 서술자(descriptor)를 사용하여 시점 변화에 강인한 특징점 정합 기법을 제안한다. 기존의 FAST 기법은 영상의 에지 부분을 따라서 불필요하게 특징점을 많이 추출하게 되는데 이러한 단점을 주곡률(principal curvatures)을 적용하여 개선한다. 추출된 특징점을 SIFT 서술자를 통해 기술하고 시점이 다른 두 영상으부터 구해진 정합쌍에 RANSAC(RANdom SAmple Consensus) 기법을 통하여 호모그래피(homography)를 계산한다. 시점 변화에 강인한 특징점 정합을 위해서 기준 영상의 특징점들을 호모그래피 변환을 통해 변경된 좌표와 시점이 다른 영상의 특징점 좌표간의 유클리디언(Euclidean) 거리를 통해 정합쌍을 분류한다. 같은 물체나 장소에 대해 시점이 변화된 여러 영상에 대한 실험을 통해서 제안하는 정합 기법이 적은 계산량으로 기존의 특징점 정합 기법보다 우수한 성능을 보여주는 것을 확인하였다. In this paper, we propose a new feature matching algorithm which is robust to the viewpoint change by using the FAST(Features from Accelerated Segment Test) feature detector and the SIFT(Scale Invariant Feature Transform) feature descriptor. The original FAST algorithm unnecessarily results in many feature points along the edges in the image. To solve this problem, we apply the principal curvatures for refining it. We use the SIFT descriptor to describe the extracted feature points and calculate the homography matrix through the RANSAC(RANdom SAmple Consensus) with the matching pairs obtained from the two different viewpoint images. To make feature matching robust to the viewpoint change, we classify the matching pairs by calculating the Euclidean distance between the transformed coordinates by the homography transformation with feature points in the reference image and the coordinates of the feature points in the different viewpoint image. Through the experimental results, it is shown that the proposed algorithm has better performance than the conventional feature matching algorithms even though it has much less computational load.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.