Abstract
Multi-scale/orientation local image analysis methods are valuable tools for obtaining highly distinctive image-based representations. Very often, these features are generated from the responses of a bank of linear filters corresponding to different scales and orientations. Naturally, as the number of filters increases, so does the feature dimensionality. Further processing is often feasible only when dimensionality reduction is performed by subspace learning techniques, such as Principal Component analysis (PCA) or Linear Discriminant Analysis (LDA). The major problem stems from the fact that as the number of features increases, so does the computational complexity of these methods which, in turn, limits the number of scales and orientations examined. In this paper, we show how linear subspace analysis on features generated by the response of linear filter banks can be efficiently re-formulated such that complexity does not depend on the number of filters used. We describe computationally efficient and exact versions of PCA while the extension to other subspace learning algorithms is straightforward. Finally, we show how the proposed methods can boost the performance of algorithms for appearance based tracking algorithm.
Highlights
A significant amount of research in computer vision has revolved around providing efficient solutions to the following problem: given samples of a high-dimensional space estimate a low-dimensional space which preserves the intrinsic structure of the input data
The necessity for dimensionality reduction in computer vision becomes more evident if we consider standard feature extraction techniques based on multi-scale/orientation local image analysis
Inspired by [1], we show how linear subspace analysis on features generated by the response of linear filter banks can be efficiently re-formulated such that complexity does not depend on the number of filters used
Summary
A significant amount of research in computer vision has revolved around providing efficient solutions to the following problem: given samples of a high-dimensional space estimate a low-dimensional space which preserves the intrinsic structure of the input data. The necessity for dimensionality reduction in computer vision becomes more evident if we consider standard feature extraction techniques based on multi-scale/orientation local image analysis. Inspired by [1], we show how linear subspace analysis on features generated by the response of linear filter banks can be efficiently re-formulated such that complexity does not depend on the number of filters used. Our approach touches upon classical results from linear algebra previously used for the efficient implementation of PCA and the efficient calculation of inner products in the Fourier domain [15] We show how these results can be used in order to formulate computationally efficient and exact versions of PCA. We show how to apply this framework in order to formulate an appearance based tracking algorithm using arbitrarily many linear filter responses [17]
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.