Abstract
Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with multi-modal datasets. In this paper, we propose an online multi-modal robust non-negative dictionary learning (OMRNDL) algorithm to overcome this deficiency. Notably, OMRNDL casts visual tracking as a dictionary learning problem under the particle filter framework and captures the intrinsic knowledge about the target from multiple visual modalities, e.g., pixel intensity and texture information. To this end, OMRNDL adaptively learns an individual dictionary, i.e., template, for each modality from available frames, and then represents new particles over all the learned dictionaries by minimizing the fitting loss of data based on M-estimation. The resultant representation coefficient can be viewed as the common semantic representation of particles across multiple modalities, and can be utilized to track the target. OMRNDL incrementally learns the dictionary and the coefficient of each particle by using multiplicative update rules to respectively guarantee their non-negativity constraints. Experimental results on a popular challenging video benchmark validate the effectiveness of OMRNDL for visual tracking in both quantity and quality.
Highlights
Visual tracking has been widely applied in many real-world tasks, such as video surveillance, but it poses significant challenges for computer vision community
Due to the efficacy of combining multi-modal features, we integrate the multi-modal features into dictionary learning and propose an online multi-modal robust non-negative dictionary learning (OMRNDL) method
The tracking procedures for visual tracking-based sparse representation can be categorized as the template update and particle representation
Summary
Visual tracking has been widely applied in many real-world tasks, such as video surveillance, but it poses significant challenges for computer vision community. Serious appearance variations such as illumination changes and cluttered backgrounds are obstacles to performing effective tracking in complex scenarios including multiple similar targets [1]. Various tracking techniques have been proposed to tackle these challenges, and recently, a strand of works that applies dictionary learning to visual tracking has achieved great success. Mei and Ling [2] originally proposed the L1 tracker (L1T) for robustly tracking the target under the particle filter framework. L1T and its variants [3, 4] suffer from one of the following drawbacks: 1)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.