Abstract

The local contexts define the target and its surrounding background within a constrained region, and have been proved useful for visual tracking, but how to adaptively employ them for building robust models remains challenging. By using the spatial weight maps, the correlation filter (CF) methods with spatial regularization provide an alternative to exploit the local contexts for appearance modeling. However, they generally utilize naive spatial weight map functions, and fail to flexibly regulate the effects of the target and background on model learning, thereby restricting the tracking performance. In this paper, we address these issues by presenting an adaptive multiple contexts correlation filter (AMCCF) framework. In particular, a novel sigmoid spatial weight map is first proposed to control the impacts of local contexts for learning more effective CF models. Based on this, different levels of local contexts (multiple contexts) are further modeled by incorporating the spatial weight maps with different parameters into multiple CF models. To adaptively utilize the local contexts on the tracking stage, the minimal weighted confidence margin loss function with a weight prior constraint is adopted for jointly estimating the target position and adaptive fusion weights of response maps from different CF models. To validate the proposed method, extensive experiments are conducted on four tracking benchmarks. The results show that our AMCCF can adaptively leverage the local contexts for robust tracking, and performs favorably against the state-of-the-art trackers.

Highlights

  • Given only the initial position on the first frame, the goal of visual tracking is to detect the trajectory of a target with large appearance variations, such as deformation, fast motion and scale changes

  • We present a novel adaptive multiple contexts correlation filter (AMCCF) framework to exploit the potentials of local contexts for visual tracking

  • Since the parameters in Eqn (6) can flexibly control the impacts of both target and surrounding background regions on CF model learning, we develop a series of sigmoid spatial weight maps with different parameters, and further employ them to train multiple CF models with different levels of local contexts

Read more

Summary

INTRODUCTION

Given only the initial position on the first frame, the goal of visual tracking is to detect the trajectory of a target with large appearance variations, such as deformation, fast motion and scale changes. The local context may be harmful when the surrounding background regions are encountered with severe appearance changes It remains a challenging issue on how to adaptively leverage the local context information for building robust tracking models. By using the spatial weight maps, spatial regularization can enforce large penalties on CF coefficients outside the target during training, and the CF methods can make use of local contexts for robust tracking, and reduce the negative impacts of too much background on CF model learning. By using multiple controlling parameters, the proposed spatial weight map function can flexibly vary the weight change speed from the target center to surrounding background regions, and regulate the impacts of local context regions on CF model learning.

CORRELATION FILTER-BASED METHODS
CONTEXT TRACKERS
THE CF METHODS WITH SPATIAL REGULARIZATION
THE SIGMOID SPATIAL WEIGHT MAP FUNCTION
PROBLEM FORMULATION OF AMCCF FRAMEWORK
OPTIMIZATION
EXPERIMENTS
Findings
CONCLUSION
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.