Abstract

Object tracking as a low-level vision task has always been a hot topic in computer vision. It is well known that Challenges such as background clutters, fast object motion and occlusion et al. affect a lot the robustness or accuracy of existing object tracking methods. This paper proposes a reinforcement learning model based on Twin Delayed Deep Deterministic algorithm (TD3) for single object tracking. The model is based on the deep reinforcement learning model, Actor-Critic (AC), in which the Actor network predicts a continuous action that moves the target bounding box in the previous frame to the object position in the current frame and adapts to the object size. The Critic network evaluates the confidence of the new bounding box online to determine whether the Critic model needs to be updated or re-initialized. In further, in our model we use TD3 algorithm to further optimize the AC model by using two Critic networks to jointly predict the bounding box confidence, and to obtain the smaller predicted value as the label to update the network parameters, thereby rendering the Critic network to avoid excessive estimation bias, accelerate the convergence of the loss function, and obtain more accurate prediction values. Also, a small amount of random noise with upper and lower bounds are added to the action in the Actor model, and the search area is reasonably expanded in offline learning to improve the robustness of the tracking method under strong background interference and fast object motion. The Critic model can also guide the Actor model to select the best action and continuously update the state of the tracking object. Comprehensive experimental results on the OTB-2013 and OTB-2015 benchmarks demonstrate that our tracker performs best in precision, robustness, and efficiency when compared with state-of-the-art methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call