scholarly journals Adaptive Context-Aware and Structural Correlation Filter for Visual Tracking

2019 ◽  
Vol 9 (7) ◽  
pp. 1338 ◽  
Author(s):  
Bin Zhou ◽  
Tuo Wang

Accurate visual tracking is a challenging issue in computer vision. Correlation filter (CF) based methods are sought in visual tracking based on their efficiency and high performance. Nonetheless, traditional CF-based trackers have insufficient context information, and easily drift in scenes of fast motion or background clutter. Moreover, CF-based trackers are sensitive to partial occlusion, which may reduce their overall performance and even lead to failure in tracking challenge. In this paper, we presented an adaptive context-aware (CA) and structural correlation filter for tracking. Firstly, we propose a novel context selecting strategy to obtain negative samples. Secondly, to gain robustness against partial occlusion, we construct a structural correlation filter by learning both the holistic and local models. Finally, we introduce an adaptive updating scheme by using a fluctuation parameter. Extensive comprehensive experiments on object tracking benchmark (OTB)-100 datasets demonstrate that our proposed tracker performs favorably against several state-of-the-art trackers.

Sensors ◽  
2021 ◽  
Vol 21 (3) ◽  
pp. 889
Author(s):  
Hang Chen ◽  
Weiguo Zhang ◽  
Danghui Yan

Object information significantly affects the performance of visual tracking. However, it is difficult to obtain accurate target foreground information because of the existence of challenging scenarios, such as occlusion, background clutter, drastic change of appearance, and so forth. Traditional correlation filter methods roughly use linear interpolation to update the model, which may lead to the introduction of noise and the loss of reliable target information, resulting in the degradation of tracking performance. In this paper, we propose a novel robust visual tracking framework with reliable object information and Kalman filter (KF). Firstly, we analyze the reliability of the tracking process, calculate the confidence of the target information at the current estimated location, and determine whether it is necessary to carry out the online training and update step. Secondly, we also model the target motion between frames with a KF module, and use it to supplement the correlation filter estimation. Finally, in order to keep the most reliable target information of the first frame in the whole tracking process, we propose a new online training method, which can improve the robustness of the tracker. Extensive experiments on several benchmarks demonstrate the effectiveness and robustness of our proposed method, and our method achieves a comparable or better performance compared with several other state-of-the-art trackers.


Author(s):  
Libin Xu ◽  
Pyoungwon Kim ◽  
Mengjie Wang ◽  
Jinfeng Pan ◽  
Xiaomin Yang ◽  
...  

AbstractThe discriminative correlation filter (DCF)-based tracking methods have achieved remarkable performance in visual tracking. However, the existing DCF paradigm still suffers from dilemmas such as boundary effect, filter degradation, and aberrance. To address these problems, we propose a spatio-temporal joint aberrance suppressed regularization (STAR) correlation filter tracker under a unified framework of response map. Specifically, a dynamic spatio-temporal regularizer is introduced into the DCF to alleviate the boundary effect and filter degradation, simultaneously. Meanwhile, an aberrance suppressed regularizer is exploited to reduce the interference of background clutter. The proposed STAR model is effectively optimized using the alternating direction method of multipliers (ADMM). Finally, comprehensive experiments on TC128, OTB2013, OTB2015 and UAV123 benchmarks demonstrate that the STAR tracker achieves compelling performance compared with the state-of-the-art (SOTA) trackers.


2020 ◽  
Vol 39 (3) ◽  
pp. 3825-3837
Author(s):  
Yibin Chen ◽  
Guohao Nie ◽  
Huanlong Zhang ◽  
Yuxing Feng ◽  
Guanglu Yang

Kernel Correlation Filter (KCF) tracker has shown great potential on precision, robustness and efficiency. However, the candidate region used to train the correlation filter is fixed, so tracking is difficult when the target escapes from the search window due to fast motion. In this paper, an improved KCF is put forward for long-term tracking. At first, the moth-flame optimization (MFO) algorithm is introduced into tracking to search for lost target. Then, the candidate sample strategy of KCF tracking method is adjusted by MFO algorithm to make it has the capability of fast motion tracking. Finally, we use the conservative learning correlation filter to judge the moving state of the target, and combine the improved KCF tracker to form a unified tracking framework. The proposed algorithm is tested on a self-made dataset benchmark. Moreover, our method obtains scores for both the distance precision plot (0.891 and 0.842) and overlap success plots (0.631 and 0.601) on the OTB-2013 and OTB-2015 data sets, respectively. The results demonstrate the feasibility and effectiveness compared with the state-of-the-art methods, especially in dealing with fast or uncertain motion.


2019 ◽  
pp. 1-11 ◽  
Author(s):  
Tayssir Bouraffa ◽  
Liping Yan ◽  
Zihang Feng ◽  
Bo Xiao ◽  
Q. M. Jonathan Wu ◽  
...  

2018 ◽  
Vol 2018 ◽  
pp. 1-12 ◽  
Author(s):  
Jianjun Ni ◽  
Xue Zhang ◽  
Pengfei Shi ◽  
Jinxiu Zhu

Correlation filter based trackers have received great attention in the field of visual target tracking, which have shown impressive advantages in terms of accuracy, robustness, and speed. However, there are still some challenges that exist in the correlation filter based methods, such as target scale variation and occlusion. To deal with these problems, an improved kernelized correlation filter (KCF) tracker is proposed, by employing the GM(1,1) grey model, the interval template matching method, and multiblock scheme. In addition, a strict template update strategy is presented in the proposed method to accommodate the appearance change and avoid template corruption. Finally, some experiments are conducted. The proposed method is compared with the top state-of-the-art trackers, and all the tracking algorithms are evaluated on the object tracking benchmark. The experimental results demonstrate obvious improvements of the proposed KCF-based visual tracking method.


2021 ◽  
Vol 11 (3) ◽  
pp. 953
Author(s):  
Jin Hong ◽  
Junseok Kwon

In this paper, we propose a novel visual tracking method for unmanned aerial vehicles (UAVs) in aerial scenery. To track the UAVs robustly, we present a new object proposal method that can accurately determine the object regions that are likely to exist. The proposed object proposal method is robust to small objects and severe background clutter. For this, we vote on candidate areas of the object and increase or decrease the weight of the area accordingly. Thus, the method can accurately propose the object areas that can be used to track small-sized UAVs with the assumption that their motion is smooth over time. Experimental results verify that UAVs are accurately tracked even when they are very small and the background is complex. The proposed method qualitatively and quantitatively delivers state-of-the-art performance in comparison with conventional object proposal-based methods.


2020 ◽  
Vol 12 (2) ◽  
pp. 325 ◽  
Author(s):  
Yufei Zha ◽  
Min Wu ◽  
Zhuling Qiu ◽  
Jingxian Sun ◽  
Peng Zhang ◽  
...  

In urban environment monitoring, visual tracking on unmanned aerial vehicles (UAVs) can produce more applications owing to the inherent advantages, but it also brings new challenges for existing visual tracking approaches (such as complex background clutters, rotation, fast motion, small objects, and realtime issues due to camera motion and viewpoint changes). Based on the Siamese network, tracking can be conducted efficiently in recent UAV datasets. Unfortunately, the learned convolutional neural network (CNN) features are not discriminative when identifying the target from the background/clutter, In particular for the distractor, and cannot capture the appearance variations temporally. Additionally, occlusion and disappearance are also reasons for tracking failure. In this paper, a semantic subspace module is designed to be integrated into the Siamese network tracker to encode the local fine-grained details of the target for UAV tracking. More specifically, the target’s semantic subspace is learned online to adapt to the target in the temporal domain. Additionally, the pixel-wise response of the semantic subspace can be used to detect occlusion and disappearance of the target, and this enables reasonable updating to relieve model drifting. Substantial experiments conducted on challenging UAV benchmarks illustrate that the proposed method can obtain competitive results in both accuracy and efficiency when they are applied to UAV videos.


Sign in / Sign up

Export Citation Format

Share Document