Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al....

13
7128 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020 Fast Learning of Spatially Regularized and Content Aware Correlation Filter for Visual Tracking Ruize Han , Wei Feng , Member, IEEE, and Song Wang , Senior Member, IEEE Abstract— With a good balance between accuracy and speed, correlation filter (CF) has become a popular and dominant visual object tracking scheme. It implicitly extends the training samples by circular shifts of a given target patch, which serve as negative samples for fast online learning of the filters. Since all these shifted patches are not real negative samples of the target, CF tracking scheme suffers from the annoying boundary effects that can greatly harm the tracking performance, especially under challenging situations, like occlusion and fast temporal variation. Spatial regularization is known as a potent way to alleviate such boundary effects, but with the cost of highly increased time complexity, caused by complex optimization imported by spatial regularization. In this paper, we propose a new fast learning approach to content-aware spatial regularization, namely weighted sample based CF tracking (WSCF). In WSCF, specifically, we present a simple yet effective energy function that implicitly weighs different training samples by spatial deviations. With the energy function, the learning of correlation filters is composed of two subproblems with closed-form solution and can be efficiently solved in an alternate way. We further develop a content-aware updating strategy to dynamically refine the weight distribution to well adapt to the temporal variations of the target and background. Finally, the proposed WSCF is used to enhance two state-of-the-art CF trackers to significantly boost their tracking accuracy, with little sacrifice on the tracking speed. Extensive experiments on five benchmarks validate the effectiveness of the proposed approach. Index Terms—Object tracking, correlation filter, boundary effects, fast spatial regularization, temporal variations. I. I NTRODUCTION V ISUAL object tracking is a classical problem and plays an important role in computer vision, with many appli- cations in practice [1]–[3]. In visual object tracking, fast Manuscript received October 20, 2019; revised April 20, 2020; accepted May 16, 2020. Date of publication June 5, 2020; date of current version July 8, 2020. This work was supported in part by the National Natural Science Foundation of China (NSFC) under Grant U1803264, Grant 61671325, and Grant 61672376. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Husrev T. Sencar. (Corresponding author: Wei Feng.) Ruize Han and Wei Feng are with the School of Computer Science and Technology, College of Intelligence and Computing, Tianjin University, Tianjin 300350, China, also with the Key Research Center for Surface Monitoring and Analysis of Cultural Relics (SMARC), State Administration of Cultural Heritage, Tianjin 300350, China, and also with the Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin 300350, China (e-mail: [email protected]). Song Wang is with the School of Computer Science and Technology, College of Intelligence and Computing, Tianjin University, Tianjin 300350, China, also with the Key Research Center for Surface Monitoring and Analysis of Cultural Relics (SMARC), State Administration of Cultural Heritage, Tianjin 300350, China, and also with the Department of Computer Science and Engineering, University of South Carolina, Columbia, SC 29208 USA. Digital Object Identifier 10.1109/TIP.2020.2998978 learning an effective appearance model of the target is crucial for tracking accuracy and robustness [4], [5]. Many machine learning methods, e.g., support vector machines (SVM) [6], [7], subspace learning [8], online multi-instance boost- ing [9], sparse and compressive reconstruction [5], [10], correlation filter (CF) [11], [12], and convolutional neural network (CNN) [13], [14], have been developed to this end. Among all of them, correlation filter (CF) based track- ers have shown great advantages due to the balance of accuracy and speed [15]. Specifically, the popularity of CF tracking scheme mainly comes from two aspects: 1) the circular convolution operation in CF effectively implic- itly generates plenty of simulating negative samples that enables fast learning of more discriminative correlation fil- ters; 2) the Fast Fourier Transform (FFT) significantly accel- erates the computation and highly improves the efficiency of algorithm. As a result, some correlation filter (CF) based trackers [11] can run over 600 fps with a single CPU and generate quite promising tracking accuracy as well. Despite the above advantages, the circularly shifted patches contain unwanted circular boundary effects [16] and are not real negative samples of the target. Such boundary effects can severely harm the performance of the CF tracking scheme, especially under challenging situations, such as occlusion and fast temporal variation. Recently, two categories of methods are proposed to alleviate the annoying boundary effects for CF tracking scheme. The first category includes SRDCF [17] and CSR-DCF [18]. They extend the region of training patch and introduce a spatial regularization (SR) map to adapt the filter to learn from the target region while suppress the biased background region. Theoretically, such methods are equiva- lent to assigning different weights to the training samples, i.e., the samples generated by strict circular shifting will get the lower weights. Spatial regularization [17], [18] has been shown to be very effective to improve both the accuracy and robustness of CF based trackers. However, it imports a spatial regularization term that leads to complex optimization to the learning of filters, thus harms the efficiency foundation of CF tracking scheme and highly increases the time complexity. The second category of CF based trackers tackling boundary effects includes CFLB [16] and BACF [19], which generate more realistic negative training samples extracted directly from the background. The limitation of such methods is that all the positive and negative training samples are given equal weights in the learning of filters, which is not good enough for 1057-7149 © 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See https://www.ieee.org/publications/rights/index.html for more information. Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Transcript of Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al....

Page 1: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7128 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

Fast Learning of Spatially Regularized and ContentAware Correlation Filter for Visual Tracking

Ruize Han , Wei Feng , Member, IEEE, and Song Wang , Senior Member, IEEE

Abstract— With a good balance between accuracy and speed,correlation filter (CF) has become a popular and dominantvisual object tracking scheme. It implicitly extends the trainingsamples by circular shifts of a given target patch, which serve asnegative samples for fast online learning of the filters. Since allthese shifted patches are not real negative samples of the target,CF tracking scheme suffers from the annoying boundary effectsthat can greatly harm the tracking performance, especially underchallenging situations, like occlusion and fast temporal variation.Spatial regularization is known as a potent way to alleviatesuch boundary effects, but with the cost of highly increasedtime complexity, caused by complex optimization importedby spatial regularization. In this paper, we propose a newfast learning approach to content-aware spatial regularization,namely weighted sample based CF tracking (WSCF). In WSCF,specifically, we present a simple yet effective energy function thatimplicitly weighs different training samples by spatial deviations.With the energy function, the learning of correlation filters iscomposed of two subproblems with closed-form solution and canbe efficiently solved in an alternate way. We further developa content-aware updating strategy to dynamically refine theweight distribution to well adapt to the temporal variationsof the target and background. Finally, the proposed WSCF isused to enhance two state-of-the-art CF trackers to significantlyboost their tracking accuracy, with little sacrifice on the trackingspeed. Extensive experiments on five benchmarks validate theeffectiveness of the proposed approach.

Index Terms— Object tracking, correlation filter, boundaryeffects, fast spatial regularization, temporal variations.

I. INTRODUCTION

V ISUAL object tracking is a classical problem and playsan important role in computer vision, with many appli-

cations in practice [1]–[3]. In visual object tracking, fast

Manuscript received October 20, 2019; revised April 20, 2020; acceptedMay 16, 2020. Date of publication June 5, 2020; date of current versionJuly 8, 2020. This work was supported in part by the National Natural ScienceFoundation of China (NSFC) under Grant U1803264, Grant 61671325,and Grant 61672376. The associate editor coordinating the review of thismanuscript and approving it for publication was Prof. Husrev T. Sencar.(Corresponding author: Wei Feng.)

Ruize Han and Wei Feng are with the School of Computer Scienceand Technology, College of Intelligence and Computing, Tianjin University,Tianjin 300350, China, also with the Key Research Center for SurfaceMonitoring and Analysis of Cultural Relics (SMARC), State Administrationof Cultural Heritage, Tianjin 300350, China, and also with the Tianjin KeyLaboratory of Cognitive Computing and Application, Tianjin University,Tianjin 300350, China (e-mail: [email protected]).

Song Wang is with the School of Computer Science and Technology,College of Intelligence and Computing, Tianjin University, Tianjin 300350,China, also with the Key Research Center for Surface Monitoring and Analysisof Cultural Relics (SMARC), State Administration of Cultural Heritage,Tianjin 300350, China, and also with the Department of Computer Scienceand Engineering, University of South Carolina, Columbia, SC 29208 USA.

Digital Object Identifier 10.1109/TIP.2020.2998978

learning an effective appearance model of the target is crucialfor tracking accuracy and robustness [4], [5]. Many machinelearning methods, e.g., support vector machines (SVM)[6], [7], subspace learning [8], online multi-instance boost-ing [9], sparse and compressive reconstruction [5], [10],correlation filter (CF) [11], [12], and convolutional neuralnetwork (CNN) [13], [14], have been developed to thisend. Among all of them, correlation filter (CF) based track-ers have shown great advantages due to the balance ofaccuracy and speed [15]. Specifically, the popularity of CFtracking scheme mainly comes from two aspects: 1) thecircular convolution operation in CF effectively implic-itly generates plenty of simulating negative samples thatenables fast learning of more discriminative correlation fil-ters; 2) the Fast Fourier Transform (FFT) significantly accel-erates the computation and highly improves the efficiencyof algorithm. As a result, some correlation filter (CF)based trackers [11] can run over 600 fps with a singleCPU and generate quite promising tracking accuracy aswell.

Despite the above advantages, the circularly shifted patchescontain unwanted circular boundary effects [16] and are notreal negative samples of the target. Such boundary effects canseverely harm the performance of the CF tracking scheme,especially under challenging situations, such as occlusion andfast temporal variation. Recently, two categories of methodsare proposed to alleviate the annoying boundary effects forCF tracking scheme. The first category includes SRDCF [17]and CSR-DCF [18]. They extend the region of training patchand introduce a spatial regularization (SR) map to adapt thefilter to learn from the target region while suppress the biasedbackground region. Theoretically, such methods are equiva-lent to assigning different weights to the training samples,i.e., the samples generated by strict circular shifting will getthe lower weights. Spatial regularization [17], [18] has beenshown to be very effective to improve both the accuracy androbustness of CF based trackers. However, it imports a spatialregularization term that leads to complex optimization to thelearning of filters, thus harms the efficiency foundation of CFtracking scheme and highly increases the time complexity.The second category of CF based trackers tackling boundaryeffects includes CFLB [16] and BACF [19], which generatemore realistic negative training samples extracted directly fromthe background. The limitation of such methods is that allthe positive and negative training samples are given equalweights in the learning of filters, which is not good enough for

1057-7149 © 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 2: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

HAN et al.: FAST LEARNING OF SPATIALLY REGULARIZED AND CONTENT AWARE CF FOR VISUAL TRACKING 7129

Fig. 1. Top: comparative tracking results on a video sequence using thebaseline trackers Staple and BACF, and their spatially-regularized versionWSCFSt and WSCFBA enhanced by the proposed WSCF approach. Bottom:intersection-over-union (IoU) curve between the predicted and ground truthbounding boxes of the target tracked by the four trackers..

the tracking accuracy and cannot fully suppress the boundaryeffects either.

In this paper, we propose a new fast content-awarespatial-temporal regularization approach to the CF trackingscheme, namely weighted sample based CF tracking (WSCF).In WSCF, we assign different content-related weights to theimplicit circular shifting generated training samples, whilestill preserving the efficiency of filter learning. Specifically,we propose a simple yet effective energy function, with theenergy function, we can fast learn the correlation filters ineach frame by a closed form solution. We further formulatethe dynamic updating of such content-aware weight map asa constrained quadratic optimization problem. This enableour approach to well adapt to the temporal variations ofthe target and background. Since our WSCF provides ageneral way to alleviate boundary effects for the whole CFtracking scheme, it is applicable to many CF based trackers.In our experiments, we equip two state-of-the-art CF trackerswith the capability of spatial regularization and significantlyboost their tracking accuracy, without highly decreasing theirtracking speed. Extensive experiments on five benchmarksvalidate the effectiveness of the proposed WSCF approach.As shown in Fig. 1, the enhanced WSCFSt and WSCFBAtrackers by our approach clearly outperforms their respectivebaselines, i.e., Staple and BACF. Note, compared to thespatial regularization based trackers, e.g., SRDCF [17] andCSR-DCF [18], WSCFSt and WSCFBA also achieve betteraccuracy and higher speed. Compared to the real-sample basedtracker CFLB [16], the proposed approach outperforms witha large margin. Moreover, the proposed WSCF approach canenhance BACF [19] to further improve its tracking accuracy.

Our major contributions are three-fold.

• A simple yet effective bound term is introduced into theCF formulation to alleviate the boundary effects and thenew energy function can be optimized in closed form.

• A dynamic content-aware updating model is proposed toadjust the weight distribution of the samples to well adaptto the temporal variations, which can be solved by a fastALM based algorithm.

• Experimental results on five benchmarks, i.e., OTB-2013,OTB-2015, VOT-2018, TC-128 and LaSOT validate thatthe proposed approach obviously improves the trackingaccuracy of CF based trackers, while maintaining theirreal-time running speed.

The rest of this paper is organized as follows. Section IIsummarizes the related works. Our approach is elaborated inSection III. Section IV extensively evaluates and comparesthe proposed WSCF approach and state-of-the-art competitors.Finally, we conclude in Section V.

II. RELATED WORK

A. Correlation Filter (CF) Based Tracking

Bolme et al. proposed the tracker MOSSE [11], whichinaugurated the CF based tracking framework. CF frameworkshows two main advantage in handling tracking problem,i.e., the circular convolution operation extends the trainingsamples sufficiently and the FFT operation decreases thecomputational complexity significantly. Due to the good tradeoff between accuracy and speed, CF tracking was developedquickly and has shown continuous performance improvementon benchmarks in recent years [12], [20]–[29]. Two typicalstrategies have been used to obtain better performance inCF tracking – integrating more effective features and usinga more complete approach in filter learning. In the firststrategy, multi-channel feature maps are integrated to CFtracking [12], [23]. Henriques et al. [12] proposed a CF trackerwith multi-channel Histogram of Oriented Gradient (HOG)features, which can improve the tracking accuracy whilemaintaining a high running speed. Danelljan et al. appliedmulti-dimensional color attributes in DSST [23]. Li andZhu further proposed SAMF tracker [24] using the featurecombination for CF tracking. Recently, deep CNN basedfeatures have been applied to CF tracking [25], [27], whichfurther improve the tracking performance but taking morecomputation time. In the second strategy, many conceptualimprovements for filter learning in CF tracking are presented,e.g., non-linear kernelized correlation filter (KCF) used in [12],accurate scale estimation used in DSST [21], and color sta-tistics integration proposed in Staple [20]. DRT [30] proposesa novel CF-based optimization problem to jointly model thediscrimination and reliability information. Sun et al. [31]integrated ROI (region-of-interest) based pooling method intoCF tracking and proposed a novel ROI pooled correlation fil-ter (RPCF) algorithm for robust tracking. Recently, CFNet [32]models the CF tracking into an end-to-end framework byinterpreting the CF learner as a differentiable layer in a deepneural network. Wang et al. [33] developed a new unsuper-vised learning method for CF-based deep tracking. Visualobject tracking especially CF tracking has many applicationscenarios [34]. For example, Shao et al. employed the velocityfeature [35], and two complementary features, i.e., the opticalflow and the histogram of oriented gradient [36], to boostthe CF tracking in satellite videos. Although these methodsmade much progress in CF tracking over the initial MOSSE,the boundary effect problem in CF tracking is still unsolvedin these trackers.

B. Spatial Regularization for CF Tracking

One issue in CF tracking is the unwanted boundary effects,which are resulted from the unreal training samples generatedby circular convolution. Several popular methods have beenproposed to address the boundary effects by considering

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 3: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7130 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

spatial constraints. In [16] and [19], CF is learned from morereal training examples by enlarging the region of samplecollection. Galoogahi et al. [16] proposed a CF tracker withlimited boundary (CFLB) to reduce the boundary effectsin CF tracking. In [19], the background-aware correlationfilter (BACF) based tracking is developed to learn CF fromreal negative training samples extracted from the background.SRDCF [17] adopts a spatial regularization (SR) componentto penalize CF values. The spatial regularization punishesthe unreal training samples by assigning different weights todifferent samples. Based on SRDCF, Danelljan et al. [37]introduced a novel formulation for learning a convolutionoperator in the continuous spatial domain, which was furtherenhanced to tackle the problems of computational complexityand over-fitting simultaneously in ECO [38]. Moreover, basedon [38], UPDT [39] further unveils the power of deep featuresfor CF based tracking. Similarly, Lukezic et al. [18] proposeddiscriminative correlation filter with channel and spatial reli-ability (CSR-DCF) using the spatial regularization map toconstrain the filter learned from the object region. Besides,CACF [40] allows the explicit incorporation of global contextwithin CF trackers. BSCF [41] introduces spatial constraintsin CF by suppressing the background region of the target infilter learning.

C. Spatial-Temporal Regularization for CF Tracking

An important factor in tracking task is to consider thetemporal variation in tracking process, e.g., the size, appear-ance and shape change of the target over time. Severalspatial-temporal regularization based methods were proposedto alleviate the incapability of CF tracking models in handlingchangeable scenes. Most of these methods are based on thespatial regularization proposed in SRDCF [17]. Specifically,a unified formulation for CF based tracking is proposedin [42] to dynamically manage the training set by estimat-ing the quality of the samples. STRCF [43] introduces thetemporal regularization to SRDCF to provide a more robustappearance model than the baseline in the case of largeappearance variations. Similarly, STRSCF [44] integrates thespatial prior directly to the image and a temporal regularizationterm as in STRCF to improve the tracking performances.Zhang et al. [45] presented a part-based tracking frameworkby exploiting multiple SRDCFs. FOSR [46] incorporates anobject-adaptive spatial regularization model for CF tracking.More recently, CRSRCF [47] integrates temporal contentinformation of the target into the spatial constraint of [17]and DSAR-CF [48] introduces a dynamic saliency-awarespatial constraint into CF tracking. Similarly, ASRCF [49]incorporates an adaptive spatially regularized CF model tooptimize the filter while adaptively tuning the spatial regu-larization weights. In addition, SSRCF [50] proposes a selec-tive spatial regularization based on SRDCF. Besides abovetrackers extended from SRDCF, Hu et al. [51] proposed aspatial-aware temporal aggregation network to construct moreefficient features for CF tracking. Guo et al. [52] derived adynamic transformations to handle the variation of both thetarget and background.

The proposed WSCF is totally different from the aboveworks in that we introduce the weighting constraint boundterm into CF formulation to alleviate the boundary effectsby assigning different weights to the training samples. Ourformulation does not involve the elementwise multiplicationoperation on the correlation filter, and avoids the computa-tional inefficiency of classical SR. This way, the filter can besolved by the closed-form solution, which almost does notincrease the computational complexity.

III. THE PROPOSED METHOD

A. Background and Motivation

1) CF Tracking and Boundary Effects: There are two pri-mary stages in CF tracking framework, i.e., detection stageand filter learning stage. In the detection stage, an onlineupdated correlation filter f ∈ R

N×1 is applied to detect thetarget location in a search region frame by frame (z ∈ R

N×1

represents the feature vector extracted from the search region),and response vector c ∈ R

N×1 can be computed by

c = z � f, (1)

where � denotes circular convolution, and the target locationis on the peak of c. In filter learning stage, correlation filter fcan be updated by minimizing

ECF(f) = ‖x � f − y‖2 + λ‖f‖2, (2)

where x denotes the vectorized feature map extracted fromthe training samples, and y ∈ R

N×1 is the desired output(i.e., the Gaussian-shaped ground truth), λ ≥ 0 is a controlfactor.1 Optimizing Eq. (2) w.r.t. f can be efficiently solvedin frequency domain where ‘�’ becomes elementwise multi-plication, which leads to a super real-time algorithm [11]. Asdiscussed in [16], the circular convolution in Eq. (2) assumesthe cyclic shifts of x, which causes the periodic repetitions onboundary positions. The cyclic shifts of base sample generatethe unfaithful training samples as shown in Fig. 2 and harmthe discriminative power of learned filters. This is called theboundary effects that inevitably degrades the CF trackingperformance.

2) Spatial Regularization (SR) for Boundary Effects: Spatialregularization (SR) [17] is a classical method to alleviate theboundary effects in CF framework. We then revisit SR basedtracker SRDCF [17], which introduces a spatial weight mapto penalize filter values outside the object boundaries by thefollowing energy function,

ESR(f) = ‖x � f − y‖2 + λ‖w � f‖2, (3)

where � denotes the elementwise multiplication operation,and w is the vectorized spatial variant weight map. When weset f ′ = w � f , Eq. (3) can be rewritten as

E′SR(f ′) =

∥∥∥∥ f ′

w� x − y

∥∥∥∥2

+ λ∥∥f ′∥∥2

, (4)

1For simplicity, what we describe above is for one-channel feature map.In practice, it can be extended to multiple channel feature maps by usingmultiple feature-extraction algorithms [53].

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 4: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

HAN et al.: FAST LEARNING OF SPATIALLY REGULARIZED AND CONTENT AWARE CF FOR VISUAL TRACKING 7131

Fig. 2. An example frame from the video soccer (left) and the tracking results on this video has been shown in Fig. 1. An illustration of the training samplesin CF framework and the weight distribution for different training samples using the proposed WSCF and original CF framework (middle). The responsemaps generated by these two types of methods (right)..

where f ′w � x can be written as Cdiag( 1

w )f′. Each row of Cdenotes a vector of circularly shifted x, i.e., a training sampleand diag(·) is the diagonalization function for a vector. ThenCdiag( 1

w )f′ can be regarded as assigning the weights 1w to

the training samples. As discussed in [48], in the original CFenergy function of Eq. (2), the second term can be taken asλ||1 � f ||, which means both the real sample and syntheticsamples are of equal importance for filter learning. However,those synthetic samples shifted to be far from the targetcenter cannot represent the real scene. Hence, the syntheticsamples may introduce disturbance in filter learning. SRDCFgenerates w according to the shifting distance of trainingsamples, i.e., a synthetic sample with large shifting distancewill be assigned a small weight. This effectively removes theinfluence of useless synthetic samples, thus learns much morediscriminative filters. From this point, the spatial regularizationcan be regarded as assigning different weights to the trainingsamples to deal with the boundary effects.

3) Limitation of SR: We transform Eq. (3) into Fourierdomain and get

ESR(f) =∥∥∥x � f − y

∥∥∥2 + λ∥∥∥ ˆw � f

∥∥∥2, (5)

where · denotes the corresponding variable in Fourier domain.The second terms of Eq. (3) and Eq. (5) are the SR terms. Wecan see that the SR weight map w imports circular convolutionoperation into the second term of Eq. (5). This breaks thesimplicity of elementwise multiplication in Fourier domain ofthe first term in Eq. (5), which is the efficiency foundation ofCF framework. As a result, we have to use the Gauss-Seidel(GS) [17] or conjugate gradient (CG) [37], [38] algorithms toiteratively solve the corresponding linear system for learningf, whose complexity is O(N3 D3) and O(N2 D), respectively,where N and D denote the size and dimension of the featuremap x. Therefore, although the SR improves the accuracy ofCF trackers, the use of SR term also significantly deceleratesthe algorithm speed, e.g., SRDCF runs at about 4 fps withHOG features. This way, we aim to construct a new spatialregularization alike method to alleviate the boundary effectswithout breaking the elementwise multiplication in Fourierdomain for learning the filters.

B. Weighted Samples for Correlation Filter Learning

In this paper, we propose a new spatially variant weightedsample based CF model. First, the CF energy function ofEq. (2) can be rewritten as

Ew(f, y∗)=∥∥x � f − y∗∥∥2+‖f‖2+λy

∥∥∥w � (y−y∗)2∥∥∥

1, (6)

where x, f and y have the same meanings as in Eq. (2),λy ≥ 0 is a control factor and ‖·‖1 denotes the vectorL1-norm. The notation w ∈ R

N×1 denotes the vectorizedweight distribution map, y∗ ∈ R

N×1 represents the computedresult of the filter applied to the training samples, i.e., x � f .In original CF function as Eq. (2), it aims to minimize thedifference between the desired output y and computed resulty∗, by treating all the training samples equally. As shownin Fig. 2, we introduce the spatially variant weight distributionw in Eq. (6) by assigning different weights to the trainingsamples (the samples generated by seriously circular shiftwill get the lower weights) to alleviate the boundary effectsproblem of CF as in [17]. But differently we do not implementthe elementwise operation of w with f , which avoids theinefficient computation in [17].

The energy function Eq. (6) involves two variables f and y∗.With respect to y∗, the gradient of Ew can be obtainedby

∂Ew(y∗)∂y∗ = 2(f � x − y∗) + 2λy(w � (y − y∗)). (7)

By setting ∂Ew(y∗)∂y∗ = 0, we get the closed-form solution

y∗ = (1 + λyw)−1 · (x � f + λyw � y). (8)

With respect to f , the form of Eq. (6) is the same as Eq. (2),while the only difference is the replacement of y with y∗and therefore, the filter f can be quickly solved as in originalCF framework [11]. In the proposed spatially variant weightsbased CF tracking, we solve y∗ and f alternately. Both y∗ andf can be solved by the closed-form solution, which makesthe algorithm more efficient. This way, the proposed methodhandles the boundary effects in CF by assigning differentweights to the training samples and weakens the impact ofunfaithful synthetic samples, which can achieve the similareffectiveness with spatial regularization (SR) [17] as discussed

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 5: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7132 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

in Section III-A. However, different from SR, the proposedformulation does not involve the elementwise multiplicationoperation on the filter f as in Eq. (3), which breaks theFourier-domain efficient solution of CF tracking.

C. Dynamic Updating of Weight Distribution

We introduced the weight distribution w in Eq. (6). Wecan see that w is fixed in the whole tracking process afterinitialization. However, in real-world tracking tasks, objectshape is usually irregular and may change frequently in thetracking process. From this perspective, it is unconscionableto define a constant weight distribution using only the spatialdistance to the map center and fix it over time. Similar studiescan be found in recent works [48], [49], which integratethe object-adaptive/temporal-dynamic constraint to improvethe fixed SR map in [17]. In the following, we considerthe object-related and temporal-varying information into theweight distribution to obtain more reliable filters and overcomethe limitation of the fixed regularization weight distribution w.Accordingly, we attempt to improve the weight distributionfrom fixed w into dynamic d.

1) Problem Formulation: Based on Eq. (6), we introducethe dynamic weight distribution d ∈ R

N×1 and get

Ed(f, y∗, d) = ∥∥x � f − y∗∥∥2 + ‖f‖2

+λy

∥∥∥d � (y − y∗)2∥∥∥

1+ λw

∥∥∥∥d2

w

∥∥∥∥1

s.t.

{dk ≥ 0∑

k dk = μ, k = 1, 2 . . . N. (9)

where w = {wk |k = 1 . . . N}, d = {dk|k = 1 . . . N}. Differentfrom the energy function in Eq. (6), the new formulationEq. (9) is a function of three variables, i.e., the correlationfilter f, computed result y∗, and the weight distribution d.As a result, the weight map d is no longer pre-set constantsas w in Eq. (6). The two constraints ensure that the weightsin d are non-negative and summed up to μ. The last termin Eq. (9) is the regularization term on the sample weightsin d. This regularization is used to constrain the similaritybetween d and w, which is controlled by the flexibilityparameter λw > 0 and the prior sample weights wk > 0,satisfying

∑k dk = ∑

k wk = μ. The parameter λw > 0controls the adaptiveness of the sample weights dk . Changingλw leads to a different degree of flexibility in the weightsdk , which will be discussed in the later part – ‘Analysisof λw’. Note that, we use ‘d2’ rather than ‘d’ in Eq. (9).This is because dk should be reserved in the differentiationof Ed w.r.t. dk , which is shown in following Eq. (12) andEq. (13).

2) Optimization: We extract the function Ed w.r.t. d andrewrite it in terms of scalar as

Ed(d) = λy

∑k

|dk(yk − y∗k )2| + λw

∑k

| d2k

wk|

s.t.

{dk ≥ 0∑

k dk = μ, k = 1, 2 . . . N. (10)

Eq. (10) can be regarded as a constraint optimization prob-lem. To optimize Eq. (10), we use Augmented LagrangeMethod (ALM) [54] to solve for d. We temporarily ignorethe inequality constraint dk ≥ 0 and introduce Lagrangemultipliers for the equality constraint,

Ed(d)=∑

k

(λy |dk(yk − y∗k )2|+λw| d2

k

wk|)−η(

∑k

dk −μ), (11)

where η > 0 denotes the Lagrange multiplier. Differentiationw.r.t. dk gives,

∂ED

∂dk= (Lk + 2λw

dk

wk) − η, (12)

where we denote Lk = λy(yk − y∗k )2. The stationary

point is computed by setting the partial derivatives tozero,

∂ED

∂dk= 0 ⇔ dk = η − Lk

2λwwk . (13)

The Lagrange multiplier η is computed by summing bothsides of Eq. (13) over k and using

∑dk = ∑

wk = μ,∑k

dk =∑

k

η − Lk

2λwwk ⇔

μ =∑

k

η

2λwwk −

∑k

Lk

2λwwk ⇔

η = 2λw + 1

μ

∑k

Lkwk . (14)

Combining (14) and (13) we have

dk = wk + wk

2λwμ(

N∑l=1

Llwl − Lk), (15)

where Ll = λy(yl − y∗l )2.

Then we take into account the ignored constraint dk ≥ 0,and we define the constant2

δ = mink

2μwk · |wk(

N∑l=1

Llwl − Lk)|−1. (16)

This choice ensures that dk > 0, ∀k if 0 < 1λw

< δ. Theinequality constraint is thus satisfied for 0 < 1

λw< δ.

3) Analysis of λw: We analyze the effect of λw by consider-ing the extreme cases of decreasing (λw → 0) and increasing(λw → +∞) the flexibility parameter. 1) λw → 0: Thiscorresponds to removing the last term in Eq. (9), implyingno regularization on dk . For the fixed Y and Y∗, the energyfunction Eq. (9) is minimized by setting smaller values todk to those samples in which yk is close to y∗

k and largervalues to dk when yk is far from y∗

k . That is easy to makethe weighting map lose the spatial constraint, if the lastterm in Eq. (9) is removed. Therefore, it is imperative touse a regularization on the weights dk . 2) λw → +∞: Byintroducing Lagrange multipliers, from Eq. (15) it can beshown that dk → wk when λw → +∞. Thus, increasingthe parameter λw also reduces the flexibility of the weights dk

2Please see the Appendix for more details.

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 6: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

HAN et al.: FAST LEARNING OF SPATIALLY REGULARIZED AND CONTENT AWARE CF FOR VISUAL TRACKING 7133

about the prior weights wk . The constant weight distributionin Eq. (6) is therefore obtained in the limit λw → +∞ bysetting dk = wk . This way, the fixed spatial distribution wused in Section III-B is a special case of d. The dynamicweight distribution can be seen as a generalization of Eq. (6)by introducing flexible sample weights dk .

D. The New Tracking Scheme

To evaluate the influence of the proposed weight constrainedtraining samples, we choose two famous state-of-the-art CFbased trackers, i.e., Staple [20] and BACF [19] as the baselinemethods. In the following, we denote the proposed trackerbased on Staple [20] and BACF [19] as WSCFSt and WSCFBArespectively. Similar to [17], we initialize the spatial weightmap as a 2D Gaussian-shape map as shown in Fig. 2, then wis the vectorization of the weight map. For fair comparison,the proposed methods use the same features as our baselinemethods. Specifically, following BACF, we employ 31-channelHOG features [55] using 4 × 4 cells multiplied by a Hannwindow [11] in WSCFBA. For WSCFSt, we combine theHOG features and the global color histogram as in [20]. Theparameters λy are set to 0.5 and 3 for WSCFSt and WSCFBA,respectively. We set λw = 1

κδ (0 < κ < 1) to satisfy theconstraint of 0 < 1

λw< δ and δ is calculated by Eq. (16) and

κ is set as 0.9 and 0.5 in WSCFSt and WSCFBA, respectively.Besides, we fix all the other parameters of the baseline trackersfor fair comparison. We further discuss the influence of theparameters in Section IV-D.

The formal description of the proposed WSCF trackingscheme is shown in Algorithm 1. For scale estimation in thealgorithm, following the most previous CF-based trackers [17],[19], we apply the filter on multi-scale searching areas forscale estimation [24]. Specifically, the learned filter f is appliedon S searching areas with different covered regions, wherethe searching areas have been resized into the same size asthe filter and S denotes the number of scales. This returnsS correlation outputs.We employ the interpolation strategyin [17] to calculate detection scores of each output. The scalewith the maximum score among all the outputs is taken as theoptimal scale.

IV. EXPERIMENTAL RESULTS

A. Setup

1) Datasets and Metrics: The proposed method is imple-mented in MATLAB and runs on a desktop computer with anIntel Core i7 3.4 GHz CPU. We evaluate the proposed methodon five standard benchmarks: OTB-2013, OTB-2015, VOT-2018, TC-128 and LaSOT. The first two OTB datasets contains50 and 100 videos, respectively. We use two acknowledgedmetrics, i.e., precision and success rate under the OPE (one-pass evaluation) for quantitative evaluation. The precisionmetric measures the distance between the detected targetlocations and those of the ground truth. The success ratemetric measures the intersection-over-union (IoU) betweenpredicted and ground truth bounding boxes. For precisionmetric, we use a threshold, i.e., 20 pixels to judge whether atracker is successful at each frame and calculate the percentage

Algorithm 1 WSCF Tracking Scheme

of successful frames out of all the frames in each sequence. Forsuccess rate metric, we calculate average success percentagesw.r.t. different overlap ratio thresholds and obtain success plotand after that the area under curve (AUC) of each plot can becalculated as the success rate AUC score. The VOT-2018 [56]has 60 videos and re-initializes the object bounding box whenthe tracker fails to track the target. The expected averageoverlap (EAO) considers both bounding box overlap ratio– accuracy, and the failures times (re-initialization times) –robustness, which serves as the major evaluation metrics. TheVOT-2018 also provides the EAO score with standard (base-line) and real-time experimental setup. The latter requirespredicting bounding boxes faster or equal to the video frame-rate. TC-128 [57] contains 128 color sequences and it is usedto explore the ability of trackers to encode color information.LaSOT [58] dataset is a recently proposed large-scale videodataset for visual object tracking which consists of 1,400 long-term sequences. Here, we report the results on its testing subsetwhich contains 280 sequences with 690K frames. Followingthe metrics of OTB dataset, TC-128 and LaSOT also adoptthe AUC of success plot and precision at 20 pixels as theevaluation metrics.

2) Comparison Methods: We compare the proposed methodwith 10 state-of-the-art trackers. Besides our baseline trackersi.e., Staple [20] and BACF [19], there are six hand-crafted fea-tures based trackers, i.e., KCF [12], DSST [21], SAMF [24],DLSSVM [7], SRDCF [17], CSR-DCF [18], and two deepfeature based methods including HCF [25] and HDT [26].Among them, KCF, DSST, SAMF, Staple, SRDCF, CSR-DCFand BACF are classical CF based trackers which obtain theperformance improvement in filter learning. Besides, HCF andHDT are deep feature boosted CF trackers, DLSSVM is aSVM based tracker.

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 7: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7134 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

TABLE I

ATTRIBUTES BASED SUCCESS RATE AUC SCORES FOR WSCFSt, WSCFBA AND OTHER 12 STATE-OF-THE-ART TRACKERS ON OTB-2015. THE BESTTHREE RESULTS ARE MARKED IN RED, GREEN AND BLUE RESPECTIVELY. ⇑ DENOTES THE IMPROVEMENT COMPARED TO BASELINE TRACKER

Fig. 3. Precision plots (left) and success plots (right) of both the proposedand comparison methods on OTB-2013 (first row) and OTB-2015 (secondrow) benchmark. The legend contains the average distance precision score at20 pixels and the AUC of success plot of each method.

B. Comparative Results

1) Comparison Results on OTB: We evaluate the overallperformance of the proposed method and compare with thebaseline methods and other 10 state-of-the-art trackers onOTB-2013 and OTB-2015 benchmarks. As shown in the firstrow of Fig. 3, in terms of precision score, our methodsWSCFBA and WSCFSt outperform the baseline methods andachieve 3.7% and 2.4% improvement over BACF and Staple,respectively. In terms of the AUC of success plots, the pro-posed WSCFBA and WSCFSt also get better performancethan the baseline methods and achieve the state-of-the-artperformance. We can see that both WSCFBA and WSCFStoutperform the deep trackers HCF, HDT. We can see similarresults on OTB-2015 in the second row of Fig. 3. The pro-posed WSCFBA and WSCFSt outperform the baseline methodsBACF and Staple in both precision score and AUC score of

success rate. We can also see that several deep trackers, e.g.,HDT, obtain higher precision scores than the proposed method.However, our WSCFBA obtains the highest performance insuccess plot AUC score, which is a more comprehensivemetric to evaluate the object tracking performance.

2) Attribute Based Comparison: To more comprehensivelyevaluate the proposed tracker in various scenes, we presenttracking performance in terms of different attributes onOTB-2015. The 100 videos of OTB-2015 are grouped into11 subsets according to 11 attributes, i.e., occlusion (OCC),background clutter (BC), illumination variation (IV), fastmotion (FM), deformation (DEF), scale variation (SV), out-of-plane rotation (OPR), in-plane rotation (IPR), out-of-view(OV), motion blur (MB), and low resolution (LR). Table Ishows the success plot AUC of 10 comparison methods andthe proposed WSCFBA and WSCFSt on 11 subsets. In terms ofresults on 11 subsets, WSCFBA gets the best results on 7 sub-sets of OCC, BC, IV, FM, DEF, SV, and OPR and the secondbest performance on 2 subsets of IPR and LR. We can see thatWSCFSt outperforms the baseline tracker Staple on 10 subsetsexcept for LR. WSCFBA also outperforms BACF on 8 subsetsexcept for the comparable results on SV subset, and inferiorresults on OV and LR subsets. Note that, the subsets of OVand LR contain 14 and 9 sequences respectively, which are twoattributes with the fewest sequences. Such small evaluation setcan be easily biased.

3) Comparison Results on VOT: In addition to OTB bench-mark, we also evaluate the proposed method on VOT-2018.We compare WSCFSt, WSCFBA with the baseline trackers,i.e., Staple [20] and BACF [19], and five methods that par-ticipate in the VOT challenge, i.e., KCF [12], DSST [21],SAMF [24], SRDCF [17] and CSR-DCF [18]. As shownin Table II, we can see that both WSCFSt and WSCFBA out-perform the baseline methods Staple and BACF in accuracy,robustness and expected average overlap (EAO) under baselineexperiments, respectively. In terms of accuracy, WSCFStgets the best performance which improves the performanceof Staple by 3.5%. Note that, CSR-DCF provides the bestperformance on robustness and baseline EAO score. However,

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 8: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

HAN et al.: FAST LEARNING OF SPATIALLY REGULARIZED AND CONTENT AWARE CF FOR VISUAL TRACKING 7135

TABLE II

COMPARATIVE RESULTS ON VOT-2018 IN TERMS OF THE AVERAGEACCURACY (ACCURACY), AVERAGE FAILURES (ROBUSTNESS), EAO

UNDER BASELINE EXPERIMENTS AND EAO UNDER

REALTIME EXPERIMENTS

as shown in the last column, it provides a very poor EAOscore in real-time experimental setting, which is even lowerthan KCF. This is because CSR-DCF runs at ∼ 8 fps, faraway from real-time speed. The proposed WSCFSt gets thebest EAO score on real-time experiments, which can run at areal-time speed on VOT dataset.

4) Comparison Results on TC-128: We show the evaluationresults on TC-128 in Fig. 4. Clearly, with our WS basedmethod, the precision scores of BACF, and Staple get therelative improvements at 2.3% and 1.8%, respectively. Forsucess rate AUC scores, WSCFBA and WSCFSt get 2.9%and 2.2% relative improvements over their baseline methods,respectively. We can see that DLSSVM gets the best perfor-mance among the comparison methods. Note that, althoughDLSSVM gets the highest score in term of precision, for thecompositive metric – success rate AUC score, our methodscan obtain the similar (WSCFBA at 49.9%) or better (WSCFStat 51.1% ) results compared to DLSSVM.

5) Comparison Results on LaSOT: We further evaluate theproposed method on LaSOT dataset. As shown in Fig. 5,we can see that the proposed method improves both BACFand Staple with 3.3% and 2.5% relative increments on pre-cision scores, respectively. In terms of success rate AUCscore, the relative increments are 2.3% and 3.7%, respec-tively. We can also see that, compared to original baselinetracker Staple with a success rate AUC score of 24.3%,the proposed method WSCFSt gets the AUC score of 25.2%,which outperforms several trackers that are superior to Staple,e.g., CSR-DCF, SRDCF, DLSSVM and HCF, and obtains thecomparative performance with HDT. Moreover, the proposedmethod WSCFBA gets the AUC score of 26.5% exceedingHDT. Above results demonstrate that the proposed WSCFalgorithm not only improves the CF trackers on short-termsequences, e.g., OTB and TC-128 datasets, but also helpshandle challenges from long-term sequences.

C. Ablation Study

To validate the effectiveness of our method, we show theresults of ablation experiments on OTB-2013 and OTB-2015.

Fig. 4. Precision plots (left) and success plots (right) of both the proposed andcomparison methods on TC-128 benchmark. The legend contains the averagedistance precision score at 20 pixels and the AUC of success plot of eachmethod.

Fig. 5. Precision plots (left) and success plots (right) of both the proposed andcomparison methods on LaSOT benchmark. The legend contains the averagedistance precision score at 20 pixels and the AUC of success plot of eachmethod.

TABLE III

ABLATION STUDY ON OTB-2013 AND OTB-2015 VIA SUCCESSRATES (% AT IOU > 0.50 AND AUC SCORE)

Specifically, as shown in Table III, Staple and BACF thetwo baseline trackers. WSCFSt-no d and WSCFBA-no d usethe fixed weight distribution i.e., w in Eq. (6), to assign theweights for training samples. The bottom row WSCFSt andWSCFBA use the dynamically updated weight distributioni.e., d in Eq. (9), to assign the sample weights. For eachbaseline tracker, from the first and second rows we can see thatthe weighting for samples by the fixed weight distribution wcan improve the performance of CF based trackers. Moreover,from the last two rows of each tracker we can also see that thedynamic updating of the weight distribution by d can furtherpromote the improvement, which demonstrates the advantageof adding the updating of w.

To further study the effectiveness of the dynamic weightdistribution presented in Section III-C, we also evaluatethe tracking performance in terms of different attributes onOTB-2013 and OTB-2015. We select six subsets with

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 9: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7136 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

TABLE IV

ABLATION STUDY ON SIX SUBSETS WITH DIFFERENT ATTRIBUTES IN OTB-2013 AND OTB-2015 VIA SUCCESS RATES (% AUC SCORE)

TABLE V

COMPARATIVE STUDY OF DIFFERENT PARAMETER SETTINGS OF λy , κ ON

OTB-2015 VIA SUCCESS RATES (% AT IOU > 0.50 AND AUC SCORE)

corresponding attributes in which the targets or their back-grounds have frequent variations over time, i.e., scale variation(SV), in-plane rotation (IPR), out-of-plane rotation (OPR),illumination variation (IV), deformation (DEF), occlusion(OCC). As shown in Table IV, we can see that the proposedmethods, including WSCFSt and WSCFBA, using dynamicweights d perform better than both the baselines and thoseusing fixed weights w, on the selected attribute-aware sub-sets of OTB-2013 and OTB-2015. This way, the dynamicweight distribution d can better handle the cases wherethe targets/backgrounds have over-time changes, therefore itimproves the final accuracy on overall dataset as shownin Table III.

D. Algorithm Analysis

1) Parameters Selection Analysis: We investigate the per-formance changes w.r.t. different setups of the two parameters,i.e., the control parameters λy and λw in Eq. (9). The para-meter values in WSCFSt and WSCFBA are slightly differentdue to the diversity of the baseline methods and we tunethe setups of the parameters on two methods, respectively.As shown in Table V, we set λy as the original settings andenlarge/reduce it two times respectively. We can see that thetracking accuracy changes little with the huge displacementof λy . Hence, our method is not very sensitive to λy . Theparameter λw in Eq. (9) is determined by two factors, i.e., κand δ. The variable δ is calculated online in the trackingprocess, thus we only tune the hyperparameter κ . Table Vcompares the tracking results on OTB-2015 with differentsetups of κ . Clearly, the tracking accuracy changes little withdifferent κ . In practice, we can see that the tracking results

under different parameter settings all outperform the baselinetrackers, which demonstrates the robustness of the proposedmethod.

2) Qualitative Analysis: Qualitative comparisons on severalsequences are shown in Fig. 6. The two sequences dragonbabyand shaking are selected to show the robustness of trackersagainst object deformation. The targets in these two sequencesare head of a baby and a singer, respectively, both havesignificant appearance variations when moving and turning.We can see that BACF fails to track the baby (e.g., #113)and Staple fails to track the singer (e.g., #31, #295, #335).The proposed trackers WSCFSt and WSCFBA can track themcontinuously against the deformation. The second row in Fig. 6shows the tracking results on two representative sequencesskater2 and basketball, where the targets are the whole bodyof a skater and a basketball player, respectively. The shape ofthe targets also vary drastically as the motion of athletes. Wecan see that Staple tracker fails to estimate the scale of theskater (e.g., #351, #339, #373) and both Staple and BACF lostthe basketball player (e.g., #500, #666). Our trackers WSCFStand WSCFBA can track the target successfully over the wholesequence. The third row in Fig. 6 shows the target with severeor long-term occlusion. In the girl sequence, the girl can betracked well until it is occluded by a man in a long term.Three trackers, i.e. Staple, BACF and CSR-DCF mistakenlytrack the man appearing in the foreground (e.g., #457), whilethe proposed trackers WSCFSt and WSCFBA can unceasinglytrack the target. Similarly in the jogging sequence, both Stapleand BACF lost the runner after leaving the obstruction (e.g.,#070), while the proposed tracker WSCFBA can keep trackingher continuously.

3) Failure Cases and Limitations Analysis: We have shownthat proposed training sample weighting does help improvethe tracking accuracy of CF based tracking method. However,the sample weighting strategy becomes less effective when thetarget moves very fast. Specifically, as shown in the last rowof Fig. 6, when the target, e.g. the high jumper, moves fast,the center localization of the target may be far away betweenneighbor frames. As a result, the weight distribution mayassign higher weights to the training samples that contain morebackground region instead of the target region, which makesthe updated filter in frame t − 1 less effective in detectingthe target in frame t . Note that, although sample weightingstrategy has such limitation, our method, i.e. WSCF, stilloutperforms the baseline trackers on the subset of fast motion

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 10: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

HAN et al.: FAST LEARNING OF SPATIALLY REGULARIZED AND CONTENT AWARE CF FOR VISUAL TRACKING 7137

Fig. 6. Tracking results of WSCFBA, WSCFSt and other four other CF trackers on eight challenging videos in OTB-2015. The last two cases show thelimitation of WSCF based trackers in handing object fast motion.

TABLE VI

COMPARATIVE STUDY OF DIFFERENT EXPERIMENTAL SETTINGS FOR SR AND WS BASED TRACKERS ON OTB-2015 VIA SUCCESSRATES (% AUC SCORE) AND AVERAGE RUNNING SPEED (FPS)

in OTB-2015, as shown in Table I. This is because the abovesudden motion does not happen frequently in a sequence andthe proposed sample weighting strategy could help improvetracking accuracy for most of the time.

4) Comparison of WS and SR Model: To compare the pro-posed weighted samples (WS) based method for CF with thespatial regularization (SR) in SRDCF, we conduct some track-ing algorithms with different settings. Specifically, we selectKCF as the baseline and integrate the SR and WS methods toboost it, respectively. As shown in the the first row of Table VI,both SR and WS can improve the tracking performance ofKCF. Although both the SR and WS reduce the runningspeed of the baseline, the WS based method can still runsat real-time speed. However, the SR severely pulls downthe speed into 7.8 fps. Besides, we also integrate the scaleestimation (SE) method [17], [24] into the baseline for furthercomparison as shown in the second row. We can see thatthe WS based method can get comparative accuracy with SRand maintain a higher running speed. As shown in the lastrow, we compare the complete SRDCF [17] algorithm withour WS based tracker using the same scale of search region,training data and detection strategy, i.e., WSCF. We can seethat although the tracking performance of WSCF is slightlylower than SRDCF, WSCF has an approximate seven-timespeedup compared to SRDCF and achieves real-time running

Fig. 7. Precision plots (left) and success plots (right) of both the proposedmethod with deep features and state-of-the-art deep learning based methodson OTB-2015. The legend contains the average distance precision score at20 pixels and the AUC of success plot of each method.

speed. Moreover, as shown in Fig. 3, the proposed techniquescan be applied to existing trackers, e.g., Staple and BACF,to obtain further performance gain.

5) Deep Feature Based Tracker: To further assess WSCF,we exploit the deep features (Norm1 from VGG-M, Conv4-3from VGG-16) for object representation to enhance ourmethod WSCFBA. In the following, we name it as WSCFDFfor simplicity. We compare it with several most state-of-the-artdeep learning based methods, including the latest CF trackers:ECO [38], CFNet [32], and ASRCF [49]; a series of Siamese

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 11: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7138 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

TABLE VII

SUCCESS RATES (% AT IOU>0.50 AND AUC SCORE) OF WSCFSt,WSCFBA VERSUS RELATED TRACKERS, AND THECORRESPONDING WEIGHTED AVERAGE SPEEDS ON OTB-2015 DATASET

Network based trackers: SiameseFC [14], SiamDW [59], andSiamRPN++ [60]; and other methods proposed in 2019:ATOM [61], GCT [62], TADT [63], DiMP [64], GradNet [65].The comparison results on OTB-2015 are shown in Fig. 7,we can see that WSCFDF achieves comparative or betterperformance compared to these state-of-the-art deep learningbased methods in term of both the precision and success rateAUC scores.

6) Speed Analysis: Algorithm efficiency is also an impor-tant factor for the tracking task. Table VII compares severalrelated well-known CF trackers, where the speed is measuredon a desktop computer with an Intel Core i7 3.4GHz CPU. Wecan see that the proposed WSCFSt and WSCFBA can improvethe baseline Staple and BACF in terms of success rate whilewith little sacrifice on tracking speed. Our method WSCFStachieves the real-time running speed over 30 fps and WSCFBAruns near real time at an average of 16 fps, both of whichare more efficient than the spatial regularization based CFtracking methods, e.g., SRDCF, CSR-DCF. The performanceimprovement of WSCFBA is relatively small compared withWSCFSt, because the real samples used in BACF can alleviatethe boundary effects to some extent.

Note that, the proposed methods do not rely on the offlinetrained models and pre-trained deep features, which caneffectively save training time and storage space. Moreover,our method is implemented on the Matlab platform withoutany optimization strategies and can be implemented for realtime applications by further optimizing the code with GPUacceleration or parallel computing.

V. CONCLUSION

In this paper, we have proposed a fast content-aware spatialregularization approach for correlated filter (CF), one of themost popular visual object tracking schemes. Our approachimplicitly weighs circular shifted training samples and solvethe spatially regularized learning of correlation filters in closedform. This provides a new efficient spatial regularizationway for the CF tracking scheme. Furthermore, to adapt totemporal variations, we present a content-aware updatingstrategy to dynamically optimize the weight distribution bysolving a constrained quadratic optimization problem. Sincethe proposed fast content-aware spatial regularization approachis general for the CF tracking scheme, it is able to helpmany CF based trackers to alleviate the boundary effectswithout harming their original tracking speed. Particularly, ourapproach is used to improve two state-of-the-art CF track-ers resulting in very promising performance improvement.

On five benchmark datasets, OTB-2013, OTB-2015, VOT-2018, TC-128 and LaSOT, we have validated the effectivenessand superiority of our approach over various state-of-the-artcompetitors. In the future, we plan to further explore the poten-tials of our approach to more recent sophisticated CF basedtracking methods, and study its possible application withinother popular tracking scheme, such as Siamese network.

APPENDIX

We discuss the derivation of constraint condition defined inEq. (16), which is to guarantee dk > 0. From Eq. (15),

dk = wk + 1

λw· wk

2μ�L, (17)

where we denote �L = ∑Nl=1 Llwl − Lk . Given the preset

parameters wk ≥ 0 and μ ≥ 0. 1) If �L ≥ 0, dk willmonotonically increase with 1

λwand it is easy to get that

dk > 0 when 1λw

≥ 0. 2) If �L < 0, dk will monotonicallydecrease with 1

λwand dk will get the minimum value when

1λw

take the maximum. To calculate the upper bound of 1λw

,we solve

wk + 1

λw· wk

2μ�L = 0, (18)

and then we get

1

λw= 2μwk · |wk�L|−1. (19)

Thus we take δ = min 2μwk · |wk�L|−1. To sum up,the constraint condition 0 < 1

λw< δ will guarantee dk > 0.

ACKNOWLEDGMENT

The authors thank all reviewers and the associate editor fortheir valuable comments. They thank Q. Guo, P. Zhang, andZ. Chen for the mutual academic discussion. They especiallythank A. Waseem for helping them to remotely access thedevices in lab for implementing experiments during the schoolholiday.

REFERENCES

[1] A. W. M. Smeulders, D. M. Chu, R. Cucchiara, S. Calderara,A. Dehghan, and M. Shah, “Visual tracking: An experimental survey,”IEEE Trans. Pattern Anal. Mach. Intell., vol. 36, no. 7, pp. 1442–1468,Jul. 2014.

[2] C. A. Perez et al., “Automatic safety video surveillance-tracking systemto avoid vehicle-workers interaction for mining applications,” in Proc.Int. Symp. Optomechatronic Technol., Nov. 2014, pp. 23–27.

[3] R. Han et al., “Complementary-view multiple human tracking,” in Proc.AAAI Conf. Artif. Intell., 2020, pp. 1–8.

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 12: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

HAN et al.: FAST LEARNING OF SPATIALLY REGULARIZED AND CONTENT AWARE CF FOR VISUAL TRACKING 7139

[4] X. Li, W. Hu, C. Shen, Z. Zhang, A. Dick, and A. V. D. Hengel,“A survey of appearance models in visual object tracking,” ACM Trans.Intell. Syst. Technol., vol. 4, no. 4, pp. 1–48, Sep. 2013.

[5] Q. Guo, W. Feng, C. Zhou, C.-M. Pun, and B. Wu, “Structure-regularized compressive tracking with online data-driven sampling,”IEEE Trans. Image Process., vol. 26, no. 12, pp. 5692–5705, Dec. 2017.

[6] S. Avidan, “Support vector tracking,” IEEE Trans. Pattern Anal. Mach.Intell., vol. 26, no. 8, pp. 1064–1072, Aug. 2004.

[7] J. Ning, J. Yang, S. Jiang, L. Zhang, and M.-H. Yang, “Objecttracking via dual linear structured SVM and explicit feature map,” inProc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016,pp. 4266–4274.

[8] D. A. Ross, J. Lim, R.-S. Lin, and M.-H. Yang, “Incremental learningfor robust visual tracking,” Int. J. Comput. Vis., vol. 77, nos. 1–3,pp. 125–141, May 2008.

[9] B. Babenko, M.-H. Yang, and S. Belongie, “Visual tracking with onlinemultiple instance learning,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2009, pp. 983–990.

[10] T. Zhang, B. Ghanem, S. Liu, and N. Ahuja, “Robust visual tracking viamulti-task sparse learning,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 2042–2049.

[11] D. Bolme, J. R. Beveridge, B. A. Draper, and Y. M. Lui, “Visual objecttracking using adaptive correlation filters,” in Proc. IEEE Comput. Soc.Conf. Comput. Vis. Pattern Recognit., Jun. 2010, pp. 2544–2550.

[12] J. F. Henriques, R. Caseiro, P. Martins, and J. Batista, “High-speedtracking with kernelized correlation filters,” IEEE Trans. Pattern Anal.Mach. Intell., vol. 37, no. 3, pp. 583–596, Mar. 2015.

[13] H. Nam and B. Han, “Learning multi-domain convolutional neuralnetworks for visual tracking,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit. (CVPR), Jun. 2016, pp. 4293–4302.

[14] L. Bertinetto, J. Valmadre, J. F. Henriques, A. Vedaldi, and P. H. S. Torr,“Fully-convolutional Siamese networks for object tracking,” in Proc.Eur. Conf. Comput. Vis., 2016, pp. 850–865.

[15] Y. Wu, J. Lim, and M. H. Yang, “Object tracking benchmark,” IEEETrans. Pattern Anal. Mach. Intell., vol. 37, no. 9, pp. 1834–1848,Sep. 2015.

[16] H. K. Galoogahi, T. Sim, and S. Lucey, “Correlation filters withlimited boundaries,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.(CVPR), Jun. 2015, pp. 4630–4638.

[17] M. Danelljan, G. Hager, F. S. Khan, and M. Felsberg, “Learning spatiallyregularized correlation filters for visual tracking,” in Proc. IEEE Int.Conf. Comput. Vis. (ICCV), Dec. 2015, pp. 4310–4318.

[18] A. Lukežic, T. Vojir, L. C. Zajc, J. Matas, and M. Kristan, “Discrimina-tive correlation filter with channel and spatial reliability,” in Proc. IEEEConf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 6309–6318.

[19] H. K. Galoogahi, A. Fagg, and S. Lucey, “Learning background-awarecorrelation filters for visual tracking,” in Proc. IEEE Int. Conf. Comput.Vis. (ICCV), Oct. 2017, pp. 1135–1143.

[20] L. Bertinetto, J. Valmadre, S. Golodetz, O. Miksik, and P. H. S. Torr,“Staple: Complementary learners for real-time tracking,” in Proc.IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016,pp. 1401–1409.

[21] M. Danelljan, G. Häger, F. S. Khan, and M. Felsberg, “Accurate scaleestimation for robust visual tracking,” in Proc. Brit. Mach. Vis. Conf.,2014, pp. 1–5.

[22] T. Zhang, S. Liu, C. Xu, B. Liu, and M.-H. Yang, “Correlation particlefilter for visual tracking,” IEEE Trans. Image Process., vol. 27, no. 6,pp. 2676–2687, Jun. 2018.

[23] M. Danelljan, F. S. Khan, M. Felsberg, and J. V. D. Weijer, “Adaptivecolor attributes for real-time visual tracking,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2014, pp. 1090–1097.

[24] Y. Li and J. Zhu, “A scale adaptive kernel correlation filter trackerwith feature integration,” in Proc. Eur. Conf. Comput. Vis., 2014,pp. 254–265.

[25] C. Ma, J.-B. Huang, X. Yang, and M.-H. Yang, “Hierarchical convolu-tional features for visual tracking,” in Proc. IEEE Int. Conf. Comput.Vis. (ICCV), Dec. 2015, pp. 3074–3082.

[26] Y. Qi et al., “Hedged deep tracking,” in Proc. IEEE Conf. Comput. Vis.Pattern Recognit. (CVPR), Jun. 2016, pp. 4303–4311.

[27] M. Danelljan, G. Hager, F. S. Khan, and M. Felsberg, “Convolutionalfeatures for correlation filter based visual tracking,” in Proc. IEEE Int.Conf. Comput. Vis. Workshop (ICCVW), Dec. 2015, pp. 58–66.

[28] T. Zhang, C. Xu, and M.-H. Yang, “Multi-task correlation particle filterfor robust object tracking,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit. (CVPR), Jul. 2017, pp. 4335–4343.

[29] N. Wang, W. Zhou, Y. Song, C. Ma, and H. Li, “Real-time correlationtracking via joint model compression and transfer,” IEEE Trans. ImageProcess., vol. 29, pp. 6123–6135, 2020.

[30] C. Sun, D. Wang, H. Lu, and M.-H. Yang, “Correlation tracking viajoint discrimination and reliability learning,” in Proc. IEEE/CVF Conf.Comput. Vis. Pattern Recognit., Jun. 2018, pp. 489–497.

[31] Y. Sun, C. Sun, D. Wang, Y. He, and H. Lu, “ROI pooled correlationfilters for visual tracking,” in Proc. IEEE/CVF Conf. Comput. Vis.Pattern Recognit. (CVPR), Jun. 2019, pp. 5783–5791.

[32] J. Valmadre, L. Bertinetto, J. Henriques, A. Vedaldi, and P. H. S. Torr,“End-to-end representation learning for correlation filter based tracking,”in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017,pp. 2805–2813.

[33] N. Wang, Y. Song, C. Ma, W. Zhou, W. Liu, and H. Li, “Unsuperviseddeep tracking,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.(CVPR), Jun. 2019, pp. 1308–1317.

[34] B. Du, S. Cai, and C. Wu, “Object tracking in satellite videos basedon a multiframe optical flow tracker,” IEEE J. Sel. Topics Appl. EarthObserv. Remote Sens., vol. 12, no. 8, pp. 3043–3055, Aug. 2019.

[35] J. Shao, B. Du, C. Wu, and L. Zhang, “Tracking objects from satellitevideos: A velocity feature based correlation filter,” IEEE Trans. Geosci.Remote Sens., vol. 57, no. 10, pp. 7860–7871, Oct. 2019.

[36] J. Shao, B. Du, C. Wu, and L. Zhang, “Can we track targets from space?A hybrid kernel correlation filter tracker for satellite video,” IEEE Trans.Geosci. Remote Sens., vol. 57, no. 11, pp. 8719–8731, Nov. 2019.

[37] M. Danelljan, A. Robinson, F. S. Khan, and M. Felsberg, “Beyondcorrelation filters: Learning continuous convolution operators for visualtracking,” in Proc. Eur. Conf. Comput. Vis., 2016, pp. 472–488.

[38] M. Danelljan, G. Bhat, F. S. Khan, and M. Felsberg, “ECO: Efficientconvolution operators for tracking,” in Proc. IEEE Conf. Comput. Vis.Pattern Recognit. (CVPR), Jul. 2017, pp. 6638–6646.

[39] G. Bhat, J. Johnander, M. Danelljan, F. S. Khan, and M. Felsberg,“Unveiling the power of deep tracking,” in Proc. Eur. Conf. Comput.Vis., 2018, pp. 483–498.

[40] M. Mueller, N. Smith, and B. Ghanem, “Context-aware correlation filtertracking,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR),Jul. 2017, pp. 1396–1404.

[41] Z. Chen, Q. Guo, L. Wan, and W. Feng, “Background-suppressed cor-relation filters for visual tracking,” in Proc. IEEE Int. Conf. MultimediaExpo (ICME), Jul. 2018, pp. 1–6.

[42] M. Danelljan, G. Hager, F. S. Khan, and M. Felsberg, “Adaptive decon-tamination of the training set: A unified formulation for discriminativevisual tracking,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.(CVPR), Jun. 2016, pp. 1430–1438.

[43] F. Li, C. Tian, W. Zuo, L. Zhang, and M.-H. Yang, “Learningspatial-temporal regularized correlation filters for visual tracking,” inProc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018,pp. 4904–4913.

[44] B. Li, C. Liu, J. Liu, H. Gao, X. Song, and W. Liu, “Visual trackingusing spatial-temporal regularized support correlation filters,” in Proc.Int. Conf. Commun., Image Signal Process., 2018, pp. 1–7.

[45] D. Zhang et al., “Part-based visual tracking with spatially regularizedcorrelation filters,” Vis. Comput., vol. 36, no. 3, pp. 509–527, Mar. 2020.

[46] P. Zhang, Q. Guo, and W. Feng, “Fast and object-adaptive spatialregularization for correlation filters based tracking,” Neurocomputing,vol. 337, pp. 129–143, Apr. 2019.

[47] R. Han, Q. Guo, and W. Feng, “Content-related spatial regularizationfor visual object tracking,” in Proc. IEEE Int. Conf. Multimedia Expo(ICME), Jul. 2018, pp. 1–6.

[48] W. Feng, R. Han, Q. Guo, J. Zhu, and S. Wang, “Dynamic saliency-aware regularization for correlation filter-based object tracking,” IEEETrans. Image Process., vol. 28, no. 7, pp. 3232–3245, Jul. 2019.

[49] K. Dai, D. Wang, H. Lu, C. Sun, and J. Li, “Visual tracking via adap-tive spatially-regularized correlation filters,” in Proc. IEEE/CVF Conf.Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 4670–4679.

[50] Q. Guo, R. Han, W. Feng, Z. Chen, and L. Wan, “Selective spatialregularization by reinforcement learned decision making for objecttracking,” IEEE Trans. Image Process., vol. 29, pp. 2999–3013, 2020.

[51] T. Hu, L. Huang, X. Liu, and H. Shen, “Real time visual tracking usingspatial-aware temporal aggregation network,” 2019, arXiv:1908.00692.[Online]. Available: http://arxiv.org/abs/1908.00692

[52] Q. Guo, W. Feng, C. Zhou, R. Huang, L. Wan, and S. Wang, “Learningdynamic siamese network for visual object tracking,” in Proc. IEEE Int.Conf. Comput. Vis. (ICCV), Oct. 2017, pp. 1763–1771.

[53] H. K. Galoogahi, T. Sim, and S. Lucey, “Multi-channel correlation fil-ters,” in Proc. IEEE Int. Conf. Comput. Vis., Dec. 2013, pp. 3072–3079.

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.

Page 13: Fast Learning of Spatially Regularized and Content Aware ...songwang/document/tip20c.pdfBolme et al. proposed the tracker MOSSE [11], which inaugurated the CF based tracking framework.

7140 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020

[54] S. Boyd, “Distributed optimization and statistical learning via the alter-nating direction method of multipliers,” Found. Trends Mach. Learn.,vol. 3, no. 1, pp. 1–122, 2010.

[55] N. Dalal and B. Triggs, “Histograms of oriented gradients for humandetection,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. PatternRecognit. (CVPR), Jun. 2005, pp. 886–893.

[56] M. Kristan et al., “The sixth visual object tracking VOT2018 challengeresults,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 3–53.

[57] P. Liang, E. Blasch, and H. Ling, “Encoding color information for visualtracking: Algorithms and benchmark,” IEEE Trans. Image Process.,vol. 24, no. 12, pp. 5630–5644, Dec. 2015.

[58] H. Fan et al., “LaSOT: A high-quality benchmark for large-scalesingle object tracking,” in Proc. IEEE/CVF Conf. Comput. Vis. PatternRecognit. (CVPR), Jun. 2019, pp. 5374–5383.

[59] Z. Zhang and H. Peng, “Deeper and wider siamese networks for real-time visual tracking,” in Proc. IEEE/CVF Conf. Comput. Vis. PatternRecognit. (CVPR), Jun. 2019, pp. 4591–4600.

[60] B. Li, W. Wu, Q. Wang, F. Zhang, J. Xing, and J. Yan, “SiamRPN++:Evolution of siamese visual tracking with very deep networks,” in Proc.IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019,pp. 4282–4291.

[61] M. Danelljan, G. Bhat, F. S. Khan, and M. Felsberg, “ATOM: Accuratetracking by overlap maximization,” in Proc. IEEE/CVF Conf. Comput.Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 4660–4669.

[62] J. Gao, T. Zhang, and C. Xu, “Graph convolutional tracking,” in Proc.IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019,pp. 4649–4659.

[63] X. Li, C. Ma, B. Wu, Z. He, and M.-H. Yang, “Target-aware deeptracking,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.(CVPR), Jun. 2019, pp. 1369–1378.

[64] G. Bhat, M. Danelljan, L. Van Gool, and R. Timofte, “Learningdiscriminative model prediction for tracking,” in Proc. IEEE/CVF Int.Conf. Comput. Vis. (ICCV), Oct. 2019, pp. 6182–6191.

[65] P. Li, B. Chen, W. Ouyang, D. Wang, X. Yang, and H. Lu, “GradNet:Gradient-guided network for visual object tracking,” in Proc. IEEE/CVFInt. Conf. Comput. Vis. (ICCV), Oct. 2019, pp. 6162–6171.

Ruize Han received the B.S. degree in mathe-matics and applied mathematics from the HebeiUniversity of Technology, China, in 2016, and theM.E. degree in computer technology from TianjinUniversity, China, in 2019. He is currently pursuingthe Ph.D. degree with the College of Intelligence andComputing, Tianjin University. His major researchinterest is visual intelligence, specifically includingmulti-camera video collaborative analysis and visualobject tracking. He is also interested in solving pre-ventive conservation problems of cultural heritagesvia artificial intelligence.

Wei Feng (Member, IEEE) received the Ph.D.degree in computer science from the City Universityof Hong Kong in 2008. From 2008 to 2010, he wasa Research Fellow with The Chinese University ofHong Kong and City University of Hong Kong.He is currently a Professor with the School ofComputer Science and Technology, College of Com-puting and Intelligence, Tianjin University, China.His major research interests are active robotic visionand visual intelligence, specifically including activecamera relocalization and lighting recurrence, gen-

eral Markov random fields modeling, energy minimization, active 3D sceneperception, SLAM, and generic pattern recognition. Recently, he focuses onsolving preventive conservation problems of cultural heritages via computervision and machine learning. He is an Associate Editor of Neurocomputingand Journal of Ambient Intelligence and Humanized Computing.

Song Wang (Senior Member, IEEE) received thePh.D. degree in electrical and computer engineeringfrom the University of Illinois at Urbana Champaign(UIUC), Champaign, IL, USA, in 2002. He wasa Research Assistant with Image Formation andProcessing Group, Beckman Institute, UIUC, from1998 to 2002. In 2002, he joined the Departmentof Computer Science and Engineering, Universityof South Carolina, Columbia, SC, USA, where heis currently a Professor. His current research inter-ests include computer vision, image processing, and

machine learning. He is a member of the IEEE Computer Society. He is cur-rently serving as the Publicity/Web Portal Chair of the Technical Committeeof Pattern Analysis and Machine Intelligence of the IEEE Computer Society,an Associate Editor of the IEEE TRANSACTION ON PATTERN ANALYSIS

AND MACHINE INTELLIGENCE, Pattern Recognition Letters, and ElectronicsLetters.

Authorized licensed use limited to: University of South Carolina. Downloaded on July 10,2020 at 00:00:37 UTC from IEEE Xplore. Restrictions apply.