Object tracking using a convolutional network and a structured output SVM
2018-01-08JunweiLiXiaolongZhouSixianChanandShengyongChen
Junwei Li,Xiaolong Zhou,Sixian Chan,and Shengyong Chen
©The Author(s)2017.This article is published with open access at Springerlink.com
Object tracking using a convolutional network and a structured output SVM
Junwei Li1,Xiaolong Zhou1,Sixian Chan1,and Shengyong Chen2
©The Author(s)2017.This article is published with open access at Springerlink.com
Object tracking has been a challenge in computer vision. In this paper,we present a novel method to model target appearance and combine it with structured output learning for robust online tracking within a tracking-by-detection framework.We take both convolutional features and handcrafted features into account to robustly encode the target appearance. First,we extract convolutional features of the target by kernels generated from the initial annotated frame.To capture appearance variation during tracking,we propose a new strategy to update the target and background kernel pool.Secondly,we employ a structured output SVM for re fining the target’s location to mitigate uncertainty in labeling samples as positive or negative.Compared with existing state-of-the-art trackers,our tracking method not only enhances the robustness of the feature representation,but also uses structured output prediction to avoid relying on heuristic intermediate steps to produce labelled binary samples.Extensive experimental evaluation on the challenging OTB-50 video sequences shows competitive results in terms of both success and precision rate,demonstrating the merits of the proposed tracking method.
object tracking;convolutional network;structured learning;feature extraction
1 Introduction
Visual tracking is a fundamental research problem in computer vision and robotics,with wide applications such as intelligent video surveillance,transportationmonitoring,robot–human interaction,etc.In recent years,many excellent tracking algorithms have been proposed,but it remains a challenging problem for a tracker to handle occlusion,abrupt motion,appearance variation,and background clutter.
In this paper,we propose a novel tracking method which utilizes a discriminative convolutional network[1]and HOG descriptors[2]to encode target appearance,together with a structured output support vector machine(SO-SVM)to jointly estimate target appearance.In the proposed method,tracking is formulated as binary classification and structured output tasks,to select the most likely target candidate and reject background patches.It uses an online trained structural output classifier within a particle filter framework.The convolutional filters for modeling target appearance are generated from the initial frame(annotated manually).We perform a soft-shrink operation on the output convolutional feature maps to enhance their robustness.One of the most significant advantages is that the convolutional filters are generated from both the target and its surrounding area,so fully exploiting local structure and internal geometric layout of the target.Additionally,our method employs an SO-SVM to overcome the drawback that samples used for training the classifier are all equally weighted,meaning that a negative example which overlaps significantly with the tracker’s bounding box is treated the same as one which overlaps very little.Another advantage of SO-SVM is that the labeler no longer labels the samples as positive or negative based on intuition and heuristics.
The remainder of this paper is organized as follows.Section 2 gives a brief overview of related work.Details of our tracking method are described in Section 3.Section 4 reports qualitative and quantitative experimental results.We finally conclude this paper in Section 5.
2 Related work
Most existing tracking methods belong to two categories: generative models and discriminative models.The generative approach formulates the tracking problem as minimizing reconstruction error,while the discriminative model method considers the tracking problem as a binary classification task to separate the target from the background.From another perspective, a tracker can be decomposed into two components:an online updated appearance model for feature extraction and an observation model to find the most probable target transformation.
Recent tracking methods mainly focus on designing a robust appearance model[3]to capture target appearance variation.The most popular approach based on the discriminative model casts tracking as a foreground and background separation problem,performing tracking by learning a classifier using multiple instance learning[4],P-N learning[5],online boosting[6,7],SVM[8],structured output SVMs [9], CRFs [10], probability hypothesis density methods[11,12],etc.These tracking methods first train a classifier online,inspired by statistical machine learning methods,to separate the target from the background surrounding the target location in the previous frame.Generative methods describe the target’s appearance using generative models and search for target regions that best fit the model. Various generative target appearance modeling algorithms have been proposed using sparse representation[13,14],density estimation [15,16], and incremental subspace learning[17]. In Ref.[18],generative and discriminative models were combined for more accurate online tracking. Some efficient trackers have been proposed using hand-crafted features,including Haar-like feature histograms[4,6,9,19],HOG descriptors[2],binary features[20],and covariance descriptors[21].However,such trackers do not adapt well to target appearance variation.
To overcome the shortcomings of handcrafted features in modeling object appearance,deep networks have been employed to directly learn features from raw data without resorting to manual intervention.Convolutional features have been used in many applications such as Ref.[22].In Ref.[23],Li et al.used a convolutional neural network(CNN)for visual tracking with multiple image cues as inputs.In Ref.[24],Zhou et al.used an ensemble of deep networks in combination with an online boosting method for visual tracking. Reference[25]presented a human tracking algorithm that learns a specific feature extractor with CNNs.Numerous auxiliary data are required for offline training the deep networks;the pre-trained model is then used for online visual tracking. Wang and Yeung[26]developed a deep learning tracking method that uses stacked de-noising auto-encoders to learn generic features from a large number of auxiliary images.Reference[27]used a two-layer CNN to learn hierarchical features from auxiliary video sequences;it takes into account complicated motion transformations and appearance variations in visual tracking. A drawback of all the above frameworks is that they need a large amount of auxiliary data to pre-train a deep network model;such models can be highly specific and have poor adaptive ability.
In Ref.[1]Zhang et al.incorporated convolutional networks(convolutional filters defined as normalized image patches from the first frame)which do not require auxiliary data to train filters.They achieved state-of-the-art precision.Several tracking algorithms based on hand-crafted features have been developed within a multiple instance learning framework,aiming to improving the poor ability of hand-crafted features to represent semantic level features.Grabner et al.proposed an online boosting algorithm to select features for tracking.However,these trackers[28,29]use one positive sample(i.e.,the current tracker location)and a few negative samples when updating the classifier.As the appearance model is updated with noisy and potentially misaligned examples,this often leads to the problem of tracking drift.A semi-supervised learning approach can be used in which positive and negative samples are selected via an online classifier with structural constraints.Yang et al.[30]present a discriminative appearance model based on superpixels. It is able to handle heavy occlusion and recover from drift. In Ref.[9],Hare et al.used an online structured output support vector machine(SVM)for robust tracking;it can mitigate the effect of wrongly labeled samples.Reference[31]introduced a fast tracking algorithm which exploits the circulant structure of the kernel matrix in SVM classifiers so that it can be efficiently computed by the fast Fourier transform algorithm.
3Method
In this section,we describe our proposed tracking method in detail.The tracking problem is formulated as a detection task,and the pipeline of the proposed approach is shown in Fig.1.We assume that the tracking target is manually annotated in the first frame. To model target appearance,we sample various background and foreground convolutional kernels to encode target and background structural information.When a new frame arrives,we first extract its convolutional feature map to estimate the target transformation.Secondly,we incorporate structured output learning and HOG descriptors to predict the target location and scale variation.Lastly,the tracking results are combined to jointly determine the target transformation and scale variation.
3.1 Feature extraction by convolutional network
The convolutional network includes two separate layers.Firstly,a set of background and foreground convolutional kernels are generated from a bank of filters which sample the input frame using a sliding window.Secondly,to enhance the robustness of the convolutional feature representation,all feature maps are stacked together,and the final feature vector is determined by solving a sparse representation equation.
In the initial frame,a set of samples,denotedis warped to a canonical size ofn×nin grayscale color space.Then each sample is pre-processed by subtracting the mean andL2normalization is performed,aiming to modify local brightness differences and achieve contrast normalization, respectively.A sliding window strategy is employed to generate a bank of patches with the field sizew×w.This results in a total ofimage patches sampled from the initial frame.
Fig.1 Architecture of the proposed tracking algorithm.
Following the pre-processing step,thek-means algorithm is used to select multiple convolution filterfrom the filter bank using the initial frame as the representative target filters. The remaining object filter kernelsare selected dynamically to capture target appearance variation,and are generated by clustering the target filter bank of thetth frame.are filter banks obtained from the initial and tth frame,respectively.The strategy of generating a dynamic convolutional filter is the most significant difference from Ref.[1].One obvious advantage is that our filters have the ability to adapt in the face of target occlusion and deformation,and illumination changes,which cause target appearance variation.In other words,the proposed convolutional filters are more robust in dynamic environments.
Given the target candidate imageand the ith object filter kernelsconvolution ofand the filters is denoted byand⊗denotes the convolution operator.The local filters encode stable object visual information from both the initial frame and the previous frame even though the object may experience significant appearance change from the initial frame.Thereby,we can extract more discriminative features and effectively handle the drift problem.
The background context surrounding the object provides useful information to discriminate the target.The convolutional networks select m background samples surrounding the object,and then the same cluster algorithm is used to generate background filtersthe ith background sample. An average pooling strategy is operated to summarize each filter inNext,the background kernelsthat encode the visual information and geometric layout surrounding the object are generated:
To further enhance the strength of this representation and eliminate the influence of noise,a complex cell feature map that is a 3D tensoris constructed,which stacks d different simple cell feature maps constructed with the filter set
A sparse vector c is set to approximate vec(C)by minimizing the following objective function:
where vec(C)is a column vector concatenating all the elements in C,of length(n−w+1)2d.The optimization problem in Eq.(4)has a closed form solution,as explained in Ref.[32]:
where median(vec(C)) is robust to target appearance variation and noise interference.
3.2 HOG feature extraction
Hand-crafted features are morphological,shape,statistical,or textural based representations that attempt to encode object appearance at lowlevel,and are the fundamental elements of object representation.Contrary to the high-level semantic features found by convolutional networks,which can be treated as a black-box object representation,hand-crafted features encode object appearance and effectively preserve structure information,which is very important in object tracking.In this paper,we use HOG(histograms of oriented gradients)features[33]as complementary features to jointly encode target appearance.HOG descriptors have several advantages.For example,their gradient structure is very characteristic of local shape,they are computed in a local cell with an easily controllable degree,and they are invariant to local geometric and photometric transformations.In other words,translations or rotations make little difference even if they are much smaller than the local spatial or orientation bin size.All of these advantages of HOG features play a key role in target location and scale estimation.
3.3 Structured output learning
Traditional tracking-by-detection approaches employ a classifier trained online to distinguish the target object from its surrounding background. In the tracking process,the classifier is used to estimate the object’s transformation by searching for the maximum classification score amongst a set of target candidates around the target’s location in the previous frame,typically using a sliding window or another motion model to generate target candidates.Given the estimated target location,traditional tracking methods generate a set of binary labelled training samples to update the classifier online.This tracking framework raises a number of issues.Firstly,it is not clear how to label the training samples in a principled manner.One popular way is to utilize predefined rules such as the distance between a sample and the estimated target candidate to determine whether a sample should be labelled as positive or negative.Secondly,the goal of a classifier is to predict a binary label instead of a structured output.However,the objective for a tracker is to estimate the object’s transformation accurately.In Ref.[6],Ma et al.formulated the tracking problem as structured output prediction to mitigate the gap between binary classification and accurate target transformation determination.
When a new frame arrives,the ultimate goal for a tracker is to estimate the target position(a 2D rectangle)in the current frame.To capture target appearance variation,the classifier is updated online based on the newly estimated target appearance aroundand the corresponding samples∈The classifier is trained on the example pairs(x,z),where z=±1 is a binary label,and makes its prediction according to z=sign(h(x)),where h:X→R is the classification confidence function which maps from feature space X to a real target confidence value R.Let pt−1denote the estimated bounding box at time(t−1).The objective is to estimate a transformation pt∈Y,so the new position of the object is approximated by the composition pt=pt−1° yt.Y is the search space.Mathematically,the estimation process is converted to searching for the position change relative to the previous frame by solving:
To overcome the above two issues arising from traditional classifiers,we utilize the structured output SVM(SO-SVM)framework to estimate object location changes.The output space is thus the space of all transformations Y instead of the confidence labels.Thus we introduce an SO-SVM based discriminant function F:
SO-SVM performs a maximization step in order to predict the object transformation,while the discriminant function F includes the label y explicitly,meaning it can be incorporated into the learning algorithm.The model update procedure is performed on a labelled example pair
Function F measures the compatibility between(x,y)pairs,and gives high scores to those which are well matched. We restrict F to be of the form F(x,y)= 〈w,φ(x,y)〉,where φ(x,y)is a joint kernel map.The parameters can be learned in a large-margin framework from a set of example pairs{(x1,y1),...,(xn,yn)}by minimizing a convex objective function:
where δ(φi(y))= φ(xi,yi)− φ(xi,y),and Δ(yi,y)is a loss function.The value of Δ(yi,y)decreases towards 0 as y and yibecome more similar.The optimization aims to ensure the value of F(xi,yi)is greater than F(xi,y)for any y /= yi,by a margin which depends on a loss function Δ.The loss function plays an important role in our approach,as it allows us to address the issue raised previously of all samples being treated equally.
Using standard Lagrangian duality techniques,Eq.(8)can be converted into an equivalent dual form:
which leads to a much simpler expression for the dual problem in Eq.(11)and corresponding discriminant function Eq.(12):
3.4 Tracking algorithm
The proposed tracking algorithm is formulated within a particle filter framework.Given the observation setthe goal is to determine the maximize the a posteriori probabilityusing Bayes Theorem:
wheredenotes the target state translation (xt,yt)and scalest,andare the motion model that predicts statestbased on the previous statest−1and the likelihood of observation respectively.
The sparse feature vectorcin Eq.(5)is used as the object feature template.It is updated incrementally to accommodate appearance changes over time for robust visual tracking.We use temporal low-pass filtering to update the tracking model:
whereis a learning parameter,ctis the target template in thetth frame,andis the sparse representation of the tracked object in framet−1.Note that a significant innovation compared with the strategy in Ref.[1]is that the convolutional filters for extracting the object template are updated based on the newly tracked target:
模型中非期望产出的处理。在DEA模型的投入产出要素中,地区生产总值为期望产出,碳排放为非期望产出,期望产出越大越好,非期望产出越少越好,违反了方程的一致性,必须进行处理。本文以非期望产出作为投入的方法处理碳排放问题。
where clusterd(·)denotes a clustering operation withdclssses,andptare the image patches generated by a sliding window within the tracked object region.One of the advantages of this operation is that we can both preserve the original target appearance as well as capture new object variation,preventing target drift.
4 Experiments
We have evaluated our proposed tracking algorithm on a public dataset [34]which includes 50 video sequences categorized with 11 attributes based on different challenging factors including illumination variation(IV),out-of-plane rotation(OPR),scale variation(SV),occlusion(OCC),deformation(DEF),motion blur(MB),fast motion(FM),in-plane rotation(IPR),out-of-view(OV),background clutter(BC),and low resolution(LR).We compared the proposed tracking algorithm on OTB-50 against other methods including SCM[35],Struck’s method[9],TLD[20],MIL[3],and CT[19].In addition,we also compared our method with a state-of-the-art method convolutional network based tracker(CNT).For quantitative evaluation,we used a success plot and a precision plot for one-pass evaluation(OPE)protocols. All 50 videos were processed using the same parameter values during the tracking process,without modification.
Fig.2 One-pass evaluation.Top:average precision plot.Bottom:success rate plot.
The results in Fig.2 compare our tracking framework and CNT,Struck’s method,TLD,MIL,SCM,CT.It is clear that the combination of convolutional features and HOG features plays an important role in robust object tracking. For both success rate and precision rate for the OTB-50 dataset,our method achieves the maximum area under the curve. Unlike the CNT tracker,our tracking method updates the convolutional filter during the tracking process to capture target appearance variation,taking into account both the original target appearance and any target variation.In addition,the combination of HOG features and a structural output SVM improves the success and precision rates of our tracker by 13.4%and 12.2%respectively.In contrast,by adding convolutional features,our tracker enhances the performance of Struck’s method by 7.6%and 2.0%in terms of overall success rate and precision rate respectively.
To analyze the strengths and weaknesses of the proposed algorithm,we further evaluate the trackers on videos with 11 attributes.Figure 3 shows success rate plots for videos with different attributes,while Fig.4 shows the corresponding precision plots.In the success rate evaluation,our tracking algorithm ranks first in 8 out of 11 attributes.Meanwhile,for the video sequences with occlusion,deformation,and fast motion,our tracking method is ranked second,with the SCM and Struck trackers achieving the best performance–they employe useful background information to train discriminative classifiers.In the precision plots in Fig.4,our tracking algorithm is ranked first in 6 out of 11 attributes,namely scale variation,out-of-plane rotation,in-plane rotation,illumination variation,motion blur,and background cluster,while our tracker is ranked second for the other 5 attributes.
Figures 5 and 6 show some tracking results on some challenge image sequences for 7 trackers.Thebasketball,deer,andsoccervideo sequences contain illumination change,pose variation,and fast motion.In the above 3 sequences,the CNT tracker fails around frames 134,6,and 79,respectively. Atbasketballframe 531,all other trackers(CT,MIL,SCM,Struck,TLD)lose the target.Thecokeandfreeman4sequences contain significant out-of-plane rotation,occlusion,and pose variation. Tracking results on thefreeman4sequence show that most trackers drift away from the target when it is heavily occluded. These tracking results prove the effectiveness and robustness of the proposed featurere presentation(a combination of HOG and convolutional features)and structural output learning.The proposed tracking method can cope with target appearance variation in the tracking process by updating the object kernels over time.To make the tracker robust to target scale variation,we employ a combination of HOG descriptors and SOSVM to capture mid-level object cues.However,the time consumed is only 1.2 times greater than that used by the CNT tracker,and runs at 4.1 fps.
Fig.3 Tracker success rates for videos with different attributes,annotated with the area under the curve.The number in each title indicates the number of video sequences with a given attribute.
Fig.4 Tracker precision for videos with different attributes,annotated with the area under the curve.The number in each title indicates the number of video sequences with a given attribute.
Fig.5 Qualitative results using the proposed method on various challenging sequences(basketball,deer,liquor,soccer)having illumination variation.Frame number is shown at the top left of each frame in green.
Fig.6 Qualitative results using the proposed method on various challenging sequences(coke,doll,freeman4)having out-of-plane rotation.Frame number is shown at the top left of each frame in green.
5 Conclusions
In this paper,we have proposed a novel method to model target appearance with background and foreground convolutional filters for online tracking.To further improve tracking performance,we exploit the combination of hand-crafted features and structured output learning within a particle filter framework to jointly estimate target transformation and scale variation.Experimental results show that the proposed tracking method achieves excellent results in terms of both success rate and precision when compared to several state-of-the-art methods on public datasets.In the future,we hope to further exploit the convolutional feature representation at super-pixel level and use sparse representation to encode target appearance.
Acknowledgements
This work was supported by the National Natural Science Foundation of China (Nos.61403342,61273286,U1509207,61325019),and Hubei Key Laboratory of Intelligent Vision Based Monitoring for Hydroelectric Engineering(No.2014KLA09).
[1]Zhang,K.;Liu,Q.;Wu,Y.;Yang,M.-H.Robust visual tracking via convolutional networks without training.IEEE Transactions on Image ProcessingVol.25,No.4,1779–1792,2016.
[2]Henriques,J.F.;Caseiro,R.;Martins,P.;Batista,J.High-speed tracking with kernelized correlation filters.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.37,No.3,583–596,2015.
[3]Smeulders,A.W.M.;Chu,D.M.;Cucchiara,R.;Calderara,S.;Dehghan,A.;Shah,M.Visual tracking:An experimental survey.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.36,No.7,1442–1468,2014.
[4]Babenko,B.;Yang,M.-H.;Belongie,S.Robust object tracking with online multiple instance learning.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.33,No.8,1619–1632,2011.
[5]Kalal,Z.;Mikolajczyk,K.;Matas,J.Trackinglearning-detection.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.34,No.7,1409–1422,2012.
[6]Grabner,H.;Grabner,M.;Bischof,H.Real-time tracking via on-line boosting.In:Proceedings British Machine Vision Conference,Vol.1,47–56,2006.
[7]Grabner,H.;Leistner,C.;Bischof,H.Semi-supervised on-line boosting for robust tracking.In:Computer Vision–ECCV 2008.Forsyth,D.;Torr,P.;Zisserma,A.Eds.Springer Berlin Heidelberg,234–247,2008.
[8]Ma,Y.;Chen,W.;Ma,X.;Xu,J.;Huang,X.;Maciejewski,R.;Tung.A.K.H.EasySVM:A visual analysis approach for open-box support vector machines.Computational Visual MediaVol.3,No.2,161–175,2017.
[9]Hare,S.;Golodetz,S.;Saffari,A.;Vineet,V.;Cheng,M.-M.;Hicks,S.L.;Torr,P.H.Struck:Structured output tracking with kernels.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.38,No.10,2096–2109,2016.
[10]Ren,X.;Malik,J.Tracking as repeated figure/ground segmentation.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,1–8,2007.
[11]Zhou,X.;Li,Y.;He,B.;Bai,T.GM-PHD-based multi-target visual tracking using entropy distribution and game theory.IEEE Transactions on Industrial InformaticsVol.10,No.2,1064–1076,2014.
[12]Zhou,X.;Yu,H.;Liu,H.;Li,Y.Tracking multiple video targets with an improved GM-PHD tracker.SensorsVol.15,No.12,30240–30260,2015.
[13]Mei,X.;Ling,H.Robust visual tracking usingl1 minimization.In: Proceedings of the IEEE 12th International Conference on Computer Vision,1436–1443,2009.
[14]Simonyan,K.;Zisserman,A.Very deep convolutional networks for large-scale image recognition.arXiv preprintarXiv:1409.1556,2014.
[15]Han,B.;Comaniciu,D.;Zhu,Y.;Davis,L.S.Sequential kernel density approximation and its application to real-time visual tracking.IEEE Transactions onPatternAnalysis andMachine IntelligenceVol.30,No.7,1186–1197,2008.
[16]Jepson,A.D.;Fleet,D.J.;El-Maraghi,T.F.Robust online appearance models for visual tracking.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.25,No.10,1296–1311,2003.
[17]Ross,D.A.; Lim,J.; Lin,R.-S;Yang,M.-H.Incremental learning for robust visual tracking.International Journal of Computer VisionVol.77,Nos.1–3,125–141,2008.
[18]Zhong,W.;Lu,H.;Yang,M.-H.Robust object tracking via sparse collaborative appearance model.IEEE Transactions on Image ProcessingVol.23,No.5,2356–2368,2014.
[19]Zhang,K.;Zhang,L.;Yang,M.-H.Real-time compressive tracking.In:Computer Vision–ECCV 2012.Fitzgibbon,A.;Lazebnik,S.;Perona,P.;Sato,Y.;Schmid,C.Eds.Springer Berlin Heidelberg,864–877,2012.
[20]Kalal,Z.;Matas,J.;Mikolajczyk,K.P-N learning:Bootstrapping binary classifiers by structural constraints.In:Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,49–56,2010.
[21]Gao,J.;Ling,H.;Hu,W.;Xing,J.Transfer learning based visual tracking with Gaussian processes regression.In:Computer Vision–ECCV 2014.Fleet,D.;Pajdla,T.;Schiele,B.;Tuytelaars.Eds.Springer Cham,188–203,2014.
[22]Zhu,Z.;Liang,D.;Zhang,S.;Huang,X.;Li,B.;Hu,S.Traffic-sign detection and classification in the wild.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,2110–2118,2016.
[23]Li,H.;Li,Y.;Porikli,F.Robust online visual tracking with a single convolutional neural network.In:Computer Vision–ACCV 2014.Cremers,D.;Reid,I.;Saito,H.;Yang,M.-H.Eds.Springer Cham,194–209,2014.
[24]Zhou,X.;Xie,L.;Zhang,P.;Zhang,Y.An ensemble of deep neural networks for object tracking.In:Proceedings of the IEEE International Conference on Image Processing,843–847,2014.
[25]Fan,J.;Xu,W.;Wu,Y.;Gong,Y.Human tracking using convolutional neural networks.IEEE Transactions on Neural NetworksVol.21,No.10,1610–1623,2010.
[26]Wang,N.;Yeung,D.-Y.Learning a deep compact image representation for visual tracking. In:Proceedings of the Advances in Neural Information Processing Systems,809–817,2013.
[27]Wang,L.;Liu,T.;Wang,G.;Chan,K.L.;Yang,Q.Video tracking using learned hierarchical features.IEEE Transactions on Image ProcessingVol.24,No.4,1424–1435,2015.
[28]Avidan, S. Support vector tracking.IEEE Transactions onPatternAnalysis andMachine IntelligenceVol.26,No.8,1064–1072,2004.
[29]Collins,R.T.;Liu,Y.;Leordeanu,M.Online selection of discriminative tracking features.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.27,No.10,1631–1643,2005.
[30]Yang,F.;Lu,H.;Yang,M.-H.Robust superpixel tracking.IEEE Transactions on Image ProcessingVol.23,No.4,1639–1651,2014.
[31]Henriques,J.F.;Caseiro,R.;Martins,P.;Batista,J.Exploiting the circulant structure of tracking-bydetection with kernels.In:Computer Vision–ECCV 2012.Fitzgibbon,A.;Lazebnik,S.;Perona,P.;Sato,Y.;Schmid,C.Eds.Springer Berlin Heidelberg,702–715,2012.
[32]Elad,M.;Figueiredo,M.A.T.;Ma,Y.On the role of sparse and redundant representations in image processing.Proceedings of the IEEEVol.98,No.6,972–982,2010.
[33]Dalal,N.;Triggs,B.Histograms of oriented gradients for human detection.In:Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,Vol.1,886–893,2005.
[34]Wu,Y.;Lim,J.;Yang,M.-H.Online object tracking:A benchmark.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,2411–2418,2013.
[35]Zhong,W.;Lu,H.;Yang,M.-H.Robust object tracking via sparsity-based collaborative model.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,1838–1845,2012.
1 Zhejiang University of Technology,Hangzhou,310023,China.
2Tianjin University of Technology,Tianjin,300384,China.E-mail:csy@tjut.edu.cn
2017-02-27;accepted:2017-04-27
Junwei Li,Ph.D.,is with the College of Computer Science and Technology,Zhejiang University of Technology.He is a member of the China Computer Federation.His main research interests include object tracking, machine learning,convolutional neural networks,and object detection.
Xiaolong Zhou,Ph.D.and associate professor, is with the College of Computer Science and Technology,Zhejiang University of Technology.He is a member of the China Computer Federation,IEEE,and ACM.His main research interests are in visual tracking,gaze estimation,and pattern recognition.
SixianChan,Ph.D.,is with the College of Computer Science and Technology, Zhejiang University of Technology. Hismain research interests include visual tracking,image processing,pattern recognition,robotics,and image understanding.
Shengyong Chen,Ph.D.,professor.He is an IET fellow,an IEEE senior member,and a senior member of the China Computer Federation.His main research interests include computer vision,pattern recognition,and robotics.
Open AccessThe articles published in this journal are distributed under the terms of the Creative Commons Attribution 4.0 International License(http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use,distribution,and reproduction in any medium,provided you give appropriate credit to the original author(s)and the source,provide a link to the Creative Commons license,and indicate if changes were made.
Other papers from this open access journal are available free of charge from http://www.springer.com/journal/41095.To submit a manuscript,please go to https://www.editorialmanager.com/cvmj.
猜你喜欢
杂志排行
Computational Visual Media的其它文章
- An effective graph and depth layer based RGB-D image foreground object extraction method
- Vectorial approximations of infinite-dimensional covariance descriptors for image classification
- Editing smoke animation using a deforming grid
- Face image retrieval based on shape and texture feature fusion
- Z2traversal order:An interleaving approach for VR stereo rendering on tile-based GPUs
- Image-based clothes changing system