# Robust object tracking via online discriminative appearance modeling

- 204 Downloads

**Part of the following topical collections:**

## Abstract

A robust object tracking algorithm is proposed in this paper based on an online discriminative appearance modeling mechanism. In contrast with traditional trackers whose computations cover the whole target region and may easily be polluted by the similar background pixels, we divided the target into a number of patches and take the most discriminative one as the tracking basis. With the consideration of both the photometric and spatial information, we construct a discriminative target model on it. Then, a likelihood map can be got by comparing the target model with candidate regions, on which the mean shift procedure is employed for mode seeking. Finally, we update the target model to adapt to the appearance variation. Experimental results on a number of challenging video sequences confirm that the proposed method outperforms the related state-of-the-art trackers.

## Keywords

Visual tracking Mean shift Online learning Discriminative appearance## 1 Introduction

Visual tracking refers to the task of generating the trajectories of the moving objects in a sequence of images. It is a challenging problem in real-time computer vision due to variations of lighting condition, pose, scale, and view-point over time. In previous literature, a huge number of tracking methods have been proposed [1, 2, 3, 4, 5, 6, 7, 8, 9]. Most object trackers [10, 11, 12] search for the target in new frames with several key components: the first is object representation, such as using histogram [13, 14] or sparse representation [15] to model the appearance, using active contours to model the shape [16]; the second is a similarity measure between the reference model and candidate targets [17] and; the third is a local mode-seeking method for finding the most similar location in new frames, such as mean shift [18] or particle filter [19, 20]. Among these three components, appearance modeling of the target is of most importance for robust object tracking. However, it is exceptionally difficult to construct an appearance model with respect to all of those variations in advance.

Many tracking algorithms [16, 18, 20] are based on a fixed target model, and so are unable to track over long time intervals. To increase the robustness, some efforts have been made to employ online update algorithms to adapt to appearance changes of the target objects. Ross in [21] presents an adaptive tracking method which utilizes the incremental principal component analysis and shows robustness to large changes in pose, scale, and illumination. In [22], the authors present a method for evaluating multiple feature spaces while tracking, and a mechanism for online selection of discriminative features to improve tracking performance. The work of Grabner et.al. [23] and Parag [24] shows impressive results of using a classifier as implicit appearance model. They initially learn a binary classifier to distinguish the object of interest from the (neighboring) background and then apply it in each new frame to locate the position of the object. In [25], the authors propose a dynamic weights update mechanism for multiple cues tracking with detection responses as supervision. In [26], the authors consider visual tracking in a weakly supervised learning scenario where multiple imperfect oracles are fused to get a final accurate result. The accuracy of each tracker as well as the most likely object position are simultaneously evaluated by a probabilistic approach. A view-based subspace model is implemented in EigenTracking [27], but it requires intensive off-line learning before tracking.

From a different point of view, parametric density representations also have been used in many tracking algorithms. Han in [28] presents an online appearance modeling technique which is based on sequential density approximation and provides accurate and compact representations using Gaussian mixtures. McKenna in [29] suggests Gaussian mixture models created by an EM algorithm for histogram-based trackers, but their method requires knowledge of the number of components, which may not be known in advance. In [30, 31], a pixel-wise target model based on Gaussian distribution is proposed, and it is updated during tracking. However, this method cannot model multi-modal density functions accurately.

Other algorithms based on patches division have also been proposed. The fragment-based tracker [32] divides the target object into several regions and represents them with multiple local histograms. A vote map is used to combine the votes from all the regions in the target frame. In [33], the authors use a patch-based dynamic appearance model in junction with an adaptive Basin Hopping Monte Carlo sampling method to successfully track a non-rigid object. In [34], the authors propose a coupled-layer visual model that combines the target’s global and local appearance to address the problem of tracking objects which undergo rapid and significant appearance changes. The local layer in this model, similar as in [33], is a set of local patches that geometrically constrain the changes in the target’s appearance. This layer probabilistically adapts to the target’s geometric deformation, while its structure is updated by removing and adding the local patches. However, all above works do not consider the discriminative properties of the patches.

In this paper, we propose a robust object tracking algorithm based on an online discriminative appearance modeling mechanism. In contrast with traditional trackers whose computations cover the whole target region and may easily be polluted by the background pixels with similar feature to the foreground model, we divided the target object into a number of patches and take the most discriminative one as the tracking basis. To this patch, we consider both the photometric and spatial information and construct a discriminative target model on it. A likelihood map can be obtained by comparing the target model with candidate regions. Then, the mean shift procedure is employed for mode-seeking. Finally, the target model is updated to adapt to the appearance variation. The preliminary conference version of this work was presented in [35].

The rest of this paper is organized as follows. We review some related works in Section 2. Then, we briefly go over the mean shift framework in Section 3. In Section 4, the proposed discriminative learning-based tracking algorithm is described in detail. Experimental results on challenging video sequences are presented in Section 5, followed by conclusion in Section 6.

## 2 Related works

Given the dynamic nature of object tracking, having an online learning mechanism to update the target’s model is vital to tracking. A large body of work in the literature is dedicated to addressing this issue. In this part, we will review the significant steps that have been taken to achieve a robust target’s model.

Some early attempts (e.g., [36, 37]) updated the model by combining the old target model and the detected current appearance with a proper weight function. The weight function was foreseen to adjust the effect of the current appearance on the model. In [38], authors proposed a probabilistic target model and devised an updating scheme based on the EM algorithm. The probabilistic model took into account the stable part of the appearance (slowly varying image observations) and the possibility of losing the target due to occlusion, or noise in a unified framework for its decisions.

The concept of feature selection has been widely used for designing robust target models. A pioneer study along this school of thought is the work of Collins et al. where an online and discriminative feature selection scheme was introduced [39]. The idea of online learning for subspaces was developed by Ross et al. [40], where a low-dimensional subspace capturing the target appearance was incrementally updated using the past and current tracking results.

Inspired by the success of sparse coding in computer vision [41, 42], several sparse coding-based trackers have also been proposed [2, 3, 4, 43]. For the sparse trackers, the most popular approach for updating the target model is to learn the dictionary in an online fashion. For example, the differences between the current and previous target samples was used in [44] for adapting the dictionary. In [45], an online learning method for creating non-negative dictionaries was proposed. Since solving the optimization problems involving *ℓ*_{1} norms is computationally expensive, authors in [45] suggest to update the dictionary by gradient descent methods. In [46], a similar idea (gradient descent for updating the dictionary) was utilized albeit authors argued that a discriminative dictionary could be attained by utilizing two disjoint dictionaries to model foreground and background of the target.

## 3 The basic mean shift

*f*(

**x**), given

*n*sample points

**x**

_{i},

*i*=1,⋯,

*n*, in

*d*-dimensional space, the kernel density estimation (also known as Parzen window estimate) of

*f*(

**x**) can be written as

where *w*(**x**_{i})≥0 is the weight of the sample **x**_{i}, and *K*(**x**) is a radially symmetric kernel satisfying \(\int {k(x)dx=1}\). The bandwidth *h* defines the scale in which the samples are considered for the probability density estimation.

*h*can be calculated by mean shift method as follows:

where the kernel profile *k*(*x*) and *g*(*x*) have the relationship of *g*(*x*)=−*k*^{′}(*x*).

The kernel is recursively moved from the current location **x** to the new location *m*_{h}(**x**) according to mean shift vector, and finally converges to the nearest mode.

The mean shift hill climbing method is one of the most common methods that has been popular for years. After its introduction in the literature [47], it has been adopted to solve various computer vision problems, such as segmentation [48] and object tracking [18]. The popularity of the mean shift method is due its ease of implementation, real time response and robust tracking performance.

In the context of tracking, a sample corresponds to a pixel **x** and has an associated sample weight *w*(**x**), which defines how likely the pixel **x** belongs to an object. Given the initial object position, the traditional mean shift tracking method evaluates the new object position by computing the mean shift vector iteratively according to the Eq. (2). The bandwidth *h* defines the scale of the target candidate, i.e., the number of pixels considered in the localization process. In [18], the original mean shift tracker uses color histograms as an object representation and Bhattacharya coefficient as a similarity measure. An isotropic kernel is used as a spatial mask to smooth a histogram-based appearance similarity function between model and target candidate regions. The mean shift tracker climbs to a local mode of this smooth similarity surface to compute the translational offset of the target blob in each frame.

## 4 Method

Our goal in this section is to develop an efficient method that continually constructs and updates the discriminative target appearance model for tracking. The assumption we depend on is that the most informative object region for tracking is the same region that best discriminate between object and background classes. Due to the updating mechanism, local discrimination is sufficient for our work. The whole proposed tracking algorithm is described in detail as follows.

### 4.1 Discriminative target appearance modeling

Given a target region learned from previous views, we divide it into a number of patches from which we select the most discriminative one as the tracking basis. A larger ring of neighboring pixels surrounding the target region is chosen to represent the background. Let \({{\hat {X}}_{0}}\) denote the current location of the object and \(Y_{0}^{i}\) represent the *i*th patch, \(R_{0}^{i}\) indicate the relative position between the patch \(Y_{0}^{i}\) and the object \({{\hat {X}}_{0}}\), and *I*:**x**→**R**^{m} be the image that maps a pixel **x**=[*x**y*]^{T}∈**R**^{2} to a value, where the value is a scalar in the case of a grayscale image (*m*=1) or a three element vector for an RGB image (*m*=3).

We use the augmented variance ratio (AVR), the ratio of the between class variance to the within class variance, to measure the discriminative power of a patch as in [22]. For each patch, we compute the histogram on it as well as the background. By normalizing their histograms, we can get a discrete probability density *p*(*j*) for the patch, and density *q*(*j*) for the background, where index *j* ranges from 1 to *b*, the number of histogram buckets.

*j*can be given by

where *δ* is a small value (set to 0.001) that prevents dividing by zero or taking the log of zero. It is obvious that the log likelihood maps the object and background region into positive values for colors distinctive to the object, and negative for colors associated with the background. Colors that are shared by both object and background tend towards zero.

*L*(

*j*) can be computed to quantify the separability of the patch and background classes:

defines the variance of *L*(*j*) with respect to a discrete probability density function *a*(*j*).

Since we would like the log likelihood values of pixels on the object and background to both be tightly clustered while the two clusters should ideally be spread apart as much as possible, the denominator of the variance ratio enforces that the within class variances should be small for both object and background classes, while the numerator rewards cases where values associated with object and background are widely separated.

*man*sequence.

### 4.2 Likelihood map for tracking

*X*

_{l}, we can get the candidate model \({{T}_{l}}=({{X}_{l}},{{Y}_{l}},{{\hat {R}}_{0}})\) where

*Y*

_{l}is the respective patch with \({{\hat {Y}}_{0}}\) in

*X*

_{l}, determined according to \({{\hat {R}}_{0}}\). Then, we compute the likelihood of the location

*X*

_{l}. Let

*S*={(

*x*,

*y*,

*I*(

*x*,

*y*))|(

*x*,

*y*)∈

*Y*},

*S*∈

**R**

^{m+2}, denote the super patch of

*Y*, whose pixel is a vector containing the pixel coordinates coupled with their image measurements. This super form of patch enables us to both consider the photometric and the spatial information simultaneously. By warping patch \({{\hat {Y}}_{0}}\) to \({{\hat {S}}_{0}}\) whereas

*Y*

_{l}to

*S*

_{l}, the likelihood of the location

*X*

_{l}can be measured by

where DIS function returns the normalized sum of squared differences between the patch in the candidate region and that in the target model, and *λ* denotes the weighting parameter that is set to 25.

Then, a new image composed of likelihood values of all candidate locations becomes the “likelihood map” used for tracking.

### 4.3 Mode-seeking

*w*(

*X*

_{l}), we can evaluate the translation of the object centroid by computing the mean shift vector

*Δ*

*X*, such that \({{\hat {X}}_{1}}={{\hat {X}}_{0}}+\Delta X\), using the following:

*X*

_{l}is specified by:

### 4.4 Online update

The algorithm iterates through each subsequent frame of the video, extracting new discriminative patch of object, and constructing new target appearance model. However, adaptively updating target model in this manner may promote the occurrence of model drift. To avoid this problem, we take both the current observation and original reference model into account.

[DJWFFXGRAPHICS]s13634-019-0646-0flba[DJWFFXGRAPHICS]

where *I*(*x*,*y*) denotes the current image while *I*^{(ref)}(*x*,*y*) represents the reference image of initial frame, and the proportional coefficient *ω* decide how much one trusts the reference template versus the current observation. Based on the description above, the complete proposed algorithm can be summarized as Algorithm 1. Given an input image and the initial target position in the image, the proposed algorithm first divides the target region into several patches. For each patch, compute the distribution of the target and background and then compute the variance ratio. The patch with the largest variance ration is selected to reconstruct the target. At the next frame, we repeat this process and estimate the target position based on a target moving step.

## 5 Results and discussions

In this section, we use several challenging video sequences taken from moving cameras to illustrate the advantage of our proposed method. We use HSV color space, kernel with Epanechnikov profile and 15×15 patch size. We set the values of all parameters of the proposed method by manually adjusting their values to achieve a desired tracking performance.

Next, we use another three video sequences with different environment and tracking challenges to further evaluate the performance of the proposed algorithm. In the first video sequence, a tiger table lamp is held and swayed behind a bunch of plant. As the table lamp moves and opens its mouth to reveal the bulb, dramatic appearance change as well as severe occlusion occurs. The second video sequence describes a woman walking in the street, with many other stuff and sheltering cases in the background. The third video sequence is a high jump match, which contains a player with fast and drastic motion.

## 6 Conclusion

A robust object-tracking algorithm is proposed in this paper based on an online discriminative appearance modeling mechanism. By dividing the target into a number of patches, we extract the most discriminative piece of the target as the tracking basis. With the consideration of both the photometric and spatial information, a discriminative target model is constructed base on it. Then, a likelihood map can be obtained by comparing the target model with candidate regions, on which the mean shift procedure is employed for mode seeking. Experiment results have confirmed the effectiveness and robustness of our method.

## Notes

### Acknowledgements

This work acknowledged the Editor and anonymous Reviewers.

### Authors’ contributions

WL conceived the method and developed the algorithm. XS conceived the method, oversaw the project, and wrote the first draft. DL assembled formulations, and drafted the manuscript. All authors read and approved the final manuscript.;

### Funding

This research is supported in part by the National Natural Science Foundation (grant nos. 61602128 and 61702138) and the Shandong Province Natural Science Foundation (grant no. ZR2016FQ13) of China, and in part by the China Postdoctoral Science Foundation (grant no. 2017M621275 and 2018T110301), in part by the Young Scholars Program of Shandong University, Weihai (no.1050501318006).

### Ethics approval and consent to participate

Not applicable.

### Consent for publication

All the data (including individual details, images or videos) in the paper are all from public datasets.

### Competing interests

The authors declare that they have no competing interests.

## References

- 1.L. Zhang, W. Wu, T. Chen, N. Strobel, D. Comaniciu, Robust object tracking using semi-supervised appearance dictionary learning. Pattern Recogn. Lett.
**62:**, 17–23 (2015).CrossRefGoogle Scholar - 2.S. Zhang, H. Zhou, F. Jiang, X. Li, Robust visual tracking using structurally random projection and weighted least squares. IEEE Trans. Circ. Syst. Video Technol.
**25**(11), 1749–1760 (2015).CrossRefGoogle Scholar - 3.S. Zhang, X. Lan, H. Yao, H. Zhou, D. Tao, X. Li, A biologically inspired appearance model for robust visual tracking. IEEE Trans. Neural Netw. Learn. Syst.
**28**(10), 2357–2370 (2017).MathSciNetCrossRefGoogle Scholar - 4.Y. Qi, L. Qin, J. Zhang, S. Zhang, Q. Huang, M. -H. Yang, Structure-aware local sparse coding for visual tracking. IEEE Trans. Image Process.
**27**(8), 3857–3869 (2018).MathSciNetCrossRefGoogle Scholar - 5.S. Zhang, Y. Qi, F. Jiang, X. Lan, P. C. Yuen, H. Zhou, Point-to-set distance metric learning on deep representations for visual tracking. IEEE Trans. Intell. Transp. Syst.
**19**(1), 187–198 (2018).CrossRefGoogle Scholar - 6.X. Lan, S. Zhang, P. C. Yuen, R. Chellappa, Learning common and feature-specific patterns: A novel multiple-sparse-representation-based tracker. IEEE Trans. Image Process.
**27**(4), 2022–2037 (2018). https://doi.org/10.1109/TIP.2017.2777183.MathSciNetCrossRefGoogle Scholar - 7.Y. Yao, X. Wu, L. Zhang, S. Shan, W. Zuo, in
*Proceedings of the European Conference on Computer Vision (ECCV)*. Joint representation and truncated inference learning for correlation filter based tracking, (2018), pp. 552–567. https://doi.org/10.1007/978-3-030-01240-3_34.CrossRefGoogle Scholar - 8.Y. Qi, S. Zhang, L. Qin, H. Yao, Q. Huang, J. Lim, M. -H. Yang, Hedging deep features for visual tracking. IEEE Trans. Pattern Anal. Mach. Intell.
**41**(5), 1116–1130 (2019).CrossRefGoogle Scholar - 9.X. Lan, M. Ye, R. Shao, B. Zhong, P. C. Yuen, H. Zhou, Learning modality-consistency feature templates: A robust rgb-infrared tracking system. IEEE Trans. Ind. Electron.
**66**(12), 9887–9897 (2019). https://doi.org/10.1109/TIE.2019.2898618.CrossRefGoogle Scholar - 10.J. Kwon, K. M. Lee, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Visual tracking decomposition, (2010), pp. 1269–1276. https://doi.org/10.1109/cvpr.2010.5539821. - 11.C. Bao, Y. Wu, H. Ling, H. Ji, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Real time robust l1 tracker using accelerated proximal gradient approach, (2012), pp. 1830–1837. https://doi.org/10.1109/cvpr.2012.6247881. - 12.X. Sun, J. Zhang, Z. Xie, et al., Active-matting-based object tracking with color cues. SIViP (2014). https://doi.org/10.1007/s11760-014-0637-4.CrossRefGoogle Scholar
- 13.S. He, Q. Yang, R. W. H. Lau, J. Wang, M. H. Yang, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Visual tracking via locality sensitive histograms, (2013), pp. 2427–2434. https://doi.org/10.1109/cvpr.2013.314. - 14.S. M. N. Shahed, J. Ho, M. H. Yang, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Visual tracking with histograms and articulating blocks, (2008). https://doi.org/10.1109/cvpr.2008.4587575. - 15.X. Wang, Y. Wang, W. Wan, J. Hwang, Object tracking with sparse representation and annealed particle filter. SIViP (2014). https://doi.org/10.1109/icig.2013.81.
- 16.A. Elgammal, R. Duraiswami, L. Davis, in
*IEEE Conference on Computer Vision and Pattern Recognition*. Probability tracking in joint feature-spatial spaces, (2003), pp. 781–788. https://doi.org/10.1109/cvpr.2003.1211432. - 17.L. Sevilla-Lara, E. Learned-Miller, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Distribution fields for tracking, (2012), pp. 1910–1917. https://doi.org/10.1109/cvpr.2012.6247891. - 18.D. Comaniciu, V. Ramesh, P. Meer, Kernel-based object tracking. IEEE Trans. Pattern Anal. Mach. Intell.
**25**(5), 234–240 (2003).CrossRefGoogle Scholar - 19.K. Nummiaro, E. Koller-Meier, L. V. Gool, An adaptive color-based particle filter. Image Vis. Comput.
**21**(1), 99–110 (2003).CrossRefGoogle Scholar - 20.P. Perez, C. Hue, J. Vermaak, M. Gangnet, in
*IEEE Conference on European Conference on Computer Vision*. Color-based probabilistic tracking, (2002), pp. 661–675. https://doi.org/10.1007/3-540-47969-4_44.CrossRefGoogle Scholar - 21.D. Ross, J. Lim, R. Lin, M. Yang, Incremental learning for robust visual tracking. IEEE Trans. Int. J. Comput. Vis.
**77**(1-3), 125–141 (2008).CrossRefGoogle Scholar - 22.R. T. Collins, L. Yanxi, M. Leordeanu, Online selection of discriminative tracking features. IEEE Trans. Pattern Anal. Mach. Intell.
**27**(10), 1631–1643 (2005).CrossRefGoogle Scholar - 23.H. Grabner, H. Bischof, On-line boosting and vision (IEEE, 2006). https://doi.org/10.1109/cvpr.2006.215.
- 24.T. Parag, F. Porikli, A. Elgammal, Boosting adaptive linear weak classifiers for online learning and tracking, (2008). https://doi.org/10.1109/cvpr.2008.4587556.
- 25.G. Jia, Y. Tian, Y. Wang, T. Huang, M. Wang, in
*ACM Conference on Multimedia*. Dynamic multi-cue tracking with detection responses association, (2010), pp. 1171–1174. https://doi.org/10.1145/1873951.1874179. - 26.B. Zhong, H. Yao, S. Chen, R. Ji, X. Yuan, S. Liu, W. Gao, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Visual tracking via weakly supervised learning from multiple imperfect oracls, (2010), pp. 1323–1330. https://doi.org/10.1109/cvpr.2010.5539816. - 27.M. J. Black, A. Jepson, in
*IEEE Conference on European Conference on Computer Vision*. Eigentracking: Robust matching and tracking of articulated objects using a view-based representation, (1996), pp. 610–619. https://doi.org/10.1007/bfb0015548.Google Scholar - 28.B. Han, L. Davis, in
*IEEE Conference on International Conference on Computer Vision*. On-line density-based appearance modeling for object tracking, (2005), pp. 1492–1499. https://doi.org/10.1109/iccv.2005.181. - 29.S. McKenna, Y. Raja, S. Gong, Tracking colour objects using adaptive mixture models. Image Vis. Comput. J.
**17:**, 223–229 (1999).CrossRefGoogle Scholar - 30.B. Frey, in
*IEEE Conference on Computer Vision and Pattern Recognition*. Filling in scenes by propagating probabilities through layers into appearance models, (2000), pp. 185–192. https://doi.org/10.1109/cvpr.2000.855818. - 31.H. T. Nguyen, M. Worring, R. van den Boomgaard, in
*IEEE Conference on International Conference on Computer Vision*. Occlusion robust adaptive template tracking, (2001), pp. 678–683. https://doi.org/10.1109/iccv.2001.937587. - 32.A. Adam, E. Rivlin, I. Shimshoni, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Robust fragments-based tracking using the integral histogram, (2006), pp. 798–805. https://doi.org/10.1109/cvpr.2006.256. - 33.J. Kwon, K. M. Lee, in
*IEEE Conf. Computer Vision and Pattern Recognition*. Tracking of a non-rigid object via patch-based dynamic appearance modeling and adaptive basin hopping monte carlo sampling, (2009), pp. 1208–1215. https://doi.org/10.1109/cvprw.2009.5206502. - 34.L. Cehovin, M. Kristan, A. Leonardis, in
*IEEE International Conference on Computer Vision*. An adaptive coupled-layer visual model for robust visual tracking, (2011). https://doi.org/10.1109/iccv.2011.6126390. - 35.X. Sun, H. Yao, S. Zhang, B. Zhong, in
*IEEE Conference on Pervasive Computing Signal Processing and Applications*. On-line discriminative appearance modeling for robust object tracking, (2010), pp. 78–81. https://doi.org/10.1109/pcspa.2010.28. - 36.P. Pérez, C. Hue, J. Vermaak, M. Gangnet,
*Color-based probabilistic tracking, ECCV’*, (2002). https://doi.org/10.1007/3-540-47969-4_44.CrossRefGoogle Scholar - 37.A. Adam, E. Rivlin, I. Shimshoni, Robust fragments-based tracking using the integral histogram, 798–805 (2006). https://doi.org/10.1109/cvpr.2006.256.
- 38.A. Jepson, D. Fleet, T. EI-Maraghi, Robust online appearance models for visual tracking. TPAMI.
**25**(10), 1296–1311 (2003).CrossRefGoogle Scholar - 39.R. Collins, Y. Liu, M. Leordeanu, On-line selection of discriminative tracking features. TPAMI.
**27**(10), 1631–1643 (2005).CrossRefGoogle Scholar - 40.D. Ross, J. Lim, R. Lin, M. Yang, Incremental learning for robust visual tracking. IJCV.
**77:**, 125–141 (2008).CrossRefGoogle Scholar - 41.P. Wilf, S. Zhang, S. Chikkerur, S. A. Little, S. L. Wing, T. Serre, Computer vision cracks the leaf code. Proc. Natl. Acad. Sci. U.S.A.
**113**(12), 3305–3310 (2016). https://doi.org/10.1073/pnas.1524473113.CrossRefGoogle Scholar - 42.H. Zhu, X. Huang, S. Zhang, P. C. Yuen, Plant identification via multipath sparse coding. Multimed. Tools Appl.
**76**(3), 4599–4615 (2017). https://doi.org/10.1007/s11042-016-3538-4.CrossRefGoogle Scholar - 43.S. Zhang, X. Lan, Y. Qi, P. C. Yuen, Robust visual tracking via basis matching. IEEE Trans. Circ. Syst. Video Technol.
**27**(3), 421–430 (2017).CrossRefGoogle Scholar - 44.S. Zhang, H. Yao, H. Zhou, X. Sun, S. Liu, Robust visual tracking based on online learning sparse representation. Neurocomputing.
**100**(1), 31–40 (2013).CrossRefGoogle Scholar - 45.N. Wang, J. Wang, D. -Y. Yeung, Online robust non-negative dictionary learning for visual tracking. ICCV (2013). https://doi.org/10.1109/iccv.2013.87.
- 46.C. Gong, K. Fu, A. Loza, Q. Wu, J. Liu, J. Yang, Discriminative object tracking via sparse representation and online dictionary learning. IEEE Trans. Cybern.
**44**(4), 539–553 (2014).CrossRefGoogle Scholar - 47.K. Fukunaga, L. Hostetler, The estimation of the gradient of a density function, with applications in pattern recognition. IEEE Trans. Inf. Theory.
**21**(1), 32–40 (1975).MathSciNetCrossRefGoogle Scholar - 48.D. Comaniciu, P. Meer, in
*IEEE Conference on International Conference on Computer Vision*. Mean shift analysis and applications, (1999), pp. 1197–1203. https://doi.org/10.1109/iccv.1999.790416.

## Copyright information

**Open Access** This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.