Abstract
We propose an online multi-target tracker that exploits both high- and low-confidence target detections in a Probability Hypothesis Density Particle Filter framework. High-confidence (strong) detections are used for label propagation and target initialization. Low-confidence (weak) detections only support the propagation of labels, i.e. tracking existing targets. Moreover, we perform data association just after the prediction stage thus avoiding the need for computationally expensive labeling procedures such as clustering. Finally, we perform sampling by considering the perspective distortion in the target observations. The tracker runs on average at 12 frames per second. Results show that our method outperforms alternative online trackers on the Multiple Object Tracking 2016 and 2015 benchmark datasets in terms tracking accuracy, false negatives and speed.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Multi-target tracking-by-detection performs temporal association of target detections to estimate trajectories, while compensating for miss-detections and rejecting false-positive detections. Trajectories can be generated online [1], offline [2] or with a short latency [3]. Online trackers estimate the target state at each time instant as detections are produced. In case of miss-detections, online trackers may rely on predictive models to continue tracking until a matching detection is found [4]. Offline trackers use both past and future detections and can therefore better cope with miss-detections using re-identification [5].
An effective filter for online state estimation is the Probability Hypothesis Density (PHD) filter, which can cope with clutter, spatial noise and miss-detections [6, 7]. The PHD filter estimates the state of multiple targets by building a positive and integrable function over a multi-dimensional state whose integral approximates the expected number of targets [6, 8]. The posterior function can be computed based on a Bayesian recursion that leverages the set of (noisy) detections and it is approximated using Sequential Monte Carlo for computational efficiency via a set of weighted random samples (particles) [8]. This approximation is known as the PHD Particle Filter (PHD-PF) and involves four main steps [6–8]: the prediction of particles over time; the update of the weights of the particles based on new detections; the resampling step to avoid that only few particles monopolize the whole mass; and state estimation. A PHD filter needs an additional mechanism to provide target identity information. For example, the particles can be clustered and labeled after resampling to enable the temporal association of the clusters with previous states [7, 9, 10]. This additional mechanism is computationally expensive and error prone.
In this paper, we formulate an early association strategy between trajectories and detections after the prediction stage, which allows us to perform target estimation and state labeling without any additional mechanisms. Our online multi-target tracker exploits both strong (certain) and weak (uncertain) detections. Strong detections have a higher confidence score and are used for initialization and tracking. Weak detections have a lower confidence score and are used only to support the continuation of an existing track when strong detections are missing. We also introduce a perspective-dependent sampling mechanism to create newborn particles depending on their distance from the camera. Figure 1 shows the block diagram of the proposed online multi-target tracker.
In summary, our contributions include (i) a strategy for the effective exploitation of low-confidence (weak) detections; (ii) a procedure to label states via an early association strategy; (iii) the exploitation of perspective in prediction, update and newborn particle generation. The tracker works on average at 12 frames per second (fps) on an i7 3.40 GHz, 16 GB RAM computer, without any parallelization. We validate our tracking pipeline without using any appearance features and compare our method against state-of-the-art alternatives on the MOT15 and MOT16 benchmark datasets.
2 Strong and Weak Detections
Let a (large) set of target detections \(\mathbf {z}^*_k \in Z^*_k\), ideally without false negatives but potentially with multiple false positives, be generated at each time step k (Fig. 2(a)). These detections can be produced for example by running in parallel multiple detectors, by changing the operational parameters of a detector or with a combination of these two approaches. During this ‘over-detection’ process a target is likely to generate multiple overlapping detections. Overlapping detections produced by the same target may be combined into a single detection \(\mathbf {z}_k \in Z_k\) using, for example, non-maxima suppression [11, 12] (Fig. 2(b)) forming the set of combined detections \(Z_k\). Let each combined target detection be defined as
where \((x_{k}, y_{k})\) is the center and \((w_{k}, h_{k})\) are the width and height of the bounding box of the detection on the image plane. Let each \(\mathbf {z}_k\) be associated to a detection confidence-score \(s_{k} \in [0,1]\).
We categorize the set \(Z_k\) based on \(s_{k}\) into two subsets: strong (certain) and weak (uncertain) detections (Fig. 2(c)). This categorization can be obtained using the score confidences or via learning certain metrics on a training dataset [13]. Strong detections \(Z_k^+ = \{ \mathbf {z}_k^+ : s_{k} \ge \tau _s \}\), where \(\tau _s\) is a confidence threshold, are more likely to be true positives. We will use strong detections for trajectory initialization and for tracking existing targets. Weak detections \(Z_k^- = \{ \mathbf {z}_k^- : s_{k} < \tau _s \}\) are potential false positives. We will use weak detections for tracking existing targets to shorten the prediction time and to maintain the tracking uncertainty low. The value of \(\tau _s\) influences the ratio between the false positives and the false negatives, as we discuss in Sect. 6.1.
3 Perspective-Dependent Prediction
Let \(\Lambda _k\) be the set of existing identities at time k whose elements are \(\lambda \in \Lambda _k\). Let the state be defined as
where \((x_{k,\lambda }, y_{k,\lambda })\) is the center, \((\dot{x}_{k,\lambda }, \dot{y}_{k,\lambda })\) are the horizontal and vertical components of the velocity, \((w_{k,\lambda }, h_{k,\lambda })\) are the width and height and \(\lambda \) is the identity of the estimated state. Let the set of all estimated states at k be \(X_{k}\) whose elements are \(\mathbf {x}_{k,\lambda } \in X_k\). The elements of this set are obtained at each time step from the set of all existing particles \(\mathcal {X}_k\) whose elements are \(\mathbf {x}_{k,\lambda }^i \in \mathcal {X}_k\), where \(\mathbf {x}_{k,\lambda }^i\) is the \(i^{th}\) particle.
The prediction step assumes the motion of a target to be independent from the others and propagates each particle \(\mathbf {x}_{k-1,\lambda }^i\) as
where \(G_k\) is an affine transformation defined as
where \(\mathbf {0}\) and \(\mathbf {I}\) are the zero and identity matrices, respectively. \(N_k\) is an additive Gaussian noise defined as \(N_k = \left( n^x_k, n^{\dot{x}}_k, n^y_k, n^{\dot{y}}_k, n^w_k, n^h_k \right) ^T\), where each component of \(N_k\) is an independent Gaussian variable with zero mean and standard deviation proportional to the bounding box size in the previous frame.
As a target moving at constant velocity produces a smaller apparent displacement on the image plane when it is farther from the camera, we improve the model in Eq. 4 by considering the effect of foreshortening. To this end, we model \(N_k\) as a function of the distance from the camera. Specifically, we set the standard deviation of the noise for the horizontal and vertical components to be proportional to the width \(w_{k-1,\lambda }\) and height \(h_{k-1,\lambda }\) of the state, respectively.
In addition to the above, target acceleration variations, noisy detections and camera motion may generate erroneous predictions. To address these problems, instead of relying only on the previous time step [7], we average the past M states over a longer time interval \([t-M,t-1]\). Therefore, \(A_k\) and \(B_k\) dynamically update the position and velocity via the average velocity in the previous M frames:
where \(u_{k,\lambda }, v_{k,\lambda }\) are the average horizontal and vertical velocities of the estimated state \(\mathbf {x}_{k,\lambda }\), respectively, whose values are computed as
where \(M = \min \left( M_{k,\lambda }, M_{max} \right) \), \(M_{k,\lambda }\) is the number of time steps since the target \(\mathbf x _{k,\lambda }\) was initialized and \(M_{max}\) the maximum number of time steps.
The weights of the particles, \(\pi _{ k|k-1 }^{ i }\), are not modified during the prediction state, therefore
where \(L_{k-1}\) is the number of existing particles at \(k-1\).
4 Labeling
4.1 Early Association (EA)
The PHD-PF estimates the state of each target without labels (i.e. without identity) [9]. Let \(\pi _{k-1}^i\) be the weight associated to particle \(\mathbf {x}_{k-1, \lambda }^{ i } \). The PHD-PF posterior \(D_{k-1|k-1}(\cdot )\) is approximated as
where \(\delta (\cdot )\) is the Kronecker’s delta function. Various works have been published aiming to address the lack of identities in the PHD-PF: (i) clustering after resampling the particles on the right-hand side of Eq. 8 [7], (ii) keeping a separate tracker for each target and then perform ‘peak-to-track’ association [14], (iii) combining clustering techniques with the introduction of hidden identifiers to the samples of the PHD [8, 10, 15]. These solutions are computationally expensive and may introduce estimation errors. To avoid these problems, we move the association stage earlier in the pipeline.
We associate the elements of \(Z^+_k\) and \(Z^-_k\) to the predicted states using the Hungarian algorithm [16]. We refer to this association as early association because, unlike [7, 8, 15], it is performed before the update and resampling stages. The association cost, \(\omega _k\), between a detection \(\mathbf {z}_k\) and the predicted state \(\mathbf {x}_{k|k-1,\lambda }\) is
where \(d_l(\cdot )\) and \(d_s(\cdot )\) are the Euclidean distances between the position and bounding box size elements, respectively. \(Q_l\) is the diagonal of the image (i.e. the maximum position variation) and \(Q_s\) is the area of the image (i.e. the maximum size variation). Note that we multiply the normalized distances instead of averaging them to penalize when they are dissimilar (e.g. when two targets are far from each other in the scene but appear close to each other on the image plane).
When a trajectory is not associated to any (strong or weak) detections, the state is estimated using existing particles only. When the trajectory is not associated to any detections for a certain temporal interval (V frames, see Sect. 6.1), the state will be discarded before the EA and therefore the weight of its particles will gradually decrease toward zero.
EA enables the tracker to generate newborn particles that inherit the properties of its associated state (inheritance) or that produce a new identity (initialization).
4.2 Inheritance
Strong detections \(Z_k^+\) generate \(J_k\) newborn particles to repopulate the area around existing states. The newborn particles are added to the \(L_{k-1}\) existing particles. In [7, 8], the newborn particles are created from a newborn importance function \(p_k(\cdot )\) [7], which can be independently modeled from the estimated states, as a Gaussian process:
where \(|\cdot |\) is the cardinality of a set, \(\mathcal {N(\cdot )}\) is a Gaussian distribution and \(\Sigma \) is the covariance matrix. The covariance matrix can be dynamically updated based on parameters as detection size or video frame rate (see Sect. 6.1). Each newborn particle has an associated weight, \(\pi ^i_k\), defined as
where \(\gamma _k(\cdot )\) is the birth intensity, which is assumed to be constant when no prior knowledge about the scene is available [7]. Typically, \(J_k\) is chosen to have, on average, \(\rho \) particles per newborn target [8]. The process described in Eq. 10 could create newborn particles that are dissimilar from the corresponding state as they are independently created.
Unlike [15, 17] that included identities in the state, we consider the identity \(\lambda \in \Lambda _k\) as attribute of the state and propagate it over time. Therefore, \(|\Lambda _k|\) is the estimated number of targets at time k.
Let \(\hat{Z}^+_k\) and \(\hat{Z}^-_k\) be the sets that contain, respectively, strong and weak detections that are associated to one of the predicted states, i.e. to an existing trajectory. Let \(\hat{Z}_k =\hat{Z}^+_k \cup \hat{Z}^-_k\) be the set of detections that inherit the identity of the corresponding trajectories.
We create newborn particles from \(\hat{Z}_k\) and inherit properties from their associated predicted states: the position and bounding box size are created from detections, whereas velocity and identity are inherited from the associated states.
We use a mixture of Gaussians as importance function to sample position and bounding box size elements from the detections as
where
and
where \(\Sigma _k\) is a time-variant standard deviations matrix that is defined based on the size of the detection bounding box and the weight of the newborn particles is calculated as in Eq. 11. This solution allows us to address the perspective distortion during the generation of newborn particles. The values of these standard deviations are learned from a training dataset (see Sect. 6.1).
The velocities and the identity are inherited from the trajectory as
where \((\dot{x}_{k-1,\lambda }, \dot{y}_{k-1,\lambda } )\) are the velocity components of a trajectory \(X_{\lambda }\) (i.e. each state with identity \(\lambda \) for all k) and \(n_k^{\dot{x}}\) and \(n_k^{\dot{y}}\) are Gaussian noises that model the velocity variations of a target.
Figure 3 shows the benefit of weak detections. Without weak detections, miss-detections produce false negative trajectories and identity switches (Fig. 3, second row). When weak detections are used the targets are correctly tracked (Fig. 3, third row).
4.3 Initialization
While un-associated weak detections are discarded after EA, un-associated strong detections form the set \(\check{ Z }_{ k }=Z^+_k \backslash \hat{ Z }^+_k \) and initialize new target identities. Newborn particles associated to a new target are generated in a limited volume of the state space around the un-associated strong detections. The same new identity is assigned to each newborn particle.
We treat spawning targets as new targets. The newborn importance function \(p_k(\cdot )\) in Eq. 10 can regulate where targets are likely to spawn or enter in a scene [18]. Each detection in \( \check{ Z }_{ k }\) initializes a new trajectory and generates newborn particles using a mixture of Gaussians as
where \(\Sigma _k\) is defined in Eq. 14 and the weights of the particles are calculated as in Eq. 11.
Figure 4 shows an example of how newborn particles are created. The target on the right is initialized because of the presence of an un-associated strong detection. The target with identity number 2 is localized with a weak detection. The weak detection in this case is a false positive that is discarded because it is not associated with any predicted states.
5 Perspective-Dependent Update, Resampling and State Estimation
Let the set of particles that share the same identity be \(\mathcal {X}_{ k,\lambda }\) whose elements are \(\mathbf { x }_{ k,\lambda }^{ i } \in \mathcal {X}_{ k,\lambda }\). After new detections are generated, the weights of the particles, \(\pi _k^i\), are recalculated for allowing the particles to update the estimation [7, 9, 19].
The weights at k are updated as
where \(p_M\) is the probability of miss-detection, \(\kappa _k (\cdot )\) is the clutter intensity associated to a detection \(\mathbf {z}_k\) [7, 19], \(C_k(\mathbf {z}_k)\) is defined as
and \(g_k(\mathbf {z}_k | \mathbf {x}_{k,\lambda }^i)\) is the likelihood function defined as
where C is defined in Eq. 13. The likelihood function, \(g_k(\cdot )\) computes the location and bounding box similarities. Unlike [7, 10] where \(\Sigma =\Sigma _k\) is fixed, we define \(\Sigma _k\) in Eq. 14 as a time-variant matrix that regulates the location and bounding box similarity between particles and detections (i.e. the particles of an object far from the camera will be less spread than those of a closer object due to the perspective). Figure 5 shows examples of the use of the proposed perspective-dependent approach.
After the update step, resampling helps avoiding the degeneracy problem [20]. The standard multinomial resampling [8, 20] splits particles proportionally to their weights, frame-by-frame independently. Because newborn particles have in general a lower weight than existing particles, new targets may not be initialized due to repetitive deletion of their particles during resampling. To allow newborn particles to grow over time and reach a comparable weight to that of existing particles, newborn particles are resampled independently from existing particles using a Multi-stage Multinomial Resampling step [7]. Finally, each state \(\mathbf { x }_{ k, \lambda } \in X_k\) is estimated as the average of all resampled particles sharing the same identity:
6 Results
6.1 Experimental Setup
We validate the proposed trackerFootnote 1, the Early Association Probability Hypothesis Density Particle Filter (EA-PHD-PF), and compare it against state-of-the-art online tracking methods on the MOT15 and MOT16 benchmark datasets (motchallenge.net) [21, 22]. We use the public detections provided by the MOT benchmark and our private detections produced by combining detections from state-of-the-art person detectors. We refer to the tracker using the public detections from MOT benchmark as EA-PHD-PF(Pub) and to the tracker using the private detections as EA-PHD-PF(Priv).
In the specific implementation presented here, the combined detection has position and bounding box size equal to the weighted average of the position and bounding box size of the detections that contributed to the combination. We use detections generated by Discriminatively Trained Deformable Part Models (DTDPM) [12], Scale Dependent Pooling (SDP) [23], Aggregate Channel Features (ACF) [24] trained on INRIA (ACF-I) and Caltech (ACF-C) datasets. We reward detections generated by the combination of a larger number of detectors (possible true positives) and penalize isolated detections (possible false positives). We normalize the confidence score of each detector using the 99\(^{th}\) percentile of the detection scores generated by each detector over the training set (and truncating to 1). Then we combine all detections via voting when their overlap area divided by their union area exceeds \(\tau _f=1/3\). Given the normalized detection confidence of each detection, \(s_j\), the confidence score of the combined detection, \(s_{k} \in [0,1]\), is \(s_{k} = \frac{U}{D^2}{\sum _{j=1}^{U}s_j}\), where U is the number of contributing detectors and D is the total number of detectors.
We allow to perform association between detections and predicted states only if their overlap area divided by their union area exceeds \(\tau _a=1/3\). The parameter that controls when a trajectory will not seek for more detections is \(V={{\lceil }{f}{\rceil }}/{1s}\), where f is the frame-rate of the video sequence. The parameter that controls the maximum possible number of frames to consider in the prediction model is \(M_{ max }={{\lceil }{f/2}{\rceil }}/{1s}\).
We train the parameters of our method on the MOT15 and MOT16 training datasets and then use these parameters in MOT15 and MOT16 testing sequences, respectively. For the set of public detections \(\tau _s=0.39\) in MOT15 and \(\tau _s=0.20\) in MOT16. For the set of private detections \(\tau _s=0.35\) in both datasetsFootnote 2. The number of particles per target, \(\rho \), is 500. The standard deviation values used for the prediction, update and newborn particle generation are modelled as a function of the bounding box size as
where \(g \in [1,G]\) indicates a state element of a ground-truth trajectory, \(std(\cdot )\) is the standard deviation operation, \(\frac{ d^2(\cdot ) }{ dk^2 }\) is the second derivative that quantifies the noise in the variation of x, y, w and h over time, and \(\frac{ d^3(\cdot ) }{ dk^3 }\) is the third derivative that quantifies the noise in the variation of \(\dot{x}\) and \(\dot{y}\) over time. We use the bounding box size at time k, \(w_{ x_{ k } }\) and \(h_{ x_{ k } }\), in order to adapt the noise to the scale of the bounding box. Note that estimated states are used in the prediction step (Sect. 3), whereas detections are used in the generation of newborn particles (Sect. 4) and update step (Sect. 5).
The evaluation measures are Multiple Object Tracking Accuracy (MOTA), Multiple Object Tracking Precision (MOTP) [25], False Alarm per Frame (FAF), Mostly Tracked targets (MT), Mostly Lost targets (ML) [26], Fragmented trajectories (Frag), False Positives (FP), False Negatives (FN), Identity Switches (IDS) and tracker speed in Hz. For a detailed description of each metric, please refer to the MOT website and [21].
6.2 Discussion
Table 1 compares the tracking results of our proposed method using both public and private detections with other online trackers submitted to the MOT15 and MOT16 benchmarkFootnote 3. The upper part of the table (MOT15) shows that EA-PHD-PF(Priv) outperforms AMPL, LKDAT_CNN, MDP_SubCNN and justry, in terms of MOTA. The number of FN and the ML percentage are overall lower than the other trackers. This is due to the ability of EA-PHD-PF(Priv) to robustly perform state estimation exploiting weak detections without relying on the prediction only when (strong) detections are missing. The higher number of IDS compared to the other methods is due to the fact that we rely only on the position and size of the bounding box inferred from the detections and we are not using any appearance models to discriminate nearby targets. Moreover, we do not model spawning targets. Therefore, identity switches are more likely in crowded scenes, as shown in Fig. 6. The bottom part of Table 1 (MOT16) shows that EA-PHD-PF(Priv) outperforms AMPL, olCF and OVBT in terms of MOTA, FN and FP. However, the number of IDS is higher than AMPL and olCF, because the features they use are better able to discriminate targets. The results using public detections rank our tracker EA-PHD-PF(Pub) at half-rank overall as it generates a high amount of FN.
Figure 7 compares sample tracking results using public (first row) and private (second row) detections. We can observe along the first row how the target firstly initialized as 115 is then reinitialized, lost and reinitialized again due to the high number of FN in the public dataset. However, the (same) target firstly initialized as 41 in the second row is correctly tracked along the whole sequence. We can observe the presence of false-positive trajectories (i.e. green, purple and red targets in Fig. 7b). These false-positive trajectories are difficult to remove because they are caused by persistent false-positive detections appearing for a few consecutive frames and the confidence scores of those detections are as high as those of true positive detections. With EA-PHD-PF(Priv) these detections are filtered out without adding any false-negative trajectories.
7 Conclusion
We presented an online multi-target tracker that exploits strong and weak detections in a Probability Hypothesis Density Particle Filter framework. Strong detections are used for trajectory initialization and tracking. Weak detections are used for tracking existing targets only to reduce the number of false negatives without increasing the false positives. Moreover, we presented a method to perform early association between trajectories and detections, which eliminates the need for a clustering step for labeling. Finally, we exploited perspective information in prediction, update and newborn particle generation. Results show that our method outperforms alternative online trackers on the Multiple Object Tracking 2016 and 2015 benchmark datasets in terms tracking accuracy, false negatives and speed. The tracker works at an average speed of 12 fps. Future work will involve using appearance features, such as color histograms, to reduce trajectory fragmentation.
Notes
- 1.
Results are available at: http://www.eecs.qmul.ac.uk/~andrea/eamtt.html.
- 2.
Larger values of \(\tau _s\) reduce the number of false positives and lead to a more conservative initialization of the trajectories.
- 3.
Last accessed on \(10^{th}\) August 2016.
References
Solera, F., Calderara, S., Cucchiara, R.: Learning to divide and conquer for online multi-target tracking. In: Proceedings of International Conference on Computer Vision, Santiago, CL, December 2015
Wang, B., Wang, G., Chan, K., Wang, L.: Tracklet association with online target-specific metric learning. In: Proceedings of Computer Vision and Pattern Recognition, Columbus, OH, USA, June 2014
Poiesi, F., Cavallaro, A.: Tracking multiple high-density homogeneous targets. IEEE Trans. on Circ. Syst. Video Technol. 25(4), 623–637 (2015)
Possegger, H., Mauthner, T., Roth, P., Bischof, H.: Occlusion geodesics for online multi-object tracking. In: Proceedings of Computer Vision and Pattern Recognition, Columbus, OH, USA, June 2014
Shu, G., Dehghan, A., Oreifej, O., Hand, E., Shah, M.: Part-based multiple-person tracking with partial occlusion handling. In: Proceedings of Computer Vision and Pattern Recognition, Rhode Island, USA, June 2012
Mahler, R.: A theoretical foundation for the Stein-Winter Probability Hypothesis Density (PHD) multitarget tracking approach. In: Proceedings of MSS National Symposium on Sensor and Data Fusion, San Diego, CA, USA, June 2002
Maggio, E., Taj, M., Cavallaro, A.: Efficient multitarget visual tracking using random finite sets. IEEE Trans. Circ. Syst. Video Technol. 18(8), 1016–1027 (2008)
Vo, B.N., Singh, S., Doucet, A.: Sequential Monte Carlo implementation of the PHD filter for multi-target tracking. In: Proceedings of Information Fusion, vol. 2, Queensland, AU, July 2003
Mahler, R.: PHD filters of higher order in target number. IEEE Aerosp. Electron. Syst. Mag. 43(4), 1523–1543 (2007)
Panta, K., Vo, B., Singh, S.: Improved probability hypothesis density (PHD) filter for multitarget tracking. In: 2005 3rd International Conference on Intelligent Sensing and Information Processing, December 2005
Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: Proceedings of Computer Vision and Pattern Recognition, San Diego, CA, USA, June 2005
Felzenszwalb, P.F., Girshick, R., Ramanan, D.: Object detection with discriminatively trained part based models. IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1627–1645 (2010)
Smedt, F.D., Goedeme, T.: Open framework for combinated pedestrian detection. In: Proceedings of Computer Vision, Imaging and Computer Graphics Theory and Applications, Berlin, GE, March 2015
Lin, L., Bar-Shalom, Y., Kirubarajan, T.: Data association combined with the Probability Hypothesis Density Filter for multitarget tracking. In: Proceedings of SPIE, August 2004
Panta, K., Vo, B.N., Singh, S., Doucet, A.: Probability Hypothesis Density filter versus multiple hypothesis tracking. In: Proceedings of SPIE, August 2004
Kuhn, H., Yaw, B.: The Hungarian method for the assignment problem. Naval Res. Logistics Q. 2, 83–97 (1955)
Poiesi, F., Mazzon, R., Cavallaro, A.: Multi-target tracking on confidence maps: an application to people tracking. Comput. Vis. Image Underst. 117(10), 1257–1272 (2013)
Maggio, E., Cavallaro, A.: Learning scene context for multiple object tracking. IEEE Trans. Image Process. 18(8), 1873–1884 (2009)
Vo, B.N., Singh, S., Doucet, A.: Sequential Monte Carlo methods for multitarget filtering with random finite sets. IEEE Aerosp. Electron. Syst. Mag. 41(4), 1224–1245 (2005)
Arulampalam, M., Maskell, S., Gordon, N., Clapp, T.: A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Sig. Process. 50(2), 174–188 (2002)
Leal-Taixé, L., Milan, A., Reid, I., Roth, S., Schindler, K.: MOTChallenge 2015: towards a benchmark for multi-target tracking, April 2015. arXiv:1504.01942 [cs]
Milan, A., Leal-Taixé, L., Reid, I., Roth, S., Schindler, K.: MOT16: a benchmark for multi-object tracking, March 2016. arXiv:1603.00831 [cs]
Yang, F., Choi, W., Lin, Y.: Exploit all the layers: fast and accurate CNN object detector with scale dependent pooling and cascaded rejection classifiers. In: Proceedings of Computer Vision and Pattern Recognition, Las Vegas, NV, USA, June 2016
Dollar, P., Appel, R., Belongie, S., Perona, P.: Fast feature pyramids for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 36(8), 1532–1545 (2014)
Kasturi, R., Goldgof, D., Soundararajan, P., Manohar, V., Garofolo, J., Bowers, R., Boonstra, M., Korzhova, V., Zhang, J.: Framework for performance evaluation of face, text, and vehicle detection and tracking in video: data, metrics, and protocol. IEEE Trans. Pattern Anal. Mach. Intell. 31(2), 319–336 (2009)
Li, Y., Huang, C., Nevatia, R.: Learning to associate: HybridBoosted multi-target tracker for crowded scene. In: Proceedings of Computer Vision and Pattern Recognition, Miami, FL, USA, June 2009
Xiang, Y., Alahi, A., Savarese, S.: Learning to track: online multi-object tracking by decision making. In: Proceedings of International Conference on Computer Vision, Santiago, CL, December 2015
Acknowledgements
This work was supported in part by the ARTEMIS JU and the UK Technology Strategy Board (Innovate UK) through the COPCAMS Project, under Grant 332913.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Sanchez-Matilla, R., Poiesi, F., Cavallaro, A. (2016). Online Multi-target Tracking with Strong and Weak Detections. In: Hua, G., Jégou, H. (eds) Computer Vision – ECCV 2016 Workshops. ECCV 2016. Lecture Notes in Computer Science(), vol 9914. Springer, Cham. https://doi.org/10.1007/978-3-319-48881-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-319-48881-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-48880-6
Online ISBN: 978-3-319-48881-3
eBook Packages: Computer ScienceComputer Science (R0)