1 Introduction

Visual target tracking is a subtask of computer vision, and many advanced methods have been explored in this research area. It has numerous applications in many domains, including visual navigation, intelligent video surveillance system, intelligent human–computer interaction, medical diagnosis. Deep learning demonstrates powerfulness in extracting and processing semantic features, and can model the appearance of object by learning multimedia information. Inspired by this, many successful applications of deep learning have been achieved in computer vision, such as image segmentation, object detection, image classification, target tracking, and so on.

Since 2013, the deep learning framework represented by SAE (stack auto-encoding) [1, 2], CNN (convolution neural network) [3, 4] and Siamese [5, 6] has become the main backbone network of tracking algorithm. Deep learning has been showing great success on object tracking. DLT [1] for the first time introduced deep network to break the bottleneck of traditional tracking model. After that, CNN has been brought to enhance the target learning capability of tracker [3], owing to the invariance principle in nonlinear changes such as translation, scale change and rotation. With the continuous research of depth structure, Tao et al. [5] successfully applied Siamese network as the backbone network of tracking algorithm and made greatly progress in speed. However, it is weakness in many practical applications due to challenges such as illumination changes, partial occlusion, motion blur and low resolution, which obstruct the robust of tracking model. In recent years, the optimization trend of visual target tracking focuses on deepening neural network and improving feature extraction strategy, but ignoring the key role of loss function in model optimization. In computer vision such as target detection, recognition and semantic segmentation, the loss function can measure the performance of training model by comparing the difference between predictive value and actual data.

In this paper, we take advantage of the recent progress in bounding box (BBox) regression loss and to propose a novel multivariate intersection over union (MIOU) loss in SiamRPN [7] tracking framework. The proposed method can deal with the non-overlapping case between target box and prediction box, and speed up the convergence rate of the training model. In summary, this work has the following steps. Firstly, MIOU regression includes three important geometric factors in BBox regression: overlapping area ratio, center distance ratio and the aspect ratio, which can better reflect the coincidence degree of the target box and prediction box. Secondly, we improve the definition of aspect ratio loss to avoid gradient explosion and improve the optimization performance of prediction box. Finally, extensive experiments on OTB2015 and VOT2016 benchmark are carried out to validate our method effectiveness.

2 Related work

It is difficult to design a tracking model with both strong robustness and high precision. Therefore, many theoretical methods have been introduced to solve the tracking problem, such as classifier [8,9,10], sparse representation [11,12,13], saliency detection [14, 15], feature selection [16,17,18,19] and deep learning [20, 21]. Based on off-line training and online fine-tuning, prior depth trackers achieve better results than traditional methods, and the online fine-tuning timely adjustment parameters to adapt the change of target better. However, despite the favorable performance of deep learning on object tracking, it is still limited by many difficulties, including insufficient training samples, the foreground-background class imbalance, and high computational complexity in terms of time and space. Therefore, online depth methods are hardly meeting the requirements of real-time tracking.

In recent years, Siamese network has been introduced to solve the tracking problem. As an end-to-end off-line training network, Siamese network learns the matching function from external data and finds the candidate patch matching the target in the subsequent frame search area. It can achieve real-time tracking without model updating or online fine-tuning. SiamFC [22] uses the Siamese structure and makes full convolution matching in the detection frame according to the template frame. The tracking speed reaches 86 fps, which has aroused widespread concern and accelerated the application of Siamese network in object tracking. In order to address the weakness of model robustness, SiamFC++ [23] proposed four guidelines: decomposition of classification and state estimation, non-ambiguous scoring, prior knowledge-free and estimation quality assessment, which effectively improved the generalization of the tracker. Graph convolutional tracking (GCT) [24] constructed a graph convolution tracking framework base on the Siamese structure, which acquired more sufficient and stable characteristic from detection frame by combining the temporal and spatial context information, and the experimental results showed that the accuracy is improved greatly.

SiamRPN [7] tracker contains Siamese network and region proposal network (RPN). RPN subnetwork uses multi-dimensional features to quickly generate target recommendation area, and obtains K anchor points according to different preset aspect ratio. The introduction of RPN makes the network not affected by multi-scale regression calculation in target tracking, and improves the tracking speed and accuracy. However, SiamRPN is vulnerable to the case of object occlusion, background clutters and motion blur. SiamRPN++ [25] mainly improves the performance of feature extraction network, solves the problem that the network deepening destroys the translation invariance, and realizes the Siamese tracking driven by ResNet network. DaSiamRPN [26] generates semantic negative sample pairs in the training process and expands the training dataset to solve the problem of poor system recognition caused by unbalanced distribution of training data. A new interference awareness module is designed to capture targets by using context information and time information. The SiamMask [27] enhances loss monitoring by adding binary segmentation task, thus reducing the distance between target tracking and Vos (visual object segment). The trained learning model can achieve class independent object tracking and segmentation only depending on an initial boundary box. The deeper and wider SiamRPN [28] designs deeper and wider backbone network to improve the capability of Siamese tracker.

Fig. 1
figure 1

Because \(\ell _1\)-smooth is sensitive to the scale of bounding box, which cause failing to reflect the intersection information between real box (black) and prediction box (red) in the same value. Moreover, GIOU loss is transformed to IOU loss when real box (black) surrounds the prediction box (red), owing to the heavily relying on intersection over union (IOU)

Although the trackers based on SiamRPN achieve good performance in many database evaluations, they use \(\ell _1\)-smooth [29] loss in location regression, which does not consider the correlation of the four corners of the bounding box, and multiple bounding boxes may have the same loss value. To alleviate the problem of class imbalance, Vital [30] adopts a high-order cost-sensitive loss to decrease the effect of easily negative samples successfully. Recently, the loss function of bounding box regression has been optimized. The n-normal form loss function represented by \(\ell _1\)-smooth [29] is very sensitive to the scale change of bounding boxes and cannot optimize the case of non-overlapping case, which is easy to cause the gradient to disappear. As shown in Fig. 1a, multiple detection boxes have the same \(\ell _1\)-smooth [29] loss value, but the IOU may vary greatly. In order to further improve the generalization performance of regression, scholars have proposed IOU loss [31] and GIOU loss [32]. When the prediction box and the target box do not intersect (non-overlapping), the IOU loss is 0. At this time, the loss function is not differentiable, so IOU loss cannot optimize the case of two boxes not intersecting. The GIOU loss can solve this problem, but because of the strong dependence on intersection over union, the convergence speed is slow. In reference [33], by directly minimizing the distance between the center points of two bounding boxes, a distance intersection over union (DIOU) is proposed, which solves the problem of slow convergence. In addition, the authors also proposed the complete intersection (CIOU) loss of three important geometric variables: overlap area ratio, center distance ratio and aspect ratio. However, CIOU [33] uses the square of the angle difference of aspect ratio to measure the scale loss, so it has the problems of gradient explosion and non-co-directional optimization of the border, like \(\frac{\partial \delta }{\partial {w}}=-\frac{h}{w}\frac{\partial \delta }{\partial {h}}\).

This paper analyzes the three factors that affect the location loss regression: the overlap area ratio, the center distance ratio and the aspect ratio of box height and width. We remove the square term of the angle difference corresponding to the aspect ratio, so as to avoid the gradient explosion problem and optimize the location regression performance better. The improved loss function (MIOU) is introduced into the regression branch of SiamRPN tracker, and achieves good performance.

The structure of the paper is as follows: firstly, the research background is sorted out in the introduction, and the related work is reviewed in the second part. Then, in the “Proposed method” section, we describe our method in detail, including the construction of network, the design of new geometric loss metrics, target class and target location. The experimental process and results are given in “Experimental results”. Finally, in the conclusion and prospect part, the work of this paper is summarized and prospected.

3 Proposed method

The framework of SiamRPN tracker contains Siamese subnetwork and region proposal subnetwork (RPN), where RPN network is constructed by two branches: classification loss and bounding box regression loss. We use ResNet50 [34] as the backbone instead of AlexNet in the original SiamRPN In training, ResNet50 pays more attention to rich semantic information, breaks the space invariance limit of connected subnetwork, and helps the tracker to better adapt to the scene of target appearance change. In addition, we propose a multivariate intersection over union (MIOU) loss to replace the \(\ell _1\)-smooth regression metric in the original RPN subnetwork and improve the tracking robustness.

Fig. 2
figure 2

Method network structure diagram

3.1 Network framework

As shown in Fig. 2, the target image Z(x) and the search region P(x) are input into two subnetworks of Siamese module, respectively. Meanwhile, they share the weights with the same structure during training. Considering the difference between classification and location, RPN is further divided into classification branch (cls) and regression branch (reg). In cls branch, we regard the classification problem as a qualitative output, and the regression problem is considered as a quantitative output in reg branch. So the outputs of Siamese subnetwork are fed into branch (cls) and (reg) individually. In detail, the classification branch convolutes \(p(x)_{cls}\) with \(z(x)_{cls}\) as convolution kernel, and the output channel number of \(A_{w\times h\times 2t}^{cls}\) is 2t, which indicates the positive and negative probability of candidate samples. Similarly, \(z(x)_{reg}\) and \(p(x)_{reg}\) produce the sensor \(A_{w\times h\times 4t}^{reg}\) of 4t channels after correlation operation. We refer the regression result \((d_x,d_y,d_w,d_h)\) as the four coordinates offsets of candidate targets. The specific operation process of the two tasks is as follows:

$$\begin{aligned} A_{w\times h\times 4t}^{cls}= & {} z(x)_{cls} \otimes p(x)_{cls}, \end{aligned}$$
(1)
$$\begin{aligned} A_{w\times h\times 4t}^{reg}= & {} z(x)_{reg} \otimes p(x)_{reg} . \end{aligned}$$
(2)

3.2 Classification loss

The classification loss of SiamRPN is cross-entropy (CE) loss. Cross-entropy method [35] is a unified method of reliability analysis and stochastic optimization design proposed by Rubinstein in 1997. Its essence is to transform the optimization problem into a small probability event estimation problem by using the optimal sampling probability density function instead of the original function of random variables based on Monte Carlo simulation. Cross-entropy can be directly used as the evaluation model of loss function, and the best training model is when the cross-entropy is minimum.

In the training, SiamRPN uses binary cross-entropy loss function for classification, assuming that the probability score of the \(i^{th}\) sample is \(p_{i}\). The tag value is \(y_{i}\) (\(y_{i} = 1\) means the sample is positive, otherwise, \(y_{i}= 0\)), and the calculation process is as follows:

$$\begin{aligned} L(y_{i},p_{i})=-((y_{i}\cdot \log (p_{i}))+(1-y_{i})(1-p_{i})) \end{aligned}$$
(3)

If the total number of samples in class branch is N, then the classification loss is:

$$\begin{aligned} L_{cls}=\frac{1}{N}\sum _{i=1}^{N}L(y_{i},p_{i}) \end{aligned}$$
(4)

3.3 Multivariate intersection over union

There are three important geometric factors in border regression: overlap area, center distance and aspect ratio. DIOU [33] does not include aspect ratio factor, and the metric of aspect ratio in CIOU [33] loss measures the scale consistency by the square of the angle difference between the two bounding boxes, which is similar to the \(L_2\) loss principle and has the problems of gradient explosion and instability. To solve the above problems, this paper redefines the loss formula to measure the aspect ratio, effectively avoids the gradient explosion, and improves the robustness.

In Fig. 3, let \(B_g\) and \(B_p\) represent the target box and prediction box, respectively, and the position of the box consisted of the coordinates of the two vertices in the lower left corner and the upper right corner, where \((\tilde{x_1},\tilde{y_1},\tilde{x_2},\tilde{y_2}) \) denotes the position of \(B_g\), and \((x_1,y_1,x_2,y_2) \) is represented by \(B_p\) . In addition, \(b_g\) and \( b_p\) are the center points of \(B_g\) and \(B_p\), and \(\rho \) represents the Euclidean distance of them. Noting that \(B_C\) denotes the smallest convex shape of \(B_g\) and \(B_p\), c denotes the diagonal Euclidean distance of \(B_C\).

The coordination \((X_{C1},Y_{C1},X_{C2},Y_{C2}) \) of \(B_C \) is as follows:

$$\begin{aligned} X_{C1}= & {} \min {(\tilde{x_1},x_1)},Y_{C1}=\min {(\tilde{y_1},y_1)}, \end{aligned}$$
(5)
$$\begin{aligned} X_{C2}= & {} \max {(\tilde{x_2},x_2)},Y_{C2}=\max {(\tilde{y_2},y_2)}. \end{aligned}$$
(6)

We use I to denote the maximum intersection box between \(B_g\) and \(B_p\) , and the coordinates \((X_{I1},Y_{I1},X_{I2},Y_{I2})\) of I come from the following formula:

$$\begin{aligned} X_{I1}= & {} \max {(\tilde{x_1},x_1)},Y_{I1}=\max {(\tilde{y_1},y_1)}, \end{aligned}$$
(7)
$$\begin{aligned} X_{I2}= & {} \min {(\tilde{x_2},x_2)},Y_{I2}=\min {(\tilde{y_2},y_2)}. \end{aligned}$$
(8)

The IOU is the ratio of the intersection and union of the area of two rectangular boxes; the calculation process is as follows:

$$\begin{aligned} A_g= & {} (\tilde{x_2}-\tilde{x_1})\times (\tilde{y_2}-\tilde{y_1}), \end{aligned}$$
(9)
$$\begin{aligned} A_p= & {} (x_2-x_1)\times (y_2-y_1), \end{aligned}$$
(10)
$$\begin{aligned} A_I= & {} (X_{I2}-X_{I1})\times (Y_{I2}-Y_{I1}), \end{aligned}$$
(11)
$$\begin{aligned} A_u= & {} A_g+A_p-A_I, \end{aligned}$$
(12)
$$\begin{aligned} IOU= & {} \frac{A_I}{A_u}, \end{aligned}$$
(13)

where \(A_g \) and \( A_p\) are the area of \(B_g\) and \(B_p\), respectively, \(A_I\) means the area of intersection between \(B_g\) and \(B_p\) , and \(A_u\) is the area formed by the union of two bounding boxes.

Fig. 3
figure 3

Schematic diagram of target box and prediction box

The formula of center distance ratio of target box and prediction box is as follows:

$$\begin{aligned} R_{dis}=\frac{\rho ^2}{c^2}, \end{aligned}$$
(14)

where \(\rho \) is the Euclidean distance of the center point of two boxes, and c is the diagonal distance of the smallest external rectangle. \(R_{dis}\) is the penalty term of the center point distance.

In order to take the aspect ratio of prediction frame into account, we define \(\theta \) reflects the difference of aspect ratio between \(B_g\) and \(B_p\), as shown in Fig. 4. And \(\theta _g\) denotes the inclination angle of the target box, while \(\theta _p\) represents the prediction box inclination angle. Let \(\theta _g=\arctan {\frac{w_g}{h_g}}\) and \(\theta _p=\arctan {\frac{w_p}{h_p}}\), where \(w_g\) and \(h_g\) represent the width and height of the target box, \(w_p\) and \(h_p\) are taken from the prediction box. In order to achieve the aspect ratio alignment between \(B_g\) and \(B_p\), we can see that \(\theta <0 \) when \(\theta _p<\theta _g\) in Eq.(15).

Fig. 4
figure 4

Schematic diagram of optimization in bounding box shape

$$\begin{aligned} \theta =\theta _p-\theta _g \end{aligned}$$
(15)

The formula of aspect ratio of target box and prediction box is as follows:

$$\begin{aligned} R_{asp}= & {} (\frac{\upsilon }{1-IOU+\upsilon })\delta , \end{aligned}$$
(16)
$$\begin{aligned} \upsilon= & {} \frac{4\theta ^2}{\pi ^2},\delta =\frac{8}{\pi ^2}\times \theta \times (w_p\times h_p) \end{aligned}$$
(17)

where \(\delta \) is used to evaluate the aspect ratios alignment of the bounding box. When the value of \(\delta \) is less than zero, it means prediction box \(B_p\) rotates counterclockwise during regression optimization. On the contrary, \(B_p\) rotates clockwise when \(\theta >0\), since \(\theta _p>\theta _g\). This optimization process in bounding box shape can be visualized in Fig. 4.

In Eq. (17), \(\delta \) is linearly related to angle difference \(\theta \) and box area \(w_p\times h_p\) to avoid gradient explosion. At the same time, in the process of scale optimization, there is no reverse relationship between the gradient \(\frac{\partial \delta }{\partial w}\) and \(\frac{\partial \delta }{\partial h}\). The gradient of \(\delta \) w.r.t. w and h is as follows:

$$\begin{aligned} \frac{\partial \delta }{\partial w}= & {} \frac{8}{\pi ^{2}}(arctan\frac{w_{p}}{h_{p}}-arctan\frac{w_{g}}{h_{g}}-\frac{wh}{w^{2}+h^{2}}) \end{aligned}$$
(18)
$$\begin{aligned} \frac{\partial \delta }{\partial h}= & {} \frac{8}{\pi ^{2}}(arctan\frac{w_{p}}{h_{p}}-arctan\frac{w_{g}}{h_{g}}+\frac{wh}{w^{2}+h^{2}}) \end{aligned}$$
(19)

Then, our loss function based on multivariate intersection over union (MIOU) is defined as follows:

$$\begin{aligned} MIOU = IOU-R_{dis}-R_{asp} \end{aligned}$$
(20)

The loss function of MIOU is defined as follows:

$$\begin{aligned} L_{MIOU}&=1-IOU+R_{dis}+R_{asp}\end{aligned}$$
(21)
$$\begin{aligned}&=1-IOU+\frac{\rho ^2}{c^2}+(\frac{\upsilon }{1-IOU+\upsilon })\delta \end{aligned}$$
(22)

Algorithm 1: Multivariate intersection over union metric as bounding box loss

Input: ground truth \(B_p(\tilde{x_1},\tilde{y_1},\tilde{x_2},\tilde{y_2})\) and prediction \((B_px_1,y_1,x_2,y_2)\) bounding box

Output: \(L_{MIOU}\)

1. Ensuring \(B_p\) meets the condition: \(x_2>x_1, y_2>y_1:\)

   \(x_1=\min (x_1,x_2),x_2=max(x_1,x_2)\),\(y_1=\min (y_1,y_2),y_2=max(y_1,y_2)\)

2. Calculating area of \(B_g\) and \(B_p\) in Eq.(9) and Eq.(10), getting \(A_g,A_p\).

3. Finding the coordinates of smallest enclosing box \(B_C\) in Eq.(5) and Eq.(6), getting

   \((X_{C1},Y_{C1},X_{C2},Y_{C2})\) and \(c^2=(X_{C2}-X_{C1})^2+(Y_{C2}-Y_{C1})^2\).

4. Calculating the center point of \(B_g\) and \(B_p\), \(b_g=(x_{b_g},y_{b_g}), b_p=(x_{b_p},y_{b_p})\):

   \(x_{b_g}=\frac{\tilde{x_1}+\tilde{x_2}}{2}, y_{b_g}=\frac{\tilde{y_1}+\tilde{y_2}}{2}\),\(x_{b_p}=\frac{x_1+x_2}{2}, y_{b_p}=\frac{y_1+y_2}{2}\)

5. Calculating the Euclidean distance between \(b_g\) and \(b_p\), \( \rho ^2=(x_{b_p}-x_{b_g})^2+(y_{b_p}-y_{b_g})^2\)

6. Calculating the center distance ratio \(R_{dis}\):   \(R_{dis}=\frac{\rho ^2}{c^2}\)

7. Finding the coordinates of intersection I between \(B_g\) and \(B_p\) in Eq.(7) and Eq.(8),getting:

   \((X_{I1},Y_{I1},X_{I2},Y_{I2})\)

8. Calculating area of \(A_I\), \(A_u\) in Eq.(11), Eq.(12), \(IOU=\frac{A_I}{A_u}\).

9. Calculating aspect ratio \(R_{asp}=(\frac{\upsilon }{1-IOU+\upsilon })\delta ,\) where \(\upsilon \) and \(\delta \) were calculated in Eq.(17).

10. Calculating the MIOU loss: \(L_{MIOU}=1-IOU+R_{dis}+R_{asp}\).

Table 1 The distribution of 11 challenging attributes in the OTB2015 dataset: illumination variation (IV), scale variation (SV), occlusion (Occ), deformation (Def), motion blur (MB), in-plane rotation (IPR), out-of-plane rotation (OPR), out of view (OV), background clutters (BC), low resolution (LR), and fast motion (FM)

In the bounding box regression loss, the overlap area ratio and center distance ratio reflect the relative position relationship between the target box and the prediction box. According to these two loss functions, we can guide the regression of the prediction bounding box and accelerate the convergence speed in the training stage. In addition, the aspect ratio of boxes can avoid the invalid regression in the case of non-overlapping or the case that the target box completely contains the prediction box, which has good scale invariance. Our method uses these three factors to carry out the bounding box regression, which avoids the gradient explosion problem, improves the convergence speed of the model training, and enhances the robustness of the tracker.

4 Experimental results

4.1 Experimental design

Since the training of Siamese network only needs image pairs, we use ILSVRC-VID dataset to train model and use OTB2015 dataset [36] to test model. ILSVRC-VID is the target detection dataset in ImageNet Large Scale Visual Recognition Competition. It includes 3862 snippets for training, 555 snippets for verification and 937 snippets for testing. Each snippet consists of 56,458 images. The ILSVRC-VID dataset has 30 categories, which are carefully selected, taking into account different scene factors, such as motion, video background interference, illumination changes, and so on. The size of the ILSVRC-VID dataset is 85GB, and the training time on the remote supercomputing server is about 3 days (CPU is 2*12 cores, Intel Xeon E5-2692 V2, 64GB memory, 1T disk storage).

The OTB2015 dataset is one of the standard datasets for target tracking, which consists of 100 fully annotated videos with 11 challenging attributes, including illumination variation (IV), scale variation (SV), occlusion (Occ), deformation (Def), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-plane (OPR), out-of-view (OV), background clutters (BC) and low resolution (LR). Table 1 shows the distribution of various challenge attributes in the OTB2015 dataset. Among them, the video test scenarios covered by SV and OPR attributes are relatively wide, accounting for more than half of the total dataset. Secondly, Occ, IPR and Def account for a relatively high proportion, indicating that OTB2015 pays more attention to the test of the target’s own deformation.

During training, for each video sequence, the target that comes from the first frame is regarded as template frame and the subsequent frame is put into search branch. Among them, the template branch adjusts the input image block size to \(127 * 127\) by using the convoluted operation, while the uniform scale of image block in searching branch is \(255 * 255\). Finally, according to the overlap area ratio calculation results, when iou value of candidate patches is greater than 0.6, it is judged as positive sample, while the iou value of negative samples is set to be no more than 0.3. The learning rate is initially set to \(5*10^{-3} \) and the number of anchors is 5. Since the target deformation difference is not obvious in the tracking process, the anchor aspect ratios are set to (0.33, 0.5, 1, 2, 3), while the anchor area is constant. Finally, a total of 20 epochs are performed.

4.2 Experimental analysis

Quantitative analysis: In the performance evaluation, we mainly compare our method against the four state-of-the-art metrics including \(\ell _1\)-smooth loss, IOU loss, GIOU loss and DIOU loss simultaneously. Firstly, We choose the average center location error as evaluation standard to quantify the performance of the methods. When the effective of tracker is better, the error value is lower, otherwise, the higher. To quickly validate the effectiveness of our proposed method, we only select 10 videos sequences from OTB2015 dataset. Table 2 shows the center error values in 10 videos, in which bold represents the best verification results. According to the results, our method performs better than \(\ell _1\)-smooth, IOU, GIOU and DIOU.

Table 2 The average center location error (aCLE) of different regression losses on SiamRPN tracker framework
Table 3 The average success, average precision and epoch of different regression losses in SiamRPN tracker framework on 100 sequences of OTB2015
Table 4 Comparison with the various method in terms of expected average overlap (EAO), robustness (failure rate), and accuracy on theVOT2016 benchmark
Fig. 5
figure 5

The comprehensive success plots (left) and precision plots (right) of comparison loss functions on 100 sequences in OTB2015

Fig. 6
figure 6

The success plots between different algorithms in 9 sequence challenges, including fast motion, occlusion, scale variation, motion blur, illumination variation, low resolution, deformation, out-of-view, out-of-plane rotation

Fig. 7
figure 7

Qualitative results of the proposed method (red), \(\ell _1\)-smooth loss (blue), IOU loss (cyan), GIOU loss (green) and DIOU loss (yellow) (football, subway, surfer, Skating1, skiing, liquor) on OTB2015

In Table 2, the scene attributes contained in the video sequences are labeled: illumination variation (IV), scale variation (SV), occlusion (Occ), deformation (Def), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-plane (OPR), out-of-view (OV), background clutters (BC) and low resolution (LR).

In order to further verify the influence of three variables (overlap area ratio, center distance ratio and aspect ratio) on boundary regression, we compared the tracking effect and training iterations under different regression variable combinations on SiamRPN tracker with 100 sequences of OTB2015, as shown in Table 3, where IOU, \(R_{dis}\) and \(R_{asp}\) represent the loss functions of overlap area ratio, center distance ratio and aspect ratio, respectively, and epoch is the number of iterations with the best performance during training. The smaller the number of iterations, the faster the convergence speed. Otherwise, the regression optimization takes a long time. It can be seen that the performance of SiamRPN+(1-IOU+\(R_{dis}\)) is better than that of the single one, and the number of iterations is reduced by one. The performance of SiamRPN+(1-IOU+\(R_{asp}\)) is improved in average precision, but the average success and epoch are not improved. The performance of SiamRPN+(1-IOU+\(R_{dis}\)+\(R_{asp}\)) (ours) is better than other models. This shows that \(R_{dis}\) plays a major role in accelerating convergence, and \(R_{asp}\) plays a major role in improving tracking accuracy, and the comprehensive performance of loss measure of three geometric variables is the best.

To further enhance the quantitative analysis, we use the vot2016 dataset to compare the tracker performance of our method with other algorithms. VOT2016 is a benchmark containing 60 video sequences. The common evaluation of VOT2016 includes accuracy (average overlap while tracking successfully), robustness (failure times) and expected average overlap (EAO). EAO is used to evaluated the overall performance which takes account of both accuracy and robustness. The bold in Table 4 means the better performance of method. In EAO and accuracy, the higher the value, the better the performance of the algorithm. On the contrary, the lower the robustness is, the less time the tracking fails, which means the method is more robust. Table 4 shows that our method (MIOU) is able to outperform the trackers in robustness and EAO.

Figure 5 shows the overall tracking success plots and precision plots for all 5 loss functions on 100 sequences in OTB2015. The success score and precision score of our approach are 0.610 and 0.845. The curves of these five methods are very close, but our method is \(0.7\%\) and \(2.2\%\) higher than the second method in success score and precision score, respectively. We set the error threshold of 20 pixels in precision plots, and the area under curve values of success plots represents the overlap rate between the prediction box and target bounding box.

Figure 6 shows success plot of different algorithms on 9 challenging attributes, including fast motion, occlusion, scale variation, motion blur, illumination variation, low resolution, deformation, out-of-view and out-of-plane rotation. Our method outperforms the other metrics trackers significantly in terms of 9 challenges, especially in occlusion, scale variation and motion blur, owing to provide regression direction in distance and shape of bounding box. Since GIOU loss and \(\ell _1\)-smooth loss have strong laziness on intersection over union calculation, it shows slow convergence and easy divergence of training. However, our method is less sensitive to the 9 challenges, which performs more generalization and robustness.

Qualitative analysis:

We illustrate the qualitative results in five different methods on a subset of 6 sequences in Fig. 7. The sequences of Football and Subway contain serious Occ (Occlusion), Def (deformation) and BC (Background Clutters), \(\ell _1\)-smooth [29] and DIOU [33] occur tracking drift in \(\sharp 109\) of Football, but our method, IOU [31] and GIOU [32] keep tracking successfully in the end. Sequences Surfer is typical of target SV (scale variation) and FM (fast motion), as we can see that GIOU has the problem of serious scale tracking failure, while other trackers perform well in these challenges. In sequences Skating1, many trackers suffer from short-term occlusion in \(\sharp 173\); however, our method and GIOU can effectively deal with the non-overlap case and reposition target in \(\sharp 200\). Moreover, our approach can identify the target in obvious illumination change. When IV (illumination change) and SV (scale variation) occur in skiing and liquor simultaneously, \(\ell _1\)-smooth [29], IOU [31], GIOU [32] and DIOU [33] are seriously affected by the susceptibility of scale, and leading to tracking failed. But our method has a good tracking effect in these cases and maintain long-term tracking. In general, the results clearly show that using our method as the bounding box regression loss performs consistently better in videos, while some failure cases occurred in \(\ell _1\)-smooth loss, IOU loss, GIOU loss and DIOU loss.

5 Conclusion and future work

SiamPRN tracking algorithm has real-time and excellent tracking performance, but it is prone to drift in the case of occlusion and non-overlapping, which is due to the \(\ell _1\)-smooth [29] loss of its bounding box regression branch. Therefore, this paper analyzes the tracking effects of IOU [31], DIOU and CIOU [33] regression loss functions in the SiamRPN framework, and proposes a multivariate intersection over union (MIOU) regression loss measurement method. MIOU uses the overlap area ratio, the center distance ratio and the aspect ratio of the boundary box, which has scale invariance and speeds up the convergence speed in the training process. On the other hand, we improve the definition of aspect ratio factor to adjust the scale alignment of the bounding box. Therefore, MIOU loss solves the problem of \(\ell _1\)-smooth loss localization failure in the case of non-overlapping, and can maintain long-term tracking. On the OTB2015 dataset and VOT2016 benchmark, experimental results show that the MIOU loss has better target location regression than other loss functions, especially for the challenges of occlusion, illumination change and fast motion.

We will further study this work. Firstly, similar objects are more sensitive to the interference of targets. We plan to combine the structural information of the target to improve the ability to identify distracters. Secondly, we find that the shape of RPN anchor has a great influence on the effectiveness of the model, so we will introduce adaptive feature fusion to refine the features based on the underlying anchor shapes.