Skip to main content
Log in

TVVS: A top-view visualization system from broadcasting soccer video

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

A holy grail for sports analytics is the top-view visualization of the game. The top-view visualization provides the actual between-player distances as opposed to the between-player distances calculated from the side and/or oblique view of a match. Related work in this area relies on multiple camera installations in the stadium or directly derive the registration map between a broadcasting video and the top-view model. Aberrating the state-of-the-art, a factor theory based approach is presented to derive the top-view visualization of the game from the broadcasting sports video. It is theoretically proved that the proposed factor theory based approach is more efficient than the state-of-the-art approach for the top-view visualization. In addition, as per the proposed approach, a model is presented for the top-view visualization by transforming the broadcasting video into a single and static camera visualization. In order to generate the single-camera visualization, the view of the entire ground is needed which is expressed as a solution to a convex optimization function, devised to explore putative matrix completions. To give pristine empirical evidence, the benchmark dataset is used and a soccer dataset has been introduced towards the end. The proposed top-view approach brings atleast 7% and 10% gains over the state-of-the-art on the benchmark and the proposed dataset respectively.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Notes

  1. https://pjreddie.com/darknet/yolo/

  2. https://pjreddie.com/darknet/install

  3. https://nhoma.github.io/

  4. https://nhoma.github.io/

  5. https://github.com/vcg-uvic/sportsfield_release

  6. https://github.com/lood339/SCCvSD

References

  1. Baker S, Matthews I (2004) Lucas-kanade 20 years on: a unifying framework. International journal of computer vision 56(3):221–255

    Article  Google Scholar 

  2. Cai JF, Candès EJ, Shen Z (2010) A singular value thresholding algorithm for matrix completion. SIAM Journal on optimization 20(4):1956–1982

    Article  MathSciNet  Google Scholar 

  3. Candès EJ, Recht B (2009) Exact matrix completion via convex optimization. Foundations of Computational mathematics 9(6):717

    Article  MathSciNet  Google Scholar 

  4. Carr P, Sheikh Y, Matthews I (2012) Point-less calibration: Camera parameters from gradient-based alignment to edge images. In: 2012 IEEE Workshop on the applications of computer vision (WACV), pp. 377–384. IEEE

  5. Chen J, Little JJ (2019) Sports camera calibration via synthetic data. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 0–0

  6. Dubrofsky E, Woodham RJ (2008) Combining line and point correspondences for homography estimation. In: International symposium on visual computing, pp. 202–213. Springer

  7. Farin D, Krabbe S, Effelsberg W, et al. (2003) Robust camera calibration for sport videos using court models. In: Storage and retrieval methods and applications for multimedia 2004, vol. 5307, pp. 80–91. International society for optics and photonics

  8. Germann M, Popa T, Keiser R, Ziegler R, Gross M (2012) Novel-view synthesis of outdoor sport events using an adaptive view-dependent geometry. In: Computer graphics forum, vol. 31, pp. 325–333. Wiley online library

  9. Ghanem B, Zhang T, Ahuja N (2012) Robust video registration applied to field-sports video analysis. In: IEEE International conference on acoustics, speech, and signal processing (ICASSP), vol. 2. Citeseer

  10. Gupta A, Little JJ, Woodham RJ (2011) Using line and ellipse features for rectification of broadcast hockey video. In: 2011 Canadian conference on computer and robot vision, pp. 32–39. IEEE

  11. Halko N, Martinsson PG, Tropp JA (2011) Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions. SIAM review 53(2):217–288

    Article  MathSciNet  Google Scholar 

  12. Harris CG, Stephens M, et al. (1988) A combined corner and edge detector. In: Alvey vision conference, vol. 15, pp. 10–5244. Citeseer

  13. Hartley R, Zisserman A (2003) Multiple view geometry in computer vision Cambridge university press

  14. Hayet JB, Piater J (2007) On-line rectification of sport sequences with moving cameras. In: Mexican international conference on artificial intelligence, pp. 736–746. Springer

  15. Hayet JB, Piater J, Verly J (2004) Robust incremental rectification of sports video sequences. In: British machine vision conference (BMVC’04), pp. 687–696. Citeseer

  16. Hess R, Fern A (2007) Improved video registration using non-distinctive local image features. In: 2007 IEEE Conference on computer vision and pattern recognition, pp. 1–8. IEEE

  17. Hobbs J, Power P, Sha L, Lucey P (2018) Quantifying the value of transitions in soccer via spatiotemporal trajectory clustering. In: MIT Sloan sports analytics conference

  18. Homayounfar N, Fidler S, Urtasun R (2017) Sports field localization via deep structured models. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5212–5220

  19. Intel (2019) Freed. https://www.intel.com/content/www/us/en/sports/technology/true-view.html. Accessed: 2019-10-3

  20. Jiang W, Higuera JCG, Angles B, Sun W, Javan M, Yi KM (2020) Optimizing through learned errors for accurate sports field registration. In: The IEEE winter conference on applications of computer vision, pp. 201–210

  21. Kendall A, Grimes M, Cipolla R (2015) Posenet: a convolutional network for real-time 6-dof camera relocalization. In: Proceedings of the IEEE international conference on computer vision, pp. 2938–2946

  22. Kim H, Hong KS (2000) Soccer video mosaicing using self-calibration and line tracking. In: Proceedings 15th International Conference on Pattern Recognition. ICPR-2000, vol. 1, pp. 592–595. IEEE

  23. Lewis AS (2003) The mathematics of eigenvalue optimization. Math Program 97(1-2):155–176

    Article  MathSciNet  Google Scholar 

  24. Liu S, Chen J, Chang CH, Ai Y (2017) A new accurate and fast homography computation algorithm for sports and traffic video analysis. IEEE Trans Circuits Syst Video Technol 28(10):2993–3006

    Article  Google Scholar 

  25. Lowe DG (2004) Distinctive image features from scale-invariant keypoints. International J Computer Vision 60(2):91–110

    Article  Google Scholar 

  26. Mahendran S, Ali H, Vidal R (2018) Convolutional networks for object category and 3d pose estimation from 2d images. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops, pp. 0–0

  27. Martinsson PG, Voronin S (2016) A randomized blocked algorithm for efficiently computing rank-revealing factorizations of matrices. SIAM J Sci Comput 38(5):S485–S507

    Article  MathSciNet  Google Scholar 

  28. Oh TH, Matsushita Y, Tai YW, So Kweon I (2015) Fast randomized singular value thresholding for nuclear norm minimization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4484–4493

  29. Okuma K, Little JJ, Lowe DG (2004) Automatic rectification of long image sequences. In: Asian conference on computer vision, vol. 9

  30. Prozone (2019) Prozone. http://prozonesports.stats.com/. Accessed: 2019-10-6

  31. Redmon J, Farhadi A (2018) Yolov3: An incremental improvement arXiv

  32. Sanyal S (2021) Who will receive the ball? predicting pass recipient in soccer videos. Journal of Visual Communication and Image Representation p 103190

  33. Sanyal S, Kundu A, Mukherjee DP (2016) On the (soccer) ball. In: Proceedings of the tenth indian conference on computer vision, graphics and image processing, pp. 1–8

  34. Sha L, Hobbs J, Felsen P, Wei X, Lucey P, Ganguly S (2020) End-to-end camera calibration for broadcast videos. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13627–13636

  35. Sha L, Lucey P, Yue Y, Wei X, Hobbs J, Rohlf C, Sridharan S (2018) Interactive sports analytics: an intelligent interface for utilizing trajectories for interactive sports play retrieval and analytics. ACM Transactions on Computer-Human Interaction (TOCHI) 25(2):1–32

    Article  Google Scholar 

  36. Sharma RA, Bhat B, Gandhi V, Jawahar CV (2018) Automated top view registration of broadcast football videos. In: IEEE Winter conference on applications of computer vision (WACV), pp. 305–313

  37. Tekin B, Sinha SN, Fua P (2018) Real-time seamless single shot 6d object pose prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 292–301

  38. Torr PH, Zisserman A (2000) Mlesac: a new robust estimator with application to estimating image geometry. Computer vision and image understanding 78(1):138–156

    Article  Google Scholar 

  39. Valada A, Radwan N, Burgard W (2018) Deep auxiliary learning for visual localization and odometry. In: 2018 IEEE International conference on robotics and automation (ICRA), pp. 6939–6946. IEEE

  40. Wang F, Sun L, Yang B, Yang S (2006) Fast arc detection algorithm for play field registration in soccer video mining. In: 2006 IEEE International conference on systems, man and cybernetics, vol. 6, pp. 4932–4936. IEEE

  41. Watanabe T, Haseyama M, Kitajima H (2004) A soccer field tracking method with wire frame model from tv images. In: 2004 International conference on image processing, 2004. ICIP’04., vol. 3, pp. 1633–1636. IEEE

  42. Watson GA (1992) Characterization of the subdifferential of some matrix norms. Linear algebra and its applications 170:33–45

    Article  MathSciNet  Google Scholar 

  43. Wen PC, Cheng WC, Wang YS, Chu HK, Tang NC, Liao HYM (2015) Court reconstruction for camera calibration in broadcast basketball videos. IEEE transactions on visualization and computer graphics 22(5):1517–1526

    Article  Google Scholar 

  44. Xiang Y, Schmidt T, Narayanan V, Fox D (2017) Posecnn:, A convolutional neural network for 6d object pose estimation in cluttered scenes. arXiv:1711.00199

  45. Yamada A, Shirai Y, Miura J (2002) Tracking players and a ball in video image sequence and estimating camera parameters for 3d interpretation of soccer games. In: Object recognition supported by user interaction for service robots, vol. 1, pp. 303–306. IEEE

  46. Zhan E, Zheng S, Yue Y, Sha L, Lucey P (2018) Generating multi-agent trajectories using programmatic weak supervision. arXiv:1803.07612

Download references

Funding

Author receives no funding from any agency for the research.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Samriddha Sanyal.

Ethics declarations

Conflict of Interests

The authors declare no competing interests.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix 1

Let F be an arbitrary approximation of F directly derived by a model from (b) into (d) in Fig. 2. Then for any F, there exists a H and T such that (HT) is a better approximation of F.

Proof

Let we have the groundtruth F that registers the players of the video frame into the top view model. Given a video frame, the problem is to approximate F. Let, the state-of-the-art method approximates F as F defined from (b) into (d). F registers players from the video frame into the top-view model with approximation error 𝜖 i.e. ∥FF∥ = 𝜖, where ∥.∥ is the Frobenius matrix norm. Given arbitrary F, we aim to construct T and H such that (HT) is a better approximation of F. More precisely, given arbitrary F, there exists T and H such that (HT) has approximation error less than 𝜖. i.e. ∥F − (HT)∥ < 𝜖. The proof is comprised of three steps. First, we construct H and T for a given F which depends on the value of a variable s. Then we derive the error bound for the proposed approximation (HT) involving s. Finally, the value of s is computed for the given 𝜖. Using the value of s, H and T can be computed specifically with approximation error less than 𝜖. Next we expound the details.

Construction of H and T The task of computing a factor form approximation of F can be divided into two steps. The first is to construct a subspace that captures the range of F. The second is to restrict the F to the subspace and compute a standard factorization of the reduced F with the help of H. Next, we discuss how to accomplish the proposed steps.

The first step can be executed with random sampling methods [27]. To understand how randomness works, let us consider F = B + E, where B captures the range of F and E is a small perturbation error during groundtruth generation process. Our aim is to obtain a basis of exact rank r that covers as much of the range of B as possible.

Let us consider the dimension of F, B and E are (a × b). In order to obtain r rank approximation of B, we fix a small number s. Then (r + s) random vectors {αi} can be generated such that:

$$ F(\alpha_{i})= B(\alpha_{i}) + E(\alpha_{i}), $$
(6)

for i = 1,..., (r + s). The perturbation E deviates the direction of each {αi} outside the range of B. Therefore, the extra s vectors enhance the chance of spanning the required subspace. Overall the general randomized algorithm to derive the H is comprised of three steps [11] as follows.

First, a random (b × (r + s)) matrix \({\mathscr{G}}\) is generated whose columns are Gaussian vectors. Thereafter compute \((F{\mathscr{G}})\). Finally, construct a matrix H whose columns form an orthonormal basis of the range \((F{\mathscr{G}})\). Once we get the H, then we can compute the other factor (HF). i.e. FH(HF). Considering T = HF, we approximate F in factor form of (HT). Next, we compute the bound of approximation error of ∥FHT∥. There after, we determine the value of s that is needed to compute H and T so that the approximation error is less than 𝜖.

Computing the error bound We aim to show:

$$ E(\|{F-H(H^{*}F)}\|)) \leq \left( 1+ \frac{r}{s-1}\right) \left( {\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}}\right) $$

where E is the expectation, \({\sum }_{i=r+1}^{\min \limits {(a,b)}} \sigma _{i}^2\) is the theoretically minimal error in approximating F by a matrix of rank r [11].

First, consider the singular value decomposition of F as \(F= U_{1} {\Sigma }_{1} V_{1}^{*}\), where U1 is a (a × r) orthonormal matrix, Σ1 is a diagonal matrix containing the non negative singular values of F and V1 is a (r × n) orthonormal matrix. We call U1 and V1 as left unitary factor and right unitary factor respectively. First partition the Σ1 = [Σ23], where the Σ2 and Σ3 are the diagonal matrix containing the first r and (br) singular values respectively. Thereafter, partition V1 = [V2|V3] into blocks containing r and br columns respectively. Define \({\mathscr{G}}_{2} = V^{*}_{2}{\mathscr{G}}\) and \({\mathscr{G}}_{3} = V^{*}_{3}{\mathscr{G}}\). Since, V2 and V3 are orthonormal, then \({\mathscr{G}}_{2}\) and \({\mathscr{G}}_{3}\) are also Gaussian. We denote the pseudoinverse of \({\mathscr{G}}_{2}\) and \({\mathscr{G}}_{3}\) as \(\hat {{\mathscr{G}}_{2}}\) and \(\hat {{\mathscr{G}}_{3}}\) respectively. \({\mathscr{G}}_{2}\) and \({\mathscr{G}}_{3}\) are non overlapping, so they are stochastically independent. Applying Holder’s inequality, we can write:

$$ E(\|{F-H(H^{*}F)}\|))\leq (E(\|{F-H(H^{*}F)}\|^{2}))^{1/2} $$
(7)

It is proved in [11] that:

$$ E(\|{F-H(H^{*}F)}\|^{2}) \leq (\|{{\Sigma}_{3}}\|^{2}_{F} + E(\|{{\Sigma}_{3}\mathscr{G}_{3} \hat{\mathscr{G}_{2}})}\|^{2}) $$
(8)

Therefore, using (7) and (8), we can write:

$$ E(\|{F-H(H^{*}F)}\|))\leq (\|{{\Sigma}_{3}}\|^{2}_{F} + E(\|{{\Sigma}_{3}\mathscr{G}_{3} \hat{\mathscr{G}_{2}})}\|^{2})^{1/2} $$
(9)

We are interested in the r ranks of the matrix. Therefore we compute \(E(\|{{\Sigma }_{3}{\mathscr{G}}_{3} \hat {{\mathscr{G}}_{2}}}\|^{2})\) by conditioning on the value of \({\mathscr{G}}_{2}\) as follows:

$$ E(\|{{\Sigma}_{3}\mathscr{G}_{3} \hat{\mathscr{G}_{2}}}\|^{2}) = E(E(\|{{\Sigma}_{3}\mathscr{G}_{3} \hat{\mathscr{G}_{2}}}\|^{2})|\mathscr{G}_{2}) $$
(10)

The Frobenious norm is unitarily invariant. i.e. for any two orthonormal matrices U1 and V1, we can write ∥U1Σ1V1∥ = ∥Σ1∥. In addition, the distribution of a Gaussian matrix is invariant under orthogonal transformations. Therefore, we can write:

$$ \begin{array}{@{}rcl@{}} E(E(\|{{\Sigma}_{3}\mathscr{G}_{3} \hat{\mathscr{G}_{2}}}\|^{2})|\mathscr{G}_{2}) & =& E(E({\Sigma}_{jk}(\sigma_{jj} [\mathscr{G}_{3}]_{jk} [\hat{\mathscr{G}_{2}}]_{kk})) \\ & =& E({\Sigma}_{jk}(\sigma^{2}_{jj} [\hat{\mathscr{G}_{2}}]^{2}_{kk})) \\ & =& E (\|{{\Sigma}_{3}}\|_{F}^{2} \|{\hat{\mathscr{G}_{2}}}\|^{2}) \\ & =& \|{{\Sigma}_{3}}\|_{F}^{2} E(\|{\hat{\mathscr{G}_{2}}}\|^{2}) \\ & =& \frac{r}{s-1} \|{{\Sigma}_{3}}\|_{F}^{2} \\ & =& \frac{r}{s-1}\left( \sum\limits_{i=r+1}^{\min{(a,b)}}{\sigma_{i}^{2}}\right) \\ \end{array} $$
(11)

Therefore, putting the expression of \(E(\|{{\Sigma }_{3}{\mathscr{G}}_{3} \hat {{\mathscr{G}}_{2}}}\|^{2})\) in the (9), we can write:

$$ E(\|{F-H(H^{*}F)}\|)) \leq \left( 1+ \frac{r}{s-1}\right) \left( \sum\limits_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}}\right) $$
(12)

In (12), \({\sum }_{i=r+1}^{\min \limits {(a,b)}} \sigma _{i}^2\) is the theoretically minimal error in approximating F by rank r [11]. Therefore, the optimal bound is missed by a factor of \((1+ \frac {r}{s-1})\). Now, our objective is to determine the value of s for a given 𝜖.

Computing the value of s Our proposed error for factorized approximation is less than 𝜖. The target rank r is strictly greater than 1. Therefore from (12), we can write:

$$ \begin{array}{@{}rcl@{}} \left( 1+ \frac{r}{s-1}\right) \left( \sum\limits_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}}\right) & <& \epsilon \\ \left( 1+ \frac{r}{s-1}\right) & <& \frac{\epsilon}{({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})} \\ \frac{r}{s-1} & <& \frac{\epsilon}{({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})} -1 \\ \frac{r}{s-1} & <& \frac{\epsilon-({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})}{({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})} \\ \frac{r({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})}{\epsilon-({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})} & <& s-1 \\ \frac{r({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})}{\epsilon-({\sum}_{i=r+1}^{\min{(a,b)}} {\sigma_{i}^{2}})}+1 & <& s \\ \end{array} $$
(13)

Therefore, we can choose \(s= \Bigg \lceil {\frac {r({\sum }_{i=r+1}^{\min \limits {(a,b)}} \sigma _{i}^2)}{\epsilon -({\sum }_{i=r+1}^{\min \limits {(a,b)}} \sigma _{i}^2)}+1}\Bigg \rceil \), where ⌈.⌉ function gives the least integer greater than or equal to the given input.

The upshot Putting the value of s we can compute the H and T = HF such that HT is an approximation of F with approximation error less than 𝜖. Thus we prove that for a given F and an arbitrary state-of-the-art approximation of F say F, we can always find an approximation in factor form of (HT) which is better than F. □

Appendix 2

Lemma 1: Let \(X^{*}_{\kappa }\) be the solution to (3) and \(X_{\infty }\) be the minimum Frobenius norm solution to (1) defined as \(X_{\infty } := \{\underset {X}{ argmin}\|{X}\|^2_F: X \text {is a solution of}\) (1)}. Then \(\lim _{\kappa \to \infty } \|{X^{*}_{\kappa }-X_{\infty }}\|_F=0\).

Proof

From the definition of \(X^{*}_{\kappa }\), we can write

$$ \|{X^{*}_{\kappa}}\|_{*} + \|{X^{*}_{\kappa}}\|^{2}_{F} \leq \|{X_{\infty}}\|_{*} + \|{X_{\infty}}\|^{2}_{F}. $$
(14)

From the definition of \(X_{\infty }\) we can write:

$$ \|{X_{\infty}}\|_{*} \leq \|{X^{*}_{\kappa}}\|_{*}. $$
(15)

From the 14 and 15, we can write:

$$ \|{X^{*}_{\kappa}}\|^{2}_{F} \leq \|{X_{\infty}}\|^{2} $$
(16)

The (16) implies that \(X^{*}_{\kappa }\) is uniformly bounded. Now, the theorem is proved if we can show that any convergent subsequence \(\{X^{*}_{\kappa _{i}}\}_{i \geq 1}\) must converge to \(X_{\infty }\).

Consider an arbitrary converging subsequence \(\{X^{*}_{\kappa _{i}}\}\) and set \(X_c= \lim _{i \to \infty } X^{*}_{\kappa _{i}}\). Since \(X^{*}_{\kappa }\) is uniformly bounded, we can write:

$$ \lim_{\kappa \to \infty} sup \|{X^{*}_{\kappa}}\| \leq \|{X_{\infty}}\|_{*} $$
(17)

and

$$ \|{X^{*}_{\kappa}}\|_{*} \leq \lim_{\kappa \to \infty} inf \|{X_{\infty}}\|_{*} $$
(18)

From Eq. 17 and 18, we can write \(\lim _{\kappa \to \infty } \|{X^{*}_{\kappa }}\| = \|{X_{\infty }}\|_{*}\), therefore, \(\|{X_{c}}\|_{*}= \|{X_{\infty }}\|_{*}\). This shows that Xc is a solution of \(\underset {X}{ argmin}\|{X}\|_{*}\). Now it follows from the definition of \(X_{\infty }\) that \(\|{X_c}\|_F \geq \|{X_{\infty }}\|_{F}\), while we also have \(\|{X_c}\|_F \leq \|{X_{\infty }}\|_F\) because of (18). Therefore, we conclude that \(\|{X_c}\|_F = \|{X_{\infty }}\|_{F}\) and thus \(X_{c} = X_{\infty }\) since \(X_{\infty }\) is unique. □

Lemma 2: For each κ > 0 and an (n1 × n2) image Y , algo. 2 converges to: \(\underset {X}{ argmin} \|{X-Y}\|^2_F+\kappa \|{X}\|_{*}\).

Proof

Essentially, we have to prove that:

\(O_{\kappa }(Y)=\underset {X}{ argmin} \|{X-Y}\|^2_F+\kappa \|{X}\|_{*}\). Let us assume \(M(X)= \underset {X}{ argmin} \|{X-Y}\|^2_F+\kappa \|{X}\|_{*}\). M is strictly convex. So there exists a unique minimizer of M. Therefore, we need to prove that the minimizer is equal to Oκ(Y ). To do this, the definition of a subgradient of a convex function is as follows [2]: Z is a subgradient of M at X0, denoted as ZM(X0), if

$$ M(X) \geq M(X_{0}) + \langle Z, (X - X_{0}) \rangle $$
(19)

for all X. Now Oκ(Y ) minimizes M if and only if 0 is a subgradient of the M at the point Oκ(Y ). i.e.

$$ 0 \in O_{\kappa}(Y)-Y + \kappa \partial \|{O_{\kappa}(Y)}\|_{*}, $$
(20)

where Oκ(Y )∥ is the set of subgradients of the nuclear norm. Let X be an arbitrary image and SVD of X = UΣVT. Then we can write [23] [42]:

\(\partial \|{X}\|_{*} = \{UV^{T} + W : W \in \mathbb {R}^{n_1 \times n_2}, U^{T}W = 0, WV = 0, \|{W}\|_{F} \leq 1\}\).

In order to show that Oκ(Y ) obeys (20), decompose the SVD of Y as: \(Y = U_{3} {\Sigma }_{3} {V_{3}^{T}} + U_{4} {\Sigma }_{4} {V_{4}^{T}}\), where U3, V3 (resp. U4, V4) are the singular vectors associated with singular values greater than κ (resp. smaller than or equal to κ ). With these notations, we have \(O_{\kappa }(Y) = U_{3} ({\Sigma }_{3} - \kappa I) {V_{3}^{T}}\).

Therefore, \(Y - O_{\kappa }(Y) = \kappa (U_{3} {V_{3}^{T}} + W)\), and thus \(W = \kappa ^{-1} U_{4} {\Sigma }_{1} {V_{4}^{T}}\).

By definition, \({U^{T}_{0}}W = 0\), WV3 = 0 and since the diagonal elements of \({\sum }_{1}\) have magnitudes bounded by κ, we also have ∥W2 ≤ 1. Hence YOκ(Y ) ∈ κOκ(Y )∥, which concludes the proof. □

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Sanyal, S. TVVS: A top-view visualization system from broadcasting soccer video. Multimed Tools Appl 81, 33613–33644 (2022). https://doi.org/10.1007/s11042-022-12605-9

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-022-12605-9

Keywords

Navigation