Abstract
Incorporating prior knowledge into a segmentation process, whether it is geometrical constraints such as volume penalisation, (partial) convexity enforcement, or topological prescriptions to preserve the contextual relations between objects, proves to improve accuracy in medical image segmentation, in particular when addressing the issue of weak boundary definition. Motivated by this observation, the proposed contribution aims to provide a unified variational framework including geometrical constraints in the training of convolutional neural networks in the form of a penalty in the loss function. These geometrical constraints take several forms and encompass level curve alignment through the integration of the weighted total variation, an area penalisation phrased as a hard constraint in the modelling, and an intensity homogeneity criterion based on a combination of the standard Dice loss with the piecewise constant Mumford–Shah model. The mathematical formulation yields a non-smooth non-convex optimisation problem, which rules out conventional smooth optimisation techniques and leads us to adopt a Lagrangian setting. The application falls within the scope of organ-at-risk segmentation in CT (computed tomography) images, in the context of radiotherapy planning. Experiments demonstrate that our method provides significant improvements (i) over existing non-constrained approaches both in terms of quantitative criteria, such as the measure of overlap, and qualitative assessment (spatial regularisation/coherency, fewer outliers), (ii) over in-layer constrained deep convolutional networks, and shows a certain degree of versatility.
This is a preview of subscription content, access via your institution.





Notes
The problem being separable with respect to the variable k, we omit the dependency in k from now on.
References
Alexandrov, O., Santosa, F.: A topology-preserving level set method for shape optimization. J. Comput. Phys. 204(1), 121–130 (2005)
Azé, D.: Éléments d’analyse convexe et variationnelle. Mathématiques pour le 2ème cycle. Ellipses (1997)
Baldi, A.: Weighted BV functions. Houston J. Math. 27(3), 683–705 (2001)
Bohlender, S., Oksuz, I., Mukhopadhyay, A.: A Survey on Shape-Constraint Deep Learning for Medical Image Segmentation (2021)
Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 3, 1–122 (2011)
Bresson, X., Esedoḡlu, S., Vandergheynst, P., Thiran, J.P., Osher, S.: Fast global minimization of the active contour/snake model. J. Math. Imaging Vis. 28(2), 151–167 (2007)
Caselles, V., Kimmel, R., Sapiro, G.: Geodesic active contours. Int. J. Comput. Vis. 22(1), 61–87 (1993)
Chambolle, A.: An algorithm for total variation minimization and applications. J. Math. Imaging Vis. 20(1), 89–97 (2004)
Chambolle, A., Pock, T.: A first-order primal–dual algorithm for convex problems with applications to imaging. J. Math. Imaging Vis. 40(1), 120–145 (2011)
Chambolle, A., Pock, T.: On the ergodic convergence rates of a first-order primal–dual algorithm. Math. Program. 159(1), 253–287 (2016)
Chambolle, A., Tan, P., Vaiter, S.: Accelerated alternating descent methods for Dykstra-like problems. J. Math. Imaging Vis. 3(59), 481–497 (2017)
Chen, X., Williams, B.M., Vallabhaneni, S.R., Czanner, G., Williams, R., Zheng, Y.: Learning active contour models for medical image segmentation. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11624–11632 (2019)
Clough, J., Byrne, N., Oksuz, I., Zimmer, V.A., Schna-bel, J.A., King, A.: A topological loss function for deep-learning based image segmentation using persistent homology. IEEE Trans. Pattern Anal. Mach. Intell. 6, 66 (2020)
Combettes, P.L., Pesquet, J.C.: Proximal Splitting Methods in Signal Processing, pp. 185–212. Springer, New York (2011)
Dolz, J., Ayed, I.B., Desrosiers, C.: Unbiased shape compactness for segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 755–763. Springer (2017)
Ekeland, I., Témam, R.: Convex Analysis and Variational Problems. Society for Industrial and Applied Mathematics (1999)
El Jurdi, R., Petitjean, C., Honeine, P., Cheplygina, V., Abdallah, F.: High-level prior-based loss functions for medical image segmentation: a survey. arXiv preprint arXiv:2011.08018 (2020)
Fu, H., Xu, Y., Lin, S., Wong, D.W.K., Liu, J.: Deepvessel: retinal vessel segmentation via deep learning and conditional random field. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 132–139. Springer (2016)
Gabay, D., Mercier, B.: A dual algorithm for the solution of nonlinear variational problems via finite element approximation. Comput. Math. Appl. 2(1), 17–40 (1976)
Ganaye, P.A., Sdika, M., Triggs, B., Benoit-Cattin, H.: Removing segmentation inconsistencies with semi-supervised non-adjacency constraint. Med. Image Anal. 58, 101551 (2019)
Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 249–256. JMLR Workshop and Conference Proceedings (2010)
Jia, F., Liu, J., Tai, X.C.: A regularized convolutional neural network for semantic image segmentation. Anal. Appl. 19(01), 147–165 (2021)
Kamnitsas, K., Ledig, C., Newcombe, V.F.J., Simpson, J.P., Kane, A.D., Menon, D.K., Rueckert, D., Glocker, B.: Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 36, 61–78 (2017)
Kass, M., Witkin, A.P., Terzopoulos, D.: Snakes: active contour models. J. Comput. Vis. 1(4), 321–331 (1988)
Kervadec, H., Dolz, J., Tang, M., Granger, E., Boykov, Y., Ayed, I.B.: Constrained-CNN losses for weakly supervised segmentation. Med. Image Anal. 54, 88–99 (2019)
Kim, B., Ye, J.C.: Mumford–Shah loss functional for image segmentation with deep learning. IEEE Trans. Image Process. 29, 1856–1866 (2020)
Lafferty, J.D., McCallum, A., Pereira, F.C.N.: Conditional random fields: probabilistic models for segmenting and labeling sequence data. In: Proceedings of the Eighteenth International Conference on Machine Learning, ICML’01, pp. 282–289. Morgan Kaufmann, San Francisco, CA, USA (2001)
Lambert, Z., Le Guyader, C., Petitjean, C.: A geometrically-constrained deep network for CT image segmentation. In: IEEE International Symposium on Biomedical Imaging (ISBI) (2021)
Lambert, Z., Petitjean, C., Dubray, B., Ruan, S.: SegTHOR: segmentation of thoracic organs at risk in CT images. In: 2020 Tenth International Conference on Image Processing Theory, Tools and Applications (IPTA), pp. 1–6 (2020)
Le Guyader, C., Vese, L.A.: Self-repelling snakes for topology-preserving segmentation models. IEEE Trans. Image Process. 17(5), 767–779 (2008)
Liu, J., Wang, X., Tai, X.C.: Deep Convolutional Neural Networks with Spatial Regularization, Volume and Star-Shape Priori for Image Segmentation (2020)
Milletari, F., Navab, N., Ahmadi, S.A.: V-net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV), pp. 565–571. IEEE (2016)
Moisan, L.: How to discretize the total variation of an image? PAMM 7(1), 1041907–1041908 (2007)
Moreau, J.J.: Fonctions convexes duales et points proximaux dans un espace hilbertien. Comptes rendus hebdomadaires des séances de l’Académie des sciences 255, 2897–2899 (1962)
Mumford, D., Shah, J.: Optimal approximation by piecewise smooth functions and associated variational problems. Comm. Pure Appl. Math. 42(5), 577–685 (1989)
Nosrati, M.S., Hamarneh, G.: Incorporating Prior Knowledge in Medical Image Segmentation: a Survey. CoRR arXiv:1607.01092 (2016)
Peng, J., Kervadec, H., Dolz, J., Ben Ayed, I., Pedersoli, M., Desrosiers, C.: Discretely-constrained deep network for weakly supervised segmentation. Neural Netw. 130, 297–308 (2020)
Rudin, L., Osher, S., Fatemi, E.: Nonlinear total variation based noise removal algorithms. Phys. D 60(1–4), 259–268 (1992)
Rupprecht, C., Huaroc, E., Baust, M., Navab, N.: Deep active contours. CoRR arXiv:1607.05074 (2016)
Ségonne, F.: Active contours under topology control-genus preserving level sets. Int. J. Comput. Vis. 79(2), 107–117 (2008)
Siu, C.Y., Chan, H.L., Lui, L.M.: Image segmentation with partial convexity shape prior using discrete conformality structures. SIAM J. Imaging Sci. 13(4), 2105–2139 (2020)
Taylor, G., Burmeister, R., Xu, Z., Singh, B., Patel, A., Goldstein, T.: Training neural networks without gradients: a scalable ADMM approach. In: Balcan, M.F., Weinberger, K.Q. (Eds.) Proceedings of the 33rd International Conference on Machine Learning, Proceedings of Machine Learning Research, vol. 48, pp. 2722–2731. PMLR (2016)
Zheng, S., Jayasumana, S., Romera-Paredes, B., Vineet, V., Su, Z., Du, D., Huang, C., Torr, P.: Conditional random fields as recurrent neural networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1529–1537 (2015)
Acknowledgements
This project was co-financed by the European Union with the European regional development fund (ERDF, 18P03390/18E01750/18P02733) and by the Hau-te-Normandie Régional Council via the M2SINUM project. The authors would like to thank the CRIANN (Centre Régio-nal Informatique et d’Applications Numériques de Normandie, France) for providing computational resources.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Proof of Theorem 2
The sets \({\mathcal {C}}_1\) and \({\mathcal {C}}_2\) are closed and convex, while the objective function is continuous and coercive. Indeed, \(\forall l \in \left\{ 1,\cdots ,L\right\} \),
Using Young’s inequality with \(\epsilon \) (valid for \(\epsilon >0\)) and stated by \(ab\le \dfrac{a^2}{2\epsilon }+\dfrac{\epsilon b^2}{2}\), we get:
or equivalently,
(To lighten the notations, when there is no ambiguity about the dimension of the mathematical objects we handle, we omit the lower index making this dimension explicit in the definition of the Euclidean norm as well as on the related scalar product.)
Taking \(\epsilon \) such that \(\dfrac{2}{\mu +2}<\epsilon <1\) yields the desired result.
To conclude, functional \({\mathcal {J}}\) is strictly convex, due to the strict convexity of functional \({\mathcal {H}}\) defined by \({\mathcal {H}}(u,v)=\dfrac{\mu }{2}\,\Vert u-s(\theta )-w\Vert ^2+\dfrac{1}{2}\,\Vert u-v\Vert ^2\), u denoting the concatenation of the \(u^l\)’s, similarly for \(s(\theta ), w\) and v. A straightforward computation gives that \(\forall (u_1,v_1)\in {\mathcal {C}}_1 \times {\mathcal {C}}_2\), \(\forall (u_2,v_2)\in {\mathcal {C}}_1 \times {\mathcal {C}}_2\),
this quantity vanishing if and only if \(u_1=u_2\) and \(v_1=v_2\).
Proof of Theorem 3
For every \(p\in {\mathcal {B}}, (u,v) \mapsto {\mathcal {L}}(u,v,p)\) is strictly convex, owing to the strict convexity of functional \({\mathcal {H}}\). For each \(p\in {\mathcal {B}}\), functional \((u,v) \mapsto {\mathcal {L}}(u,v,p)\) is continuous and coercive. To establish such a coercivity inequality, denoting by \(\kappa =\Vert {\text{ div }}\Vert =\displaystyle {\sup _{\Vert p\Vert _{Y^L} \le 1}}\,\Vert {\text{ div }}\,p\Vert _{X^L}\), we first observe that with the convention \((p^1)^1_{0,j}=(p^1)^1_{N,j}=(p^1)^2_{i,0}=(p^1)^2_{i,N}=0\) (similarly for \(p^l\) with \(l\in \left\{ 2,\cdots ,L\right\} \)) and by applying twice the inequality \((a+b)^2\le 2(a^2+b^2)\),
Thus \(\kappa \le 2\sqrt{2}\) and with suitable \(\epsilon \),
since \(p\in {\mathcal {B}}\), entailing that \(\Vert p\Vert _{Y^L}^2\le LN^2\).
Applying again Young’s inequality with \(\epsilon '>0\),
This latter inequality shows that by choosing \(\epsilon '\) suitable— which is always possible, the coercivity property is ensured. Also, the quantity \({\mathcal {L}}(u,v,p)\) is bounded below (independently of p) and one remarks that if one takes \({\tilde{u}}\) such that \({\tilde{u}}^1 \equiv 1\) and \(\forall l\in \left\{ 2,\cdots ,L\right\} , {\tilde{u}}^l\equiv 0\), and \({\tilde{v}}\) such that \(\forall l \in \left\{ 1,\cdots ,L\right\} \), \(\forall (i,j)\in {\mathcal {G}}\), \({\tilde{v}}_{i,j}^l=\dfrac{\alpha ^l}{N^2}\), \({\mathcal {L}}({\tilde{u}},{\tilde{v}},p)\) is independent of p, showing that the infimum is finite.
Then for each \(p\in {\mathcal {B}}\), functional \({\mathcal {L}}(\cdot ,\cdot ,p)\) is continuous, coercive and strictly convex, so it admits a unique minimiser in \({\mathcal {C}}_1 \times {\mathcal {C}}_2\)—\({\mathcal {C}}_1\) and \({\mathcal {C}}_2\) being closed convex sets—denoted by \((e_1(p),e_2(p))\). We denote this minimum by f(p), i.e.
Function \(p \mapsto f(p)\) is concave as the pointwise infimum of concave functions (\(\forall (u,v) \in {\mathcal {C}}_1 \times {\mathcal {C}}_2, p \mapsto {\mathcal {L}}(u,v,p)\) is concave since in fact linear). This can be proved using the hypograph of f. Also, f is upper semicontinuous as the pointwise infimum of continuous functions. It is therefore bounded above and attains its upper bound as the set \({\mathcal {B}}\) is compact at a point denoted by \({\bar{p}}\). Thus
Additionally, as \(f({\bar{p}})=\displaystyle {\min _{(u,v)\in {\mathcal {C}}_1 \times {\mathcal {C}}_2}}\,{\mathcal {L}}(u,v,{\bar{p}})\), one has, \(\forall (u,v)\in {\mathcal {C}}_1 \times {\mathcal {C}}_2\),
By concavity of \({\mathcal {L}}\) with respect to the third argument (in fact, linearity), \(\forall (u,v)\in {\mathcal {C}}_1 \times {\mathcal {C}}_2, \forall p \in {\mathcal {B}}, \forall \lambda \in ]0,1[\),
Taking as particular value \((u,v)=(e_1((1-\lambda )\,{\bar{p}}+\lambda p),e_2((1-\lambda )\,{\bar{p}}+\lambda p))=(e_{\lambda }^1,e_{\lambda }^2)\), it yields, using again that \(f({\bar{p}})=\displaystyle {\max _{p\in {\mathcal {B}}}}\,f(p)\) and the concavity (even linearity) of \({\mathcal {L}}\) with respect to the third argument,
As \({\mathcal {L}}(e_{\lambda }^1,e_{\lambda }^2,{\bar{p}}) \ge f({\bar{p}})=\displaystyle {\min _{(u,v)\in {\mathcal {C}}_1 \times {\mathcal {C}}_2}}\,{\mathcal {L}}(u,v,{\bar{p}})\), the latter inequality implies that \(\forall p \in {\mathcal {B}}\),
By virtue of the coercivity property established previously, one has, \(\forall p \in {\mathcal {B}}\) (parameters \(\epsilon \) and \(\epsilon '\) being suitably chosen),
with \({\tilde{u}}\in {\mathcal {C}}_1\) and \({\tilde{v}} \in {\mathcal {C}}_2\) as defined before, making the right-hand side independent of \(p, {\bar{p}}\) and \(\lambda \)—and constituting thus a uniform bound, and showing that \(e_{\lambda }^1\) is uniformly bounded (this was already known owing to the definition of \({\mathcal {C}}_1\)) as well as \(e_{\lambda }^2\). One can thus extract a subsequence (common extracting mapping) \(e_{\lambda _n}^1\) and \(e_{\lambda _n}^2\) with \(\lambda _n \underset{n \rightarrow +\infty }{\rightarrow } 0\) converging to some limits \({\bar{u}}\) and \({\bar{v}}\). We show next that \({\bar{u}}=e_1({\bar{p}})\) and \({\bar{v}}=e_2({\bar{p}})\).
As by definition, \({\mathcal {L}}(e_{\lambda }^1,e_{\lambda }^2,(1-\lambda )\,{\bar{p}}+\lambda p)= \min _{(u,v)\in \displaystyle {{\mathcal {C}}_1\times {\mathcal {C}}_2}}\,{\mathcal {L}}(u,v,(1-\lambda )\,{\bar{p}}+\lambda p), \forall (u,v)\in {\mathcal {C}}_1\times {\mathcal {C}}_2\),
and by linearity of \({\mathcal {L}}\) with respect to the third argument, it follows that \(\forall (u,v)\in {\mathcal {C}}_1\times {\mathcal {C}}_2\),
The quantity \({\mathcal {L}}(e_{\lambda }^1,e_{\lambda }^2,p)\) is bounded below by f(p) so that passing to the limit in the previous inequality when \(\lambda _n\) tends to 0 yields, using the continuity of \({\mathcal {L}}\),
this being true for all \((u,v)\in {\mathcal {C}}_1\times {\mathcal {C}}_2\). By uniqueness of the minimiser of \(\displaystyle {\min _{(u,v)\in {\mathcal {C}}_1\times {\mathcal {C}}_2}}\,{\mathcal {L}}(u,v,{\bar{p}})\), we deduce that \(({\bar{u}},{\bar{v}})=(e_1({\bar{p}}),e_2({\bar{p}}))\).
At last, passing to the limit in (20) yields \({\mathcal {L}}({\bar{u}},{\bar{v}},p)\le f({\bar{p}}), \forall p \in {\mathcal {B}}\), which combines with (19) and the invocation of [16, Chapter VI, Proposition 1.3] enables one to conclude that \(({\bar{u}},{\bar{v}},{\bar{p}})\) is a saddle point of \({\mathcal {L}}\).
Comment on the Fact that Fixed Points of Algorithm 3 are Saddle Points of the Associated Lagrangian
First of all, one remarks that
The first step of the algorithm can thus be rephrased as
Similarly,
The second step of the algorithm thus reads as
Now, considering a fixed point \((x^{*},y^{*})\) of the algorithm and owing to the fact that
the relation (21) gives that \(\forall x \in {\mathcal {X}}\),
while the relation (22) leads to \(\forall y \in {\mathcal {Y}}\),
By summing both inequalities, it yields
But by convexity of \(f,f(x^{*})+\langle \nabla f(x^{*}),x-x^{*}\rangle \le f(x)\), which implies that
Then \(\forall x \in {\mathcal {X}}, \forall y \in {\mathcal {Y}}\), \({\mathcal {L}}(x^{*},y)\le {\mathcal {L}}(x,y^{*})\), showing that \((x^{*},y^{*})\) is a saddle point of the associated Lagrangian denoted by \({\mathcal {L}}\) here.
Proof of Theorem 4
We recall that the considered saddle-point structure reads as

and that the general iteration of the algorithm is given by
with
The inputs are thus the points \(({\bar{x}},{\bar{y}})\) as well as the intermediate points \(({\tilde{x}},{\tilde{y}})\), while the outputs are the generated points \(({\hat{x}},{\hat{y}})\).
Here \(D_x\) and \(D_y\) are Bregman proximity/distance functions (see [10, p. 256] for further details) chosen to be \(D_x(x,{\bar{x}})=\frac{1}{2}\,\Vert x-{\bar{x}}\Vert ^2\) (respectively, \(D_y(y,{\bar{y}})=\frac{1}{2}\,\Vert y-{\bar{y}}\Vert ^2\)) in our setting, this choice being the most common one. Also, from an algorithmic viewpoint, an iteration is applied with \({\bar{x}}=x^n\), \({\bar{y}}=y^n, {\tilde{x}}=2x^{n+1}-x^n\) and \({\tilde{y}}=y^n\). Lemma 1 from [10, Lemma 1, p. 257 with proof] states that, provided the previous general iteration holds, then for any \(x\in {\mathcal {X}}\) and for any \(y\in {\mathcal {Y}}\), one has
Applying [10, Lemma 1] with \({\hat{x}}=x^{n+1}, {\hat{y}}=y^{n+1}\) and \(D_x, D_y\) as defined above, gives that \(\forall x \in {\mathcal {X}}\) and \(\forall y \in {\mathcal {Y}}\), one has:
Remarking then that \(\left\{ \begin{array}{ccc}x-x^{n+1}&{}=&{}x-x^n+x^n-x^{n+1}\\ x^{n+1}-x^n&{}=&{}x^{n+1}-x+x-x^n \end{array}\right. \), and thus
it yields
Owing to hypothesis (H) according to which \(\left( \frac{1}{\tau }-L_f\right) \frac{1}{\sigma } >\Vert K\Vert ^2\), the quantities in brackets are non-negative. To show this, let us focus on the last quantity in brackets. The same reasoning applies to the two former ones. Using Cauchy–Schwarz inequality combined with Young’s inequality with parameter \(\varepsilon >0\), one has \(\langle K(x^{n+1}-x^n),y^{n+1}-y^n\rangle \le \Vert K\Vert \,\Vert x^{n+1}-x^n\Vert \,\Vert y^{n+1}-y^n\Vert \le \dfrac{\Vert K\Vert }{2\varepsilon }\Vert x^{n+1}-x^n\Vert ^2+\dfrac{\Vert K\Vert \varepsilon }{2}\Vert y^{n+1}-y^n\Vert ^2\), so that setting \(\varepsilon =\frac{1}{{\sqrt{\sigma \,\left( \frac{1}{\tau }-L_f\right) }}}\) leads to:
Hypothesis (H) enables one to conclude that the weights balancing \(\Vert x^{n+1}-x^n\Vert ^2\) and \(\Vert y^{n+1}-y^n\Vert ^2\) are positive, or equivalently that there exists \(\xi >0\) so that
An immediate consequence is that equation (23) reduces to
By taking as particular (x, y) a saddle point \((x^{*},y^{*})\) of the Lagrangian \({\mathcal {L}}\) (whose existence is ensured in our case by Theorem 3), entailing, by definition of a saddle point, that \({\mathcal {L}}(x^{n+1},y^{*})-{\mathcal {L}}(x^{*},y^{n+1})\ge 0\), inequality (25) gives:
By summing from \(n=0\) to \(N-1\), one gets:
and using an estimate as in (24), it follows, still using Cauchy–Schwarz inequality and Young’s inequality, that
showing that the sequence \((x^n,y^n)\) is a bounded sequence. One can thus extract a subsequence \((x^{\Psi (n)},y^{\Psi (n)})\) that (strongly) converges to \((\hat{{\hat{x}}},\hat{{\hat{y}}})\) (since we work in finite dimension). Let us now come back to inequality (23). Proceeding as before with \((x,y)=(x^{*},y^{*})\) and summing the inequalities from \(n=0\) to \(N-1\) coupled with estimation (24) shows that
the latter line containing a negative quantity according again to hypothesis (H). Thus
The sequence \(({\mathcal {S}}(N))_{N\in {\mathbb {N}}^{*}}\) with general term \({\mathcal {S}}(N)=\) \(\displaystyle {\sum _{n=0}^{N-1}}\,\Vert x^{n+1}-x^n\Vert ^2\) is thus increasing, bounded above, so it converges and \(\displaystyle {\lim _{n \rightarrow +\infty }}\,(x^{n+1}-x^n)=0\) (similarly, \(\displaystyle {\lim _{n \rightarrow +\infty }}\,(y^{n+1}-y^n)=0\)), implying that \((x^{\Psi (n)-1},y^{\Psi (n)-1})\) also converges to \((\hat{{\hat{x}}},\hat{{\hat{y}}})\) (take \(n:=\Psi (n)-1\) in the previous result), which is thus a fixed point of Algorithm 3 iteration, hence a saddle point of the Lagrangian \({\mathcal {L}}\) from Appendix Appendix C.
For the last time, we come back to inequality (25) with \((x,y)=(\hat{{\hat{x}}},\hat{{\hat{y}}})\) and sum the inequalities from \(n=\Psi (n)\) to \(N-1\) with \(N>\Psi (n)\). It yields
which proves that \(x^N \rightarrow \hat{{\hat{x}}}\) and \(y^N \rightarrow \hat{{\hat{y}}}\) as N tends to \(+\infty \).
Rights and permissions
About this article
Cite this article
Lambert, Z., Le Guyader, C. & Petitjean, C. Enforcing Geometrical Priors in Deep Networks for Semantic Segmentation Applied to Radiotherapy Planning. J Math Imaging Vis 64, 892–915 (2022). https://doi.org/10.1007/s10851-022-01102-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10851-022-01102-1
Keywords
- Image segmentation
- Deep learning
- Proximal operator
- ADMM algorithm
- Douglas–Rachford algorithm
- Primal–dual algorithm