Chan–Vese Reformulation for Selective Image Segmentation

Selective segmentation involves incorporating user input to partition an image into foreground and background, by discriminating between objects of a similar type. Typically, such methods involve introducing additional constraints to generic segmentation approaches. However, we show that this is often inconsistent with respect to common assumptions about the image. The proposed method introduces a new fitting term that is more useful in practice than the Chan–Vese framework. In particular, the idea is to define a term that allows for the background to consist of multiple regions of inhomogeneity. We provide comparative experimental results to alternative approaches to demonstrate the advantages of the proposed method, broadening the possible application of these methods.


Introduction
Image segmentation is an important application of image processing techniques in which some, or all, objects in an image are isolated from the background.In other words, for an image z(x) ∈ R 2 , we find the partitioning of the image domain Ω ⊂ R 2 into subregions of interest.In the case of two-phase approaches this consists of the foreground domain Ω F and background domain Ω B , such that Ω = Ω F ∪ Ω B .In this work we concentrate on approaching this problem with variational methods, particularly in cases where user input is incorporated.Specifically, we consider the convex relaxation approach of [14,8] and many others.This consists of a binary labelling problem where the aim is to compute a function u(x) ∈ {0, 1} indicating regions belonging to Ω F and Ω B , respectively.This is obtained by imposing a relaxed constraint on the function, u ∈ [0, 1], and minimising a functional that fits the solution to the data with certain conditions on the regularity of the boundary of the foreground regions.
We will first introduce the seminal work of Chan and Vese [15], a segmentation model that uses the level set framework of Osher and Sethian [31].This approach assumes that the image z is approximately piecewise-constant, but is dependent on the initialisation of the level set function as the minimisation problem is nonconvex.The Chan-Vese model was reformulated to avoid this by Chan et al. [14], using convex relaxation methods, that has the following data fitting functional where f 1 (x) and f 2 (x) are data fitting terms indicating the foreground and background regions, respectively.In particular, in [15] and [14] these are given by It should be noted that it is common to fix λ = λ 1 = λ 2 .The introduction of binary labels to image segmentation was also proposed by Lie et al. [26], with the connections between [14] and [26] discussed in Wei et al. [44].The data fitting functional is balanced against a regularisation term.Typically, this penalises the length of the contour.This is represented by the total variation (TV) of the function [15,37], and is sometimes weighted by an edge detection function (i) Image with ground truth (ii) Foreground, c 1 = 0.15 (iii) Background, c 2 = 0.19 33,35,39].Therefore, the regularisation term is given as The convex segmentation problem, assuming fixed constants c 1 and c 2 , is then defined by min In the case where the intensity constants are unknown it is also possible to minimise F CV alternately with respect to u, c 1 , and c 2 , however, this would make the problem nonconvex and hence dependent on the initialisation of u.Functionals of this type have been widely studied with respect to two-phase segmentation [8,14,15], which is our main interest.Alternative choices of data fitting terms can be used when different assumptions are made on the image, z.Examples include [1,2,16,25,40,43].We note that multiphase approaches [9,42] are also closely related to this formulation although in this paper we focus on the two-phase problem due to associated applications of interest.It is also important to acknowledge analogous methods in the discrete setting such as [4,18,22,36].However, we do not go into detail about such methods here, although we introduce the work of [17] in §3 and compare corresponding results in §7.
In selective segmentation the idea is to apply additional constraints such that user input is incorporated to isolate specific objects of interest.It is common for the user to input marker points to form a set M , where M = {(x i , y i ) ∈ Ω , 1 ≤ i ≤ k} and from this we can form a foreground region P whose interior points are inside the object to be segmented.In the case that M is provided P will be a polygon, but any user-defined region in the foreground is consistent with the proposed method.Some examples of selective or interactive methods include [10,17,21,27,30,35,22,38,48,41].A particular application of this in medical imaging is organ contouring in computed tomography (CT) images.This is often done manually which can be laborious and inefficient and it is often not possible to enhance existing methods with training data.In cases where learning based methods are applicable, the work of Xu et al. [46] and Bernard and Gygli [5] are state of the art approaches.At this stage we define the additional constraints in selective segmentation as follows: where D(x) is some distance penalty term, such as [34,35,39], and θ is a selection parameter.Essentially, the idea is that the selection term D(x) (based on the region P formed by the user input marker set) should penalise regions of the background (as defined by the data fitting term f 2 (x)) and also pixels far from P. In this paper we choose D(x) to be the geodesic distance penalty proposed in [35].Explicitly, the geodesic distance from the region P formed from the marker set is given by: D M (x) = 0 for x ∈ P, where D 0 M (x) is the solution of the following PDE: The function q(x) is image dependent and controls the rate of increase in the distance.It is defined as a function similar to where ε D is a small non-zero parameter and β G is a nonnegative tuning parameter.We set the value of β G = 1000 and ε D = 10 −3 throughout.Note that if q(x) ≡ 1 then the distance penalty D M (x) is simply the normalised Euclidean distance, as used in [39].
A general selective segmentation functional, assuming homogeneous target regions, is therefore given by: Assuming that the optimal intensity constants c 1 and c 2 are fixed, the minimisation problem is then: Again, it is possible to alternately minimise F S (u, c 1 , c 2 ) with respect to the constants c 1 and c 2 to obtain the average intensity in Ω F and Ω B , respectively.However, in selective segmentation it is often sufficient to fix these according to the user input.In the framework of (9) the Chan-Vese terms [14,15,29] have limitations due to the dependence on c 2 .In conventional two-phase segmentation problems it makes sense to penalise deviances from c 2 outside the contour, however for selective segmentation we need not consider the intensities outside of the object we have segmented.Regardless of whether the intensity of regions outside the object is above or below c 1 , it should be penalised positively.The Chan-Vese terms cannot ensure this as they work based on a fixed "exterior" intensity c 2 and can lead to negative penalties on regions which are outside the object of interest.It is our aim in this paper to address this problem.The motivation for this work comes from observing contradictions in using piecewise-constant intensity fitting terms in selective segmentation.Whilst good results are possible with this approach, the exceptional cases lead to severe limitations in practice.This is quite common in medical imaging as demonstrated in Fig. 1, where the target foreground has a low intensity.Given that the corresponding background includes large regions of low intensity, the optimal average intensities for this segmentation problem are c 1 = 0.1534 and c 2 = 0.1878.For cases where c 1 ≈ c 2 , we see that by (1), f 1 − f 2 ≈ 0 almost everywhere in the domain Ω .This means that it is very difficult to achieve an adequate result, without an over-reliance on the user input or parameter selection.
The central premise for applying Chan-Vese type methods is the assumption that the image approximately consists of where η is noise, χ i is the characteristic function of the region Ω i , for i = F, B respectively.The idea of selective segmentation is to incorporate user input to apply constraints that exclude regions classified as foreground, based on their location in the image.We use a distance constraint which penalises the distance from the user input markers.However, a key problem for selective segmentation is that for cases where the optimal intensity values c 1 and c 2 are similar, the intensity fitting term will become obsolete as the contour evolves.This is illustrated in Fig. 3.The purpose of our approach is to construct a model that is based on assumptions that are consistent with the observed image and any homogeneous target region of interest.A common approach in selective segmentation is to discriminate between objects of a similar intensity [34,35,39].However, the fitting terms in previous formulations [24,34,35,39] aren't applicable in many cases as there are contradictions in the formulation in this context.We will address this in detail in the following section.
In this paper our main contribution is to highlight a crucial flaw in the assumptions behind many current selective segmentation approaches and propose a new fitting term in relation to such methods.We demonstrate how our reformulation is capable of achieving superior results and is more robust to parameter choices than existing approaches, allowing for more consistency in practice.In §2 we give a brief review of alternative intensity fitting terms proposed in the literature, and detail them in relation to selective segmentation.We then briefly detail alternative selective segmentation approaches to compare our method against in §3.In §4 we introduce the proposed model, focussing on a fitting term that allows for significant intensity variation in the background domain.In §5 we discuss the implementation of each approach in a convex relaxation framework, provide the algorithm in §6, and detail some experimental results in §7.Finally, in §8 we give some concluding remarks.

Related Approaches
Here, we introduce and discuss work that has introduced alternative data fitting terms closely related to Chan-Vese [15].In order to make direct comparisons, we convert each approach to the unified framework of convex relaxation [14].It is worth noting that this alternative implementation is equivalent in some respects, but that the results might differ slightly if using the original methods.We are considering these models in the terms of selective segmentation, so all formulations have the following structure: We are interested in the effectiveness of f (u) in this context, which we will focus on next.In particular, we detail various choices of f (u) from the literature that are generalisations of the Chan-Vese approach.In the following we refer to minimisers of convex formulations, such as (11), by u γ .Here, the minimiser of F(u) is thresholded for γ ∈ (0, 1) in a conventional way [14].
2.1 Region-Scalable Fitting (RSF) [25] The data fitting term from the work of Li et al. [25], known as Region-Scalable Fitting (RSF), consistent with the convex relaxation technique of [14] is given by where and K σ (x) is chosen as a Gaussian kernel with scale parameter σ > 0. The RSF selective formulation is then given as follows: The functions h 1 (x) and h 2 (x), which are generalisations of c 1 and c 2 from Chan-Vese, are updated iteratively by Using the RSF fitting term, any deviations of z from h 1 and h 2 are smoothed by the convolution operator, K σ .This allows for intensity inhomogeneity in the foreground and background of target objects.
2.2 Local Chan-Vese (LCV) Fitting [43] Wang et al. [43] proposed the Local Chan-Vese (LCV) model.In terms of the equivalent convex formulation, the data fitting term is given by where and z * = M k * z.Here, M k is an averaging convolution with k × k window.The LCV selective formulation is then given as The values c 1 , c 2 , d 1 , d 2 which minimise this functional for u γ are given by The formulation is minimised iteratively.The LCV fitting term that f 1 (x) and f 2 (x) includes an additional term weighted by the parameters α and β .The principle for the LCV model is that the difference image z * − z is a higher contrast image than z and a two-phase segmentation on this image can be computed.

Hybrid (HYB) Fitting [1]
Based on extending the LCV model, Ali et al. [1] proposed the following data fitting term, where Here, z * = M k * z, w = z * z, and w * = M k * w, with M k the averaging convolution as used in the LCV model.The values c 1 , c 2 , d 1 , d 2 are updated in a similar way to [43], with further details found in [1].The authors refer to this approach as the Hybrid (HYB) Model.The HYB selective formulation is then given as The key aim of the HYB model is to account for intensity inhomogeneity in the foreground and background of the image through the product image w.In LCV, the presence of the blurred image z * in the data fitting term deals with intensity inhomogeneity, whilst including z helps identify contrast between regions.The authors found that the product image w = z * z can improve the data fitting in both respects.Therefore they construct a LCV-type function with w rather than the original z.Their results suggest that this approach is more robust.
2.4 Generalised Averages (GAV) Fitting [2] Recently, Ali et al. [2] proposed using the data fitting terms of Chan-Vese in a signed pressure force function framework [48].They refer to this approach as Generalised Averages (GAV) as they update the intensity constants in an alternative way, detailed below.In the convex framework, we consider the selective GAV functional: where f GAV (u) = f CV (u).This is identical to the CV selective formulation (8).However, the authors propose an alternative update for the fitting constants c 1 and c 2 , given as follows: with β ∈ R. If β = 1, the approach is identical to CV.In [2] the authors assert that the proposed adjustments have the following properties.As β → ∞, c 1 and c 2 approach the maximum and minimum intensity in the foreground and background of the image, respectively.Also, as β → −∞, c 1 and c 2 approach the minimum intensity in the foreground and background of the image, respectively.For example, if a high value of β is set, c 1 will take a larger value than in CV which can be useful for selective segmentation.For example, if we consider the image in Fig. 1 we can achieve a larger c 2 value by setting β > 1 and a smaller value by setting β < 1.Therefore, there is more flexibility when using this data fitting term in selective formulations.However, it should be noted that it involves the selection of the parameter β , which can be difficult to optimise.

Alternative Selective Segmentation Models
We now introduce two recent methods that incorporate user input to perform selective segmentation.Each involves input in the form of foreground/background regions to indicate relevant structures of interest.An example of this can be seen in Fig. 18, where red regions indicate foreground and blue regions indicate background.We compare against the work of Nguyen et al. [30], which uses a similar convex relaxation framework to the proposed approach, and Dong et al. [17], which uses a variation of the random walk approach.We summarise the essential aspects of each approach in the following.

Constrained Active Contours (CAC) [30]
The authors use a probability map, P(x), from Bai and Sapiro [4] where the geodesic distances to the foreground/background regions are denoted by D F (x) and D B (x), respectively.An approximation of the probability that a point x belongs to the foreground is then given by Foreground/background Gaussian mixture models (GMM) are estimated from the user input.The terms Pr(x|F) and Pr(x|B) denote the probability that a point, x, belongs to the the foreground and background, respectively.The normalised log likelihood for each is then given by GMMs are widely used in selective segmentation [18,36,4,22,17] and the authors in [30] incorporate this idea into the framework we consider with the following data fitting term: for a weighting parameter α 0 ∈ [0, 1].It is proposed that α 0 is selected automatically as follows: where N is the total number of pixels in the image.Defining g 0 as the function g(s) applied to the image z(x) and g p applied to the GMM probability map P F (x), an enhanced edge function is defined as for a weighting parameter β 0 ∈ [0, 1], which can be set automatically in a similar way to (28).Thus, Nguyen et al. [30] define the Constrained Active Contours (CAC) Model as They obtain a solution using the split Bregman method of Goldstein et al. [19], although other methods are applicable and will yield similar results.However, that is not the focus of this paper so we omit the details here.In the results section, §7, we will compare our method against CAC to see how our data fitting term compares against a GMM-based approach.
3.2 Submarkov Random Walks (SRW) [17] We now introduce a recent selective segmentation method by Dong et al. [17] known as Submarkov Random Walks (SRW).Rather than using the continuous framework of [14], this approach is based in the discrete setting where each pixel in the image is treated as a node in a weighted graph.Random walks (RW) have been widely used for segmentation since the work of Grady [22].SRW is capable of achieving impressive results with user-defined foreground and background regions.The selective segmentation result can be obtained by assigning a label to each pixel based on the computed probabilities of the random walk approach.
For brevity, we do not provide the full details of the method here, however, further details can be found in [17].We compare SRW to our proposed approach on a CT data set in §7.4.
We now introduce essential notation to understand the approach of [17].In RW an image is formulated as a weighted undirected graph G = (V, E) with nodes v ∈ V and edges e ∈ E ⊆ V ×V .Each node v i represents an image pixel x i .An edge e i j connects two nodes v i and v j and a weight w i j ∈ W of edge e i j measures the likelihood that a random walker will cross this edge: where I i and I j are pixel intensities, with σ 0 , ε 0 ∈ R. In SRW a user indicates foreground/background regions in a similar way to CAC, as shown in Fig. 18, and can be viewed as a traditional random walker with added auxiliary nodes.
In [17], these are defined as a set of labelled nodes , and M k the number of seeds labelled l k .The prior is then constructed from the seeded nodes (defined by the user).Assuming a label l k has an intensity distribution H k (based on GMM learning), a set of auxiliary nodes is added into an expanded graph G e to define a graph with prior Ḡ.Each prior node is connected with all nodes in V and the weight, w ih k , of an edge between a prior node h k and a node v i ∈ V is proportional to u k i , the probability density belonging to H k at v i .
The authors define the probabilities of each node v i ∈ V belonging to label l k as the average reaching probability, denoted rl k i .This term incorporates the auxillary nodes introduced above and is dependent on multiple variables and parameters, including w i j (31).Further details can be found in [17].The segmentation result is then found by solving the following discrete optimisation problem: where Ri represents the final label for each node.In other words, for a two-phase segmentation problem, Ri is analogous to the discretised solution of a convex relaxation problem in the continuous setting.Comparisons in terms of accuracy can therefore be made directly, which we elaborate on further in §7.The authors also detail the optimisation procedure and aspects of dealing with noise reduction.

Proposed Model
In this section we introduce the proposed data fitting term for selective segmentation.We consider objects that are approximately homogeneous in the target region.Intrinsically, it is then assumed that the region P, provided by the user, is likely to provide a reasonable approximation of the optimal c 1 value and therefore an appropriate foreground fitting function, f 1 , is given by CV (2).For this reason, it makes sense to retain this term in the proposed approach.The contradiction is in how the background fitting function f 2 is defined.Considering piecewise-constant assumptions of the image, and many of the related approaches, the background is expected to be defined by a single constant value, c 2 .If c 1 ≈ c 2 then f 2 ≈ f 1 everywhere, and therefore the fitting term can't accurately separate background regions from the foreground.It is not practical to rely on f S (u) to overcome this difficulty as it will produce an over-dependence on the choice of M and P.This is prohibitive in practice.An alternative function f 2 must therefore be defined which is compatible with f 1 and f S (u).Here, we define a new data fitting term that penalises background objects in such a way that avoids these problems by allowing intensity variation above and below the value c 1 .In order to design a new functional, we first look at the original CV background fitting function It is clear that in an approximately piecewise-constant image this function will be small outside the target region (i.e.where the image takes values near c 2 ) and larger inside the target region.Our aim in a new fitting term is to mimic this in such a way that is consistent with selective segmentation, where regions with a 'foreground intensity' are forced to be in the background.It is beneficial to introduce two parameters, γ 1 and γ 2 , to enforce the penalty on regions of intensity in the range i.e. enforce the penalty asymmetrically around c 1 .We propose the following function to achieve this: This function takes its maximum value where z(x) = c 1 and is 0 for z(x) > c 1 − γ 1 and z(x) < c 1 + γ 2 .In Fig. 2 we provide a 1D representation of f2 (x) for various choices of γ 1 and γ 2 , with z(x) ∈ [0, 1] and c 1 = 0.5.Here, it can be seen how the proposed data fitting term acts as a penalty in relation to a fixed constant c 1 .It is analogous to CV, whilst accounting for the idea of selective segmentation with a data fitting term.The main advantage of this term is that it replaces the dependence on c 2 in the formulation, which has no meaningful relation to the solution of a selective segmentation problem.Even when the foreground is relatively homogeneous, the background may have intensities of a similar value to c 1 which will cause difficulties in obtaining an accurate solution.We detail the proposed fitting term in the following section.

New Fitting Term
We define the proposed data fitting functional as follows: for f 1 (x) = (z − c 1 ) 2 and f2 (x) as defined in (33).This is consistent with respect to the intensities of the observed object and the concept of selective segmentation.In Fig. 3 we see the difference between CV and the proposed fitting terms for given user input on a CT image.For the CT image, Here, we show the difference between the CV fitting function and the proposed approach.The target region is clearly defined by negative values in (iii).
(i) Image and P (ii) CV fitting (iii) New fitting the CV fitting terms are near 0 within the target region.This is despite there being a distinct homogeneous area with good contrast on the boundary.This illustrates the problem we are aiming to overcome.With the proposed fitting term this phenomenon should be avoided in cases like this.By defining f2 as in (33) there is no contradiction if the foreground and background intensities of the target region are similar.
For images where we assume that the target foreground is approximately homogeneous, we have generally found that fixing c 1 according to the user input is preferable.We compute c 1 as the average intensity inside the region P formed from the user input marker point set.We therefore propose to minimise the following functional with respect to u ∈ [0, 1], given a fixed c 1 : where f S is the geodesic distance computed as described earlier using (6).The minimisation problem is given as The model consists of weighted TV regularisation with a geodesic distance constraint as in [35].However, alternative constraints are possible, such as Euclidean [39], or moments [24].It is important to note that we have defined the model in a similar framework to the related approaches discussed previously.The main idea is to establish how the proposed fitting term, f PM (u), performs compared to alternative methods.Next we describe how we determine the values of γ 1 and γ 2 in the function f2 (x) automatically.This is important in practice as it avoids any additional user input or parameter dependence to achieve an accurate result.In subsequent sections we provide details of how we obtain a solution for the proposed model.

Parameter Selection
For a particular problem it is quite straightforward to optimise the choice of γ 1 and γ 2 experimentally, but we would like a method which is not sensitive to the choice of γ 1 and γ 2 and would also prefer that the user need not choose these values manually.Therefore, in this section we explain how to choose these values automatically based on justifiable assumptions about general selective segmentation problems.
To select the parameters γ 1 and γ 2 we use Otsu's method [32] to divide the histogram of image intensities into N partitions.Otsu's thresholding is an automatic clustering method which chooses optimal threshold values to minimise the intra-class variance.This has been implemented very efficiently in MAT-LAB in the function multithresh for dividing a histogram such that there are N − 1 thresholds T i .We use the thresholds from Otsu's method to find γ 1 and γ 2 as follows.There are three cases to consider, based on the value of c 1 computed from the user input: i) For each case we set the parameters as follows: (iii) Test Image 9 Choosing N too large could mean γ 1 and γ 2 are too small as the histogram would be partitioned too precisely.Generally we only ever need to consider a maximum of 3 phases for selective segmentation.If there is a large number of pixels in the image with intensity above or below c 1 the image can be considered two-phase in practice.Conversely, if a large number of pixels in the image have intensity above and below c 1 the image can essentially be considered three-phase in the context of selective segmentation.This is due to the way f2 has been defined.Therefore, we set N = 3 for all tests.In Fig. 4 we can see the Otsu thresholds chosen for various images given in this paper.They divide the peaks in the histogram well and once we know the value of c 1 (the approximation of the intensity of the object we would like to segment) we can automatically choose γ 1 and γ 2 according to this criteria.

Numerical Implementation
We now introduce the framework in which we compute a solution to the minimisation of the proposed model, as well the related models introduced in §1 and §2.All consist of the minimisation problem for X = CV, RSF, LCV, HYB, GAV, PM respectively.Minimisation problems of this type (37) have been widely studied in terms of continuous optimisation in imaging, including two-phase segmentation.A summary of such methods in recent years is given by Chambolle and Pock [13].Details of the introduction of binary labels to image segmentation can be found in Lie et al. [26] and Chan et al. [14], and our numerical scheme follows the approach in [14]: enforcing the constraint in (37) with a penalty function, and deriving the Euler-Lagrange of the regularised functional.We then solve the corresponding PDE by following a splitting scheme first applied to this kind of problem by Spencer and Chen [39].
Whilst the numerical details are not the focus of the work, it is important to note widely used alternatives.A summary of such approaches, describing major developments in this area and the connections between each method is given in a review by Wei et al. [44].
It has proved very effective to exploit the duality in the functional and avoid smoothing the TV term.A prominent example is the split Bregman approach for segmentation by Goldstein et al. [19].This is closely related to augmented lagrangian methods, a matter further discussed by Boyd et al. [7].Analogous approaches also consist of the first-order primal dual algorithm of Chambolle and Pock [12] and the max-flow/min-cut framework detailed by Yuan et al. [47].There are practical advantages in implementing such a numerical scheme for our problem, primarily in terms of computational speed.However, in the numerical tests we include we're mainly interested in accuracy comparisons.For this purpose the convex splitting algorithm of [39] is sufficient, and the extension of splitting schemes for convex segmentation problems may be of interest.Further details can be found in [39] and [35].In the following, we first discuss the minimisation of (37) in a general sense and then mention some important aspects in relation to the alternative fitting terms discussed in §2.

Finding the Global Minimiser
To solve this constrained convex minimisation problem (38) we use the Additive Operator Splitting (AOS) scheme from Gordeziani et al. [20], Lu et al. [28] and Weickert et al. [45].This is used extensively for image segmentation models [34,35,39].It allows the 2D problem to be split into two 1D problems, each solved separately, with the results combined in an efficient manner.We address some aspects of AOS in §6, with further details provided in [35,39].
A challenge with the functional (35), particularly with respect to AOS, is that this is a constrained minimisation problem.Consequently, it is reformulated by introducing an exact penalty function, ν(u), given in [14].To simplify the formulation we define r(x) = θ D(x) + f (x), f (x) is the function associated with f X (u).We introduce a new parameter, λ , which allows us to balance the data fit-ting terms to the regularisation term more reliably.To be clear, we still only have two main tuning parameters (θ and λ ) as we fix any variable parameters in f (x) according to the choices in the corresponding papers.The unconstrained minimisation problem is then given as: We rescale the data term with F (x) = r(x)/||r(x)|| ∞ .In effect this change is simply a rescaling of the parameters.This allows for the parameter choices between different models to be more consistent, as the fitting terms are similar in value.The problem (38) has the corresponding Euler-Lagrange equation (for fixed c 1 ): in Ω and ∂ u ∂ n = 0 where n is the outward unit normal.The constraint is enforced for α > λ 2 ||r(x)|| by [14].Two parameters, ε 1 and ε 2 , are introduced here.The former is to avoid singularities in the TV term and the latter is associated with the regularised penalty function ν ε 2 (u) from [39]: with b ε 2 (u) = (2u − 1) 2 + ε 2 − 1 and regularised Heaviside function The viscosity solution of the parabolic formulation of (39), obtained by multiplying the PDE by |∇u|, exists and is unique.
The general proof for a class of PDEs to which (39) belongs, is included in [35] and we refer the reader there for the details.Once the solution to (39) is found, denoted u * , we define the computed foreground region as follows: We select γ = 0.5 (although other values γ ∈ (0, 1) would yield a similar result according to Chan et al. [14]).In the following we use the binary form of the solution, u * , denoted u γ .This partitions the domain into Ω F and Ω B according to the labelling function u γ .

Implementation for Related Models
The discussion in this section so far has used the function f (x) associated with the data fitting functional f X (u).This corresponding equations for the RSF, LCV, HYB and GAV models are detailed in §2, CV is discussed in §1, and our approach is given by eqn.(34).We use this implementation to obtain selective segmentation versions of each of those models, given by (37).When these terms contain parameter choices we follow the advice in the corresponding papers as far as possible, unless we have found that alternatives will improve results.In the next section we will give the results of these models and compare them to our proposed approach.
Note.We now discuss details behind tuning parameters for the GAV model.It is noted in §2 that the GAV model requires a parameter β to adapt the c 1 and c 2 calculation.We find that it is actually better to consider c 1 and c 2 separately to achieve improved results, as sometimes we wish to tune the values to have a higher c 1 and lower c 2 (or vice-versa) simultaneously.Therefore we introduce parameters β 1 and β 2 to tune c 1 and c 2 as follows: In all experiments, we tested the following combinations of (β 1 , β 2 ): (1.5, 0.5), (2, 0), (3, −1), (4, −2), (0.5, 1.5), (0, 2), (−1, 3) and (−2, 4).For each choice, we optimised the values of λ and θ according to the procedure described in §7.1.This allowed us to select the optimal combination of (β 1 , β 2 ) for each image.

Algorithm
Here, we will discuss the algorithm that we use to minimise the selective segmentation model (37).We utilise additive operator splitting techniques to solve the minimisation problem efficiently.

An Additive Operator Splitting (AOS) Scheme
Additive Operator Splitting (AOS) [20,28,45] is a widely used method for solving PDEs with linear and non-linear diffusion terms [34,35,39] such as AOS allows us to split the two-dimensional problem into two one-dimensional problems, which we solve separately and then combine.Each one-dimensional problem gives rise to a tridiagonal system of equations which can be solved efficiently by Thomas' algorithm, hence AOS is a very efficient method for solving PDEs of this type.AOS is a semiimplicit method and permits far larger time-steps than the corresponding explicit schemes would.Hence AOS is more stable than an explicit method [45].Note here that and µ = 1.The standard AOS scheme assumes f 0 does not depend on u, however in this instance that is not the case.This requires a modification to be used for convex segmentation problems, first introduced by [39].This non-standard formulation incorporates the regularised penalty term, ν ε 2 (u), into the AOS scheme which we briefly detail next.
The authors consider the Taylor expansions of ν ε 2 (u) around u = 0 and u = 1.They find that the coefficient b of the linear term in u is the same for both expansions.Therefore, for a change in u of δ u around u = 0 and u = 1 the change in ν ε 2 (u) can be approximated by b • δ (u).To address this, the relevant interval is defined as and a corresponding update function is given as b The solution for ( 44) is then obtained by discretising the equation as follows: where A 1 and A 2 are discrete forms of ∂ x (G(u)∂ x ) and ∂ y (G(u)∂ y ), respectively (given in [39,35]).The modified AOS update is then given by where 0 .This scheme allows for more control on the changes in f 0 between iterations due to the function b and parameter ζ , and therefore leads to a more stable convergence.We refer the reader to [39] for full details of the numerical method.

The Proposed Algorithm
In Algorithm 1 we provide details of how we find the minimiser of the various selective segmentation models detailed above, defined by (37).The algorithm is in a general form to be applied to any of the approaches discussed so far.It is important to reiterate that alternative solvers to AOS are available, such as the dual formulation [3,8,11], split-Bregman [19], augmented Lagrange [6], primal dual [12], and maxflow/min-cut [47].In all experiments we use the tolerance of 10 −4 for the stopping criteria and set ε 1 = 10 −4 , ε 2 = 10 −1 and τ = 10 −2 .

Results
In this section we will present results obtained using the proposed model and compare them to using fitting terms from similar models (CV [15], RSF [25], LCV [43], HYB [1], Fig. 6: Test Images 4-6; the ground truth contours are defined in the first row and the corresponding user input marker set is shown in the second row.These are real images with some degree of intensity inhomogeneity in the foreground, with potential medical applications in mind.

Test Image 4
Test Image 5 Test Image 6 GAV [2]), detailed in §2, and additional comparisons to alternative selective models.Specifically, we compare against the work of Nguyen et al. [30] and Dong et al. [17], referred to as CAC and SRW respectively and detailed in §3.We intend to provide an overview of how effective each approach is in a number of key respects and analyse their potential for practical use in a reliable and consistent manner.Our focus is on how each fitting term can be applied to a consistent selective segmentation framework, and how robust the proposed model is overall.The key questions we consider are: (i) How sensitive are the results to variations of the parameters λ and θ ?(ii) Is the model capable of achieving accurate results?(iii) To what extent is the proposed model dependent on the user input?(iv) Does the model compare favourably against alternative selective methods?
Test Images.We will perform initial tests on the images shown in Figs.5-7.We have provided the ground truth and initialisation used for each image.Test Images 1-3 are synthetic, Test Image 4 is an MRI scan of a knee, Test Images 5-6 are abdominal CT scans, and Test Images 7-9 are lung CT scans.They have been selected to present challenges relevant to the discussion in §2.We focus on medical images as this is the application of most interest to our work.In the following we will discuss the results in terms of synthetic images (1-3) and real images (4)(5)(6)(7)(8)(9).We also test the proposed approach on a larger data set of 30 CT images (a sample of which is presented in Fig. 18), comparing against existing selective methods detailed in §3.
Measuring Segmentation Accuracy.In our tests we use the Jaccard Coefficient [23], often referred to as the Tanimoto Coefficient (TC), to measure the quality of the seg-mentation.We define accuracy with respect to a ground truth, GT , given by a manual segmentation: The Tanimoto Coefficient is then calculated as where N(•) refers to the number of points in the enclosed region.This takes values in the range [0, 1], with higher TC values indicating a more accurate segmentation.In the following we will represent accuracy visually from red (TC = 0) to green (TC = 1), with the intermediate scaling of colours used shown in Fig. 8.This will be particularly relevant in §7.2.Note.In §2.4 we mentioned the tuning of parameters in the GAV model.To be explicit the optimal (β 1 , β 2 ) pairs used in the following tests were (4,-2) for Test Images 1 and 2, (1.5,0.5) for Test Images 3,4, and 6, (2,0) for Test Image 5, and (-2,4) for Test Images 7,8, and 9. Results vary significantly as (β 1 , β 2 ) are varied, but we found these to be the best choices for each image.
The discussion of results is split into four sections, addressing the questions introduced above.First, in §7.1, we will examine the robustness to the parameters λ and θ for each model.Then, in §7.2, we will compare the optimal accuracy achieved by each method to determine what they are capable of in the context of selective segmentation for these examples.In §7.3, we will test the proposed model with respect to the user input.By randomising the input we will determine to what extent the proposed model is suitable for use in practice.Finally, in §7.4 we will compare the proposed approach to the methods introduced in §3 on an additional CT data set.This will help further establish how the In these tests we aim to demonstrate how sensitive to parameter choices each choice of fitting term is.To accomplish this we perform the segmentations for each of the models discussed (CV, RSF, LCV, HYB, GAV) and the proposed model for a wide range of parameters and compute the TC value.The parameter range used is λ , θ ∈ [1,50].Due to computational constraints, we run for each integer λ , θ between 1 and 10, and every fifth from 15 to 50.This aspect of a model's performance is vital when used in practice.The less sensitive to parameter choices a model is the more relevant it is in relation to potential applications.It should be noted that we neglect to test the selective models detailed in §3 with respect to parameter robustness as we are using the authors' implementation of each approach.Instead, we make direct comparisons in the following sections.
The TC values for the parameter sets ( λ , θ ) are presented as heatmaps in Figs.11-13.A heatmap is a convenient way to display accuracy results for hundreds of tests concisely.In Fig. 9 we give an example heatmap with the same axes used for those in Figs.11-13.For each of the combinations of parameter values ( λ , θ ) we give the TC value of the segmentation result and represent it by the appropriate colour.The corresponding colour scale is shown in Fig. 8. Qualitatively, the more green areas of the heatmap the more accurate the model is for a wider set of parameters.Example results for Test Image 5 when varying λ (with θ = 4) for the proposed model are given in Fig. 10.Here it can be seen what each accuracy result corresponds to visually.
Note.The axes have been removed from the heatmaps in Figs.11-13 for presentational clarity.However, to be explicit, the axes used in all heatmaps are the same as those in Fig. 9. ).The colours correspond to the TC value (green is TC = 1, red is TC = 0), consistent with the scale in Fig. 8.This is for Test Image 5, with the corresponding heatmap provided in Fig. 12.
Synthetic Images.These results are presented in Fig. 11.For Test Images 1-2 we see poor parameter robustness from all competing models, except for GAV which performs reasonably well.However, the proposed model has minimal parameter sensitivity for these images, with good results achieved for almost every combination of values tested.For Test Image 3 all models have a reasonable parameter range (except for RSF), however the proposed model gives better quality results for a wider parameter range.The other models achieve reasonable results here as the foreground intensity of the ground truth is greater than the background (c 1 = 0.75, c 2 = 0.49), whereas for Test Images 1-2 they are equal (c 1 = c 2 = 0.50).These results highlight the key advantage of the proposed model.
Real Images.In Fig 12 we present results for Test Images 4-6.Here, the proposed model performs in a similar way to its competitors because these images are more typical selective segmentation problems in the sense that there is a clear distinction between the foreground and background intensities.In particular, the values in each case are: Test Image 4 (c 1 = 0.85, c 2 = 0.25), Test Image 5 (c 1 = 0.70, c 2 = 0.19), and Test Image 6 (c 1 = 0.73, c 2 = 0.20).It can be seen that the proposed model is competitive compared to previous approaches.The performance is quite poor for Test Image 5, but is arguably still the best for this challenging case.In Fig. 13 we present results for Test Images 7-9.Here the proposed model outperforms previous approaches significantly for each image.This is mainly due to the type of image considered.Specifically, the true intensities are: RSF [25] LCV [43] HYB [1] GAV [2] Proposed CV [15] RSF [25] LCV [43] HYB [1] GAV [2] Proposed CV [15] RSF [25] LCV [43] HYB [1] GAV [2] Proposed

Accuracy Comparisons
Here we aim to address the question of whether each model is capable of achieving an accurate result.In other words, assuming that factors such as parameter and user input sensitivity are ignored, how successful is each approach.In Table 1 we present the optimal TC values for each model found from the tests described in the previous section, with the highest value in bold.We include values for CAC [30] and SRW [17], which we have obtained by iteratively refining the user input and running the algorithm.It is worth mentioning that we are using the authors' implementation of each method.For each image, the results presented in Table 1  ).Below we will discuss some relevant details of the results, again by splitting the test images into synthetic and real.Synthetic Images.We observe that for Test Images 1 and 2 (where c 1 = c 2 , CV, LCV, and HYB fail completely.GAV performs well, with the proposed model and RSF being the most accurate with perfect results.For Test Image 3, all models are capable of achieving a good result.It should be noted that in this case c 1 = 0.75 and c 2 = 0.49.This difference enables the other models to perform well, although the proposed model is slightly superior with a perfect result.The alternative selective models also perform well for these images, although CAC has minor errors on the boundaries of the foreground for each image. Real Images.In Table 1 we can see that the proposed model is the most successful in terms of optimal accuracy.It is worth noting some inconsistency in the other models, with all but GAV having results that fall below TC = 0.9 for at least one image.GAV performs well for Test Images 4-9, with the proposed model slightly outperforming it in each case.It is worth reminding the reader that for GAV the parameters (β 1 , β 2 ) have been refined for each example.Fixing this results in more variability in the quality of results.The proposed model has no such parameter optimisation between examples.CAC and SRW perform reasonably well for these images, although are sometimes substandard for Test Images 4-7.This is despite extensive refinement of the user input to achieve an acceptable result.We present the optimal results for Test Image 9 in Fig. 14.Here we can see how much variation there is in the quality of results for this lung CT image.CAC and SRW are competitive in this instance.Of the remaining approaches GAV is the most competitive (TC = 0.919), but is visually inadequate.Two other models (CV, HYB) fail completely.In this case, the problem looks quite straightforward and yet other fitting terms are insufficient to produce a good result.Again, the proposed model tends to be superior in cases where c 1 ≈ c 2 and is capable of achieving very good results for all the images considered.This highlight the advantages of the proposed fitting term.

User Input Randomisation
One key consideration for the practical use of selective segmentation models is that the result is not too reliant on user input.With intricate user input accurate results are almost guaranteed.However, the benefit of this kind of approach is that accuracy should be attainable with minimal, intuitive user input.One challenge in this setting is how to ascertain (i) CV [15], TC = 0.18 (ii) RSF [25], TC = 0.78 (iii) LCV [43], TC = 0.83 (iv) HYB [1], TC = 0.00 (v) GAV [2], TC = 0.92 (vi) CAC [30], TC = 0.95 (vii) SRW [17], TC = 0.96 (viii) Proposed, TC = 0.97 to what extent a method is dependent on the user input.In this section we will generalise the user input for the proposed model in order to determine how sensitive it is in this respect.By generalising in this way we will make two assumptions about the markers, M , consistent with the above considerations: (i) All points are within the target object.
(ii) Only 3 markers are selected.
We regard neither of these assumptions to be too onerous on a user, and are quite consistent with practical use.To perform this test, we randomly choose 1000 sets of 3 marker points and run each algorithm using them.The parameters λ and θ are fixed at those which gave the optimal TC values in Table 1.For each set of marker points we compute the corresponding TC value of applying the proposed model with this input.The results for each image are summarised by boxplots in Fig. 15 with examples of the worst results, excluding outliers, shown in Fig. 16.Here, it can be seen that the worst result often outperforms the optimal results of the alternative models considered, which is impressive.Below we discuss the results for the test images, by again splitting them into synthetic and real images.Based on the authors' implementation of CAC and SRW it was not possible to generalise the input in this way.Instead we make direct comparisons of input in the next section.
Synthetic Images.For the Test Images 1-3 we achieve near perfect segmentations in all cases, shown by the mean TC being between 0.99 and 1.00 in all cases (for Test Image 1, the mean is precisely 1.00) and a small variance around the mean.Therefore, we can conclude that for images of this type, where the foreground is homogeneous, our method is very robust to user input.Essentially, any reasonable set of markers should produce excellent results.It should be noted that the optimal results from comparable approaches are less than the mean result of 1000 random tests for our method (except for SRW).This can be observed in Table 1.Furthermore, these methods often fail completely.This is a key result highlighting the advantages of our method.In visually simple cases (Test Images 1-3) our new data fitting term is an improvement on existing approaches by modifying the underlying assumptions involved.
Real Images.In all cases for Test Images 4-9 the mean values show that the segmentation results are highly accurate.Also, we notice that the variances are very reasonable demonstrating the robustness of varying the user input.This is an important aspect of selective segmentation, and highlights the advantages of the proposed fitting term.For Test Images 4-6 we observe more variability in the accuracy due to minor intensity inhomogeneity in the foreground.This means randomising the user input will be more sensitive.However, we can see that the results are very good with Fig. 15: Boxplots of the TC values for 1000 random user inputs using the proposed model.We observe that the method is remarkably consistent.Even the worst results, excluding outliers, are competitive with the optimal results of the existing approaches shown in Table 1.
the mean accuracy being competitive with the optimal accuracy of comparable methods.In the case of the lung CT images (Test Images 7-9) the variance in TC values is very small, due to the homogeneity of the foreground.Again, it is important to compare the results of 1000 random results using our proposed model to the optimal result of comparable methods.For these images all of the methods (except GAV,CAC, and SRW) have at least one TC value below 0.9.However, GAV requires the tuning of additional parameters (β 1 , β 2 ) whilst the proposed model does not.The results for CAC and SRW also rely on extensive requirements of the user input to achieve this accuracy, whereas random input compares favourably here.Compared to GAV, we can see that the mean of our tests is similar to the optimal value of GAV.One exception is for Test Image 9 (shown in Fig. 14), where there is a significant gap in favour of our model.Again, from Fig. 16, we can see that the worst result of randomising the user input for the proposed model is competitive with the optimal results of the alternatives.This is one of the most encouraging aspects of the tests; the proposed model is remarkably robust to varying user input.This proves that successful results with minimal, intuitive user input is possible for a range of examples.

Alternative Selective Methods
In order to further establish the robustness of our method, we now introduce the results of testing our approach against competing interactive segmentation methods on a larger data set.The results are presented in Fig. 17, showing a boxplot of accuracy in terms of TC on a set of 30 CT images (excluding outliers).The target structure we consider is the spleen, as this consists of a relatively homogeneous foreground, appropriate for the approach considered.The data has been manually contoured providing ground truth data for the image set.We compare CAC [30] and SRW [17] against our method with five variations of user input for each image.It is worth emphasising here that the input used in the tests is identical for each approach and was not refined in any way.It was designed to mimic what a user, unfamiliar with each approach, might select intuitively.A representative example for three images is shown in Fig. 18.This shows foreground (red) and background (blue) user input regions.For our method, we define the red region as P as discussed in §1 and enforce hard constraints on the blue region.We refer to the results of the proposed approach using this input as Ours (i).We also include results of randomising the user input in an identical way to §7.3.For each image we generate 1000 simulated user input choices, which we present as Ours (ii).It is important to note that the difference between Ours (i) and (ii) is only the definition of P. The method and parameters are fixed between each.
The performance of CAC [30] is very good, as shown in Fig. 17.We have included an additional figure to highlight the difference between CAC and Ours (i) and (ii) more precisely.This is shown in Fig. 19 (this is the same as Fig. 17 with TC restricted to [0.8,1]).Here we can see that the proposed approach has a slightly better median (0.96 compared to 0.94) and is generally more consistent than CAC.This is particularly evident when considering the worst TC results of CAC (0.19) against ours (0.87).
In Fig. 17 it can be seen that our method exceeds the performance of SRW by a large margin (0.66 compared to 0.95).One possible reason for this is that the input used, as displayed in Fig. 18, is restricted to be as intuitive as possible.SRW is capable of achieving improved results with more elaborate foreground/background input.However, it is generally reliant on a trial and error approach which is not ideal in practice.This highlights an important advantage of our method.It is able to achieve a high standard of results with simple user input.This is reinforced by considering Ours (ii), where the results of 30000 random variations of the user input does not cause a drop off in accuracy compared to the 150 manual user input selections.Again, this can be seen more clearly in Fig. 19.In fact, the results for the proposed approach with the random input are slightly better than with the manual input.This underlines the robustness to user input in the model, which is a vital aspect of selective segmentation.

Conclusion
In this paper we have proposed a new intensity fitting term, for use in selective segmentation.We have compared it to fitting terms from comparable approaches (CV, RSF, LCV, HYB, GAV), in order to address an underlying problem in selective segmentation: if the foreground is approximately homogeneous what is the best way to define the intensity fitting term?Previous methods [34,35,39] involve contradictions in the formulation, which we attempt to address.
We have evaluated the success of the proposed model in four respects: parameter robustness, optimal accuracy, dependence on user input, and comparisons to competing selective models.Our focus is on medical applications, where the target object has approximately homogeneous intensity.In each way, the proposed model performs very well, particularly in cases where the true foreground and background intensities are similar.We have shown that our method is remarkably insensitive to varying user input, highlighting its potential for use in practice, and also outperforms competitive algorithms in the literature.Fig. 17: Boxplots of the TC values comparing our method to CAC [30] and SRW [17] for 30 test images.Ours (i) refers to using identical user input to CAC and SRW, with a sample shown in Fig. 18.Ours (ii) refers to 1000 random variations of the user input for each image.Fig. 18: Examples of the input used to compare our method to CAC [30] and SRW [17].Each row represents an image in the dataset and we present five variations of the input used in the tests described in §7.4.  .Here, the extent to which the proposed method outperforms CAC [30] is clearer for both types of input.

Fig. 1 :
Fig. 1: CT image with ground truth segmentation shown (green) and associated average intensity values (c 1 and c 2 ).

Fig. 4 :
Fig. 4: The histograms of intensities for some example images.The red lines are the automatic thresholds T i obtained by Otsu's thresholding with N = 3.

Fig. 5 :
Fig. 5: Test Images 1-3; the ground truth contours are defined in the first row and the corresponding user input marker set is shown in the second row.These are synthetic images with homogeneous foregrounds selected to highlight the benefits of the proposed model.

Fig. 7 :Fig. 8 :Fig. 9 :
Fig. 7: Test Images 7-9; the ground truth contours are defined in the first row and the corresponding user input marker set is shown in the second row.These are real images with approximately homogeneous foregrounds.The challenge is that the background contains substantial regions of a similar intensity.

Fig. 10 :
Fig.10: Segmentation results and TC values for the proposed model whilst varying λ (with θ = 4).The colours correspond to the TC value (green is TC = 1, red is TC = 0), consistent with the scale in Fig.8.This is for Test Image 5, with the corresponding heatmap provided in Fig.12.

Fig. 11 :
Fig. 11: Heatmaps of TC values for permutations of λ and θ .Each row and column is labelled according to the model used and the image tested.The colour is consistent with the scale in Fig. 8. Here, we present Test Images 1 -3.

Fig. 12 :
Fig. 12: Heatmaps of TC values for permutations of λ and θ .Each row and column is labelled according to the model used and the image tested.The colour is consistent with the scale in Fig. 8. Here, we present Test Images 4 -6.

Fig. 13 :
Fig. 13: Heatmaps of TC values for permutations of λ and θ .Each row and column is labelled according to the model used and the image tested.The colour is consistent with the scale in Fig. 8. Here, we present Test Images 7 -9.

Fig. 14 :
Fig.14:We present the optimal results for Test Image 9. The accuracy is represented by colour, consistent with the scale in Fig.8.The proposed model often significantly outperforms previous approaches in this case.

Fig. 16 :
Fig. 16: Results for the proposed model for each image, including TC values.The worst result, excluding outliers, of 1000 random user inputs for each example is presented.This demonstrates that the model is robust to user input, with poor results being competitive with the optimal result of competitors.
algorithms for imaging and vision" where work on this paper was undertaken.This work was supported by EPSRC grant no EP/K032208/1.The first author wishes to thank the UK EPSRC, the Smith Institute for Industrial Mathematics, and the Liverpool Heart and Chest Hospital for supporting the work through an Industrial CASE award.The second author would like to acknowledge the support of the EPSRC grant EP/N014499/1.This work was generously supported by the Wellcome Trust Institutional Strategic Support Award (204909/Z/16/Z).

Fig. 19 :
Fig. 19: Boxplots of the TC values from Fig. 17 for TC ∈ [0.8, 1].Here, the extent to which the proposed method outperforms CAC[30] is clearer for both types of input.
are the most accurate we could obtain given a reasonable level of input (comparisons with identical input are discussed in §7.4).Immediately we can see that the proposed model consistently outperforms the other models in terms of accuracy for the test images (RSF equals it for Test Image 1, SRW equals it for Test Images 1-3, and beats it for Test Image 8