Next Article in Journal
X-ray Imaging Investigation on the Gilding Technique of an Ancient Egyptian Taweret Wooden Statuette
Next Article in Special Issue
An Optimization-Based Meta-Learning Model for MRI Reconstruction with Diverse Dataset
Previous Article in Journal
A Temporal Boosted YOLO-Based Model for Birds Detection around Wind Farms
Previous Article in Special Issue
Fast Fiber Orientation Estimation in Diffusion MRI from kq-Space Sampling and Anatomical Priors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On a Variational and Convex Model of the Blake–Zisserman Type for Segmentation of Low-Contrast and Piecewise Smooth Images

Liverpool Centre of Mathematics for Healthcare and Centre for Mathematical Imaging Techniques, Department of Mathematical Sciences, University of Liverpool, Liverpool L69 7ZL, UK
*
Author to whom correspondence should be addressed.
J. Imaging 2021, 7(11), 228; https://doi.org/10.3390/jimaging7110228
Submission received: 6 September 2021 / Revised: 10 October 2021 / Accepted: 22 October 2021 / Published: 28 October 2021
(This article belongs to the Special Issue Inverse Problems and Imaging)

Abstract

:
This paper proposes a new variational model for segmentation of low-contrast and piecewise smooth images. The model is motivated by the two-stage image segmentation work of Cai–Chan–Zeng (2013) for the Mumford–Shah model. To deal with low-contrast images more effectively, especially in treating higher-order discontinuities, we follow the idea of the Blake–Zisserman model instead of the Mumford–Shah. Two practical ideas are introduced here: first, a convex relaxation idea is used to derive an implementable formulation, and second, a game reformulation is proposed to reduce the strong dependence of coupling parameters. The proposed model is then analysed for existence and further solved by an ADMM solver. Numerical experiments can show that the new model outperforms the current state-of-the-art models for some challenging and low-contrast images.

1. Introduction

Image segmentation is a widely studied yet still challenging subject, especially for new and emerging imaging modalities where Mumford–Shah and extremely strong noise may be present. Of course, extremely simple images with clear contrast, without noise and without blur may be segmented by the simple methods, such as thresholding the image intensity values.
Real-life images inevitably have noise and low contrast which poses a challenge for the simple algorithms. Variational segmentation models generally provide more robust solutions for complex images and can usually be categorised loosely into two categories: edge-based or region-based models. Well-known edge-based methods include Kass et al. [1] and Caselles et al. [2]. Region-based models are generally referred to the pioneering work of Mumford–Shah (MS) [3], with some simplified variants such as Chan–Vese [4,5] that are most widely used.
In the last few years, when mentioning segmentation of challenging images, we would automatically recommend machine-learning-based approaches such as the UNet [6] and Resnet [7]. However, such works are data-dependent, and often, networks are tailored to a specific task. Firstly, they require training data which may not be available (or reliably available) at all. Secondly, we cannot yet conduct automatic transfer learning from a subject area to another to overcome the lack of sufficient training data, e.g., aircraft identification network cannot be adapted to identification of livers in medical imaging. A reliable way of overcoming the lack of sufficient training data is by weakly or semi-supervised learning which uses a small set of training data (in a supervised way) and a larger set of data without annotations (in an unsupervised way) [8,9]. Here, ‘unsupervised’ means that a suitable segmentation model is required; developing such a model is our aim.
This paper addresses the fundamental problem of how to segment low-contrast images where image features of interest have piecewise smooth intensities. In fact, the difficulties of the two problems, namely low-contrast and piecewise smooth features, are well-known challenges. Low contrast implies that edge information by way of image gradients alone is not sufficient enough to detect weak jumps. Moreover, many well-known models such as [4] or its variants assume an input image has approximately piecewise constant intensities; piecewise smooth features imply these models cannot segment such features (or a feature would be split into sub-regions (or multiple phases) according to the intensity distribution, which means that the segmentation is already incorrect). Many approximation models based on the MS [3] can deal with segmentation of piecewise smooth features but not necessarily images displaying low contrast.
Therefore, this paper considers the Blake–Zisserman model [10] which can improve on the MS model [3]. The model [10] cannot be implemented directly and exactly, just as with the MS [3], which was never solved directly.
The rest of the paper is organised as follows. Section 2 briefly reviews related segmentation models. Section 3 introduces our new model and a game theory reformulation to facilitate subsequent solutions. Proof of the solution existence of the game formulation is given. Section 4 presents our numerical algorithm for the game formulation, and Section 5 shows numerical experiments. Brief conclusions are drawn in Section 6.

2. Related Works

The above-mentioned Mumford–Shah model [3] minimises the following:
F M S ( g , Γ ) = λ 2 Ω ( f g ) 2 d x + Ω \ Γ | g | 2 d x + ν H 1 ( Γ ) ,
given the input (possibly noisy) image f : Ω R 2 , where, most importantly, the segmentation is defined by the unknown boundary Γ , g : Ω R 2 is a piecewise smooth approximation of f, and H 1 ( Γ ) denotes the Hausdorff measure (i.e., length of the boundary). In the literature, there are many follow-up works of this paper, proposed to make revised models implementable numerically. Successful results have been obtained. See [11,12,13], among others.
However, for images that have weak edges possibly buried in noise and blur, the Mumford–Shah type models may fail to capture the ‘discontinuities of second kind’ or gradient discontinuity, which may be called the staircasing effect for gradients. The Blake–Zisserman (BZ) type model [10], though less well-known and published earlier than [3], can be very useful for a class of challenging images where MS is less effective; e.g., see [14,15]. The functional of a BZ model takes the form
F B Z ( g , Γ , Γ ) = λ 2 Ω ( f g ) 2 d x + Ω \ ( Γ Γ ) | 2 g | 2 d x + ν 1 H 1 ( Γ ) + + ν 2 H 1 ( Γ \ Γ ) ,
where g , g B V ( Ω ) . Here, Γ is the discontinuity of g . As with the original formulation (1), the BZ model (2) is theoretical, not in a readily solvable form. This paper will propose an approximate and solvable model.
Our work is motivated by Cai–Chan–Zeng [12], who derived a solvable and convex model for (1). We now review this model briefly. As a first step of reformulation of (1), Cai–Chan–Zeng [12] rewrites (1) in an equivalent form
E ( Σ , g 1 , g 2 ) = λ 2 Ω \ Σ f g 2 2 d x + μ 2 Ω \ Σ g 2 2 d x + λ 2 Σ \ Γ f g 1 2 d x + μ 2 Σ \ Γ g 1 2 d x + H 1 ( Γ ) ,
where Γ is assumed to be a Jordan curve as the boundary Σ for the closed domain Σ = I n s i d e ( Γ ) ¯ . Hence, g 1 , g 2 are defined in the inside and outside of Γ , respectively. Of course, both g 1 , g 2 can be smoothly extended to the entire domain Ω . A key observation in [12], motivated by [5], is that the term H 1 ( Γ ) , which is the length of Γ , may be approximated by Ω | g 1 | d x . Then, viewing the smooth functions g 1 , g 2 as a single function, the model by [12] is the following:
min g λ 2 Ω f g 2 d x + μ 2 Ω g 2 d x + Ω g d x .
We now propose a solvable model based on the Blake–Zisserman model (2). Assume the given image is f, and our approximation is g W 1 , 2 ( Ω ) , with g ( W 1 , 2 ( Ω ) ) 2 .
Motivated by the work of [12], we shall respectively approximate the key quantities H 1 ( Γ ) , H 1 ( Γ \ Γ ) by Ω | g | d x , Ω | ( g ) | d x . Therefore, our initial minimisation model takes the form
E B Z ( g ) = λ 2 Ω ( g f ) 2 d x + ν 1 2 Ω | 2 g | 2 d x + ν 2 2 Ω | g | 2 d x + Ω | 2 g | d x + Ω | g | d x .
While (5) is well-defined in terms of solvability, to facilitate the choice of coupling parameters, we now consider a game formulation. A game formulation encourages independent players to complete with each. Here, each player is a sub-problem in an optimisation formulation; see [16]. Here, independence means that parameters of sub-problems do not have to rely on each other.

3. The New Model and Reformulation as a Nash Game

In this work, we are interested in a particular case of a two-player game formulation. Instead of optimising the single energy (5), we consider a game reformulation, where two individuals, or ‘players’, are involved. The first player is the variable g, and the second one will be introduced by using the idea of operator splitting [17] to reduce the high-order derivatives in (5) as first-order terms and to simplify subsequent solution. The solution to this game is the Nash equilibrium, whose existence must be established. For important techniques and results in game theory and its connections to partial differential equations (PDEs) for other problems, the reader is directed to [18,19,20,21].
More precisely, let G be an approximation for vector g . Then, we propose our new model, approximating (5), as
min g E 1 ( g , G ) = λ 1 2 | | f g | | 2 2 + μ 1 2 | | g | | 2 2 + | | g | | 1 + ξ 1 2 | | G g | | 2 2 ,
min G E 2 ( G , g ) = λ 2 2 | | f G | | 2 2 + μ 2 2 | | G | | 2 2 + | | G | | 1 + ξ 2 2 | | G g | | 2 2 ,
where g W 1 , 2 ( Ω ) and G ( W 1 , 2 ( Ω ) ) 2 .
Definition 1.
A pair ( g * , G * ) in the space W = W 1 , 2 ( Ω ) × ( W 1 , 2 ( Ω ) ) 2 is called a Nash equilibrium for the game involving the two energies E 1 ( · ) and E 2 ( · ) , defined on W, if
E 1 ( g * , G * ) E 1 ( g , G * ) , g W 1 , 2 ( Ω ) , E 2 ( g * , G * ) E 2 ( g * , G ) , G ( W 1 , 2 ( Ω ) ) 2 .
One could consider only the single energy E 1 + E 2 to be optimised; however, for the theoretical analysis, the ellipticity of the sum energy is not guaranteed because of the coupling term between g and G . Hence, the existence of minimisers is not straightforward. However, we emphasise that in the game formulation, the energies E 1 ( · , G ) and E 2 ( g , · ) are partially elliptic, i.e., with respect to the variables g and G , respectively. This is a very important property which eases the proof of the existence of Nash equilibrium.
Proposition 1.
There exists a unique Nash equilibrium ( g * , G * ) W 1 , 2 ( Ω ) × ( W 1 , 2 ( Ω ) ) 2 for the two-player game involving the costs functional E 1 ( · , · ) and E 2 ( · , · ) in (6) and (7).
Proof of Proposition 1.
Since
  • E 1 ( · , G ) is partially strict convex, partially elliptic and weakly lower semi-continuous with respect to variable g,
  • E 2 ( g , · ) is partially strict convex, partially elliptic and weakly lower semi-continuous with respect to variable G ,
the proof is a straightforward and direct application of the the Nash theorem [22]. □

4. Numerical Algorithms and Implementation

In this section, we detail the numerical algorithm to solve our game model and show how we utilise the outputs to obtain a segmentation result.

4.1. Stage One: Solution of the Main Model Using ADMM

The discretised version of our two-player game model (6) and (7) is given as follows:
min g λ 1 2 | | f g | | 2 2 + μ 1 2 | | g | | 2 2 + | | g | | 1 + ξ 1 2 | | G g | | 2 2 , min G λ 2 2 | | f G | | 2 2 + μ 2 2 | | G | | 2 2 + | | G | | 1 + ξ 2 2 | | G g | | 2 2 ,
where | | g | | 1 = i Ω ( x g ) i 2 + ( y g ) i 2 and | | G | | 1 = i Ω | ( x G x ) i + ( y G y ) i | . The gradient operator = ( x , y ) is discretised using backwards differences with zero Neumann boundary conditions.
We aim to solve the coupled problem using the split-Bregman variant of the alternating direction method of multipliers (ADMM) [23], which is commonly used for problems containing L 1 regularisation. In order to do this, we introduce a new variable into each sub-problem:
min g , v λ 1 2 | | f g | | 2 2 + μ 1 2 | | g | | 2 2 + | | v | | 1 + ξ 1 2 | | G g | | 2 2 , such that v = g , min G , w λ 2 2 | | f G | | 2 2 + μ 2 2 | | G | | 2 2 + | | w | | 1 + ξ 2 2 | | G g | | 2 2 , such that w = G .
Applying split-Bregman to enforce the constraints gives us:
min g , v λ 1 2 | | f g | | 2 2 + μ 1 2 | | g | | 2 2 + | | v | | 1 + ξ 1 2 | | G g | | 2 2 + ρ 1 2 | | v g b 1 | | 2 2 , min G , w λ 2 2 | | f G | | 2 2 + μ 2 2 | | G | | 2 2 + | | w | | 1 + ξ 2 2 | | G g | | 2 2 + ρ 2 2 | | w G b 2 | | 2 2 .
We detail briefly how to solve each of the sub-problems:
g sub-problem: We aim to solve the minimisation problem for fixed v ( k ) , G ( k ) , b 1 ( k ) :
g ( k + 1 ) = arg min g λ 1 2 | | f g | | 2 2 + μ 1 2 | | g | | 2 2 + ξ 1 2 | | G ( k ) g | | 2 2 + ρ 1 2 | | v ^ ( k ) g | | 2 2 ,
where v ^ ( k ) = v ( k ) b 1 ( k ) , which amounts to solving the following equation for g:
λ 1 + ( μ 1 + ξ 1 + ρ 1 ) 2 g = λ 1 f + ξ 1 T G ( k ) + ρ 1 T ( v ( k ) b 1 ( k ) ) .
This can be solved by using discrete Fourier transforms F :
g ( k + 1 ) = F 1 F ( L ) M
where
L = λ 1 f + ξ 1 T G ( k ) + ρ 1 T ( v ( k ) b 1 ( k ) ) , M = λ 1 + ( μ 1 + ξ 1 + ρ 1 ) F ( 2 ) .
v sub-problem: We aim to solve this minimisation problem for fixed g ( k + 1 ) , b 1 ( k ) :
v ( k + 1 ) = arg min v | | v | | 1 + ρ 1 2 | | v g ( k + 1 ) b 1 ( k ) | | 2 2 ,
which is solved analytically by a generalised shrinkage formula:
v x ( k + 1 ) = max s ( k ) 1 ρ 1 , 0 s x ( k ) s ( k ) , v y ( k + 1 ) = max s ( k ) 1 ρ 1 , 0 s y ( k ) s ( k ) ,
where s 1 ( k ) = x g ( k + 1 ) + ( b 1 ( k ) ) x , s 2 ( k ) = y g ( k + 1 ) + ( b 1 ( k ) ) y and
s ( k ) = ( s 1 ( k ) ) 2 + ( s 2 ( k ) ) 2 .
The associated Bregman update is:
( b 1 ( k + 1 ) ) x = ( b 1 ( k ) ) x + x g ( k ) v x ( k + 1 ) , ( b 1 ( k + 1 ) ) y = ( b 1 ( k ) ) y + y g ( k ) v y ( k + 1 ) .
G sub-problem: We aim to solve the minimisation problem for fixed g ( k + 1 ) , w ( k ) , b 2 ( k ) :
G ( k + 1 ) = arg min G { λ 2 2 | | f G | | 2 2 + μ 2 2 | | G | | 2 2 +
ξ 2 2 | | G g ( k + 1 ) | | 2 2 + ρ 2 2 | | w ( k ) G b 2 ( k ) | | 2 2 } ,
whose solution is defined by the following:
( λ 2 + ξ 2 + ( μ 2 + ρ 2 ) 2 ) G = λ 2 f + ξ 2 g ( k + 1 ) + ρ 2 T ( w ( k ) b 2 ( k ) ) .
To find the solution G , we apply discrete Fourier transforms F :
G x ( k + 1 ) = F F ( N 1 ) P , G y ( k + 1 ) = F F ( N 2 ) P ,
where P = λ 2 + ξ 2 + ( μ 2 + ρ 2 ) F ( 2 ) , N 1 = x f + ξ 2 x g ( k + 1 ) + ρ 2 x T ( w ( k ) b 2 ( k ) ) , and N 2 = y f + ξ 2 y g ( k + 1 ) + ρ 2 y T ( w ( k ) b 2 ( k ) ) .
w sub-problem: We aim to solve the minimisation problem for fixed G ( k + 1 ) , b 2 ( k ) :
w ( k + 1 ) = arg min w | | w | | 1 + ρ 2 2 | | w G ( k + 1 ) b 2 ( k ) | | 2 2 ,
which, similar to (9), is solved by using a shrinkage formula:
w ( k + 1 ) = max | r ( k ) | 1 ρ 2 , 0 r ( k ) | r ( k ) | ,
where r ( k ) = G ( k + 1 ) + b 2 ( k ) .

4.2. Stage Two: Segmentation of f by Thresholding g

In order to acquire a segmentation result for f, we take the minimiser g from stage one and threshold it according to some suitably defined threshold parameter(s). As in [12], the advantage of this method is that changing the threshold value(s) does not require the re-computation of the optimisation done in stage one.
There are two strategies that can be employed to define the threshold(s). The first is to use the k-means algorithm, which is an automatic method that partitions a given input into K clusters, for K 2 . The second is to define the threshold value(s) manually, which generally provides better results. As the threshold values are applied after optimisation, a wide range of values can easily be tried and the best selected. In our experiments, we use manual threshold values for two-phase segmentation, whereas for multiphase segmentation with multiple threshold values, we use k-means to simplify the process.

5. Numerical Results

In this section, we display some examples of the performance of our model and compare it with a number of models, namely:
  • CRCV: Convex relaxed Chan–Vese model [5];
  • CCZ: The two-stage convex variant of the Mumford–Shah model by Cai et al. [12] given in (4);
  • CNC: The convex non-convex segmentation by Chan et al. [24];
  • T-ROF: The T-ROF model by Cai et al. [25];
and also a deep learning model.
We first show some visual comparisons, where noise is added to the original image, and then later do a quantitative analysis on a dataset. Note that all the models above (and ours) except for the CRCV model is capable of multiphase segmentation, whereas the CRCV model (in the Chan–Vese framework) is only capable of two-phase segmentation. For this reason, in the experiments, we only include the CRCV model in two-phase examples.

5.1. Qualitative Results

In Figure 1, we show an image from an ultrasound. We add additive Gaussian noise with mean 0 and standard deviation 10. We display the outputs of all the competing models, the segmentation result overlaid on the original image, and for all but the CRCV show the segmentation result after thresholding (as the segmentation result after thresholding is the binary output shown first). We see that the segmentation result from our model is better at segmenting the object in the image, noticing that our segmentation effectively segments the “tail” part at the top of the object, whereas the CCZ model fails to segment it well. The CRCV and CNC models segment the tail but fail to remove the noise. We note that the T-ROF model is the best competing model but does not quite segment all the tail.
Similarly, in Figure 2, we show another two-phase segmentation example, where we have the clean image but add Gaussian noise with mean 0 and standard deviation 25. It is clear that none of the competing models are as good as ours. Our result manages to preserve more detail in general, notably at the strand at the top, and the curved structure at the bottom of the image, without being susceptible to the noise.
In Figure 3, Figure 4 and Figure 5, we show some examples of multiphase segmentation on MRI images of the brain. In all cases, we add Gaussian noise with mean 0 and standard deviation 17 and run the noisy image as input to both for all models but the CRCV model (as this is a two-phase model only). The output is then given as input to the k-means algorithm with K = 4 . We show the clustering output in the final column of the relevant figures. We see that the segmentation result of our model is better at finding some of the finer edges; for example, the white matter segmentation from our model is in general more detailed than the segmentation from the competing ones.

5.2. Quantitative Analysis

To assess our method quantitatively, we run our model on 20 images in the Digital Retinal Images for Vessel Extraction (DRIVE) dataset (https://drive.grand-challenge.org/ accessed 25 October 2021). We use the manual segmentation image as the clean image and add additive Gaussian noise with mean 0 and standard deviation 100 to use as the input image, as shown in Figure 6, Figure 7, Figure 8 and Figure 9a,b respectively. We display the output of the competing models and our model here as well as a deep learning model (abbreviated as DL). We trained a U-Net [6] network on 15 of the images (and used the other five as validation set), where the noisy image served as input and we trained with binary cross-entropy loss function to match with the clean image. The results are good; however, we lack a large dataset to provide the impressive result that deep learning approaches usually provide.
Figure 6, Figure 7, Figure 8 and Figure 9 are four examples on the given dataset; however, we run on the 20 images available to provide some quantitative analysis. We use the DICE coefficient and the JACCARD similarity coefficient as quantitative measures to evaluate the performance of segmentation results. Given a binary segmentation result Σ from a model and ground truth segmentation G T , the DICE coefficient is given as:
D I C E ( Σ , G T ) = 2 | Σ G T | | Σ | + | G T | .
The JACCARD similarity coefficient is given as:
J A C C A R D ( Σ , G T ) = | Σ G T | | Σ | + | G T | | Σ G T | .
In Table 1, we show the mean and standard deviation values of the DICE and JACCARD scores on the dataset. We see clearly that our model is more effective than the Cai model from these results. We note that the numerical values provided for the DL method are run on all 20 images in the dataset; however, the DL was trained on 15 of these images. This is somewhat of an unfair comparison; however, we see that the numerical values for our approach are still larger than the values for the DL approach despite this. Figure 10 shows the boxplots of quantitative results on the data, for further visualisation.

6. Conclusions

In this paper, we have developed a convex relaxed game formulation of the less well-known Blake–Zisserman model in order to segment images with low contrast and strong noise. The advantages of the game formulation are that the existence of Nash equilibrium can be proved and there is less dependence on parameters for each sub-problem, i.e., parameters of each sub-problem do not rely on each other, and so can be tuned appropriately and separately. The game model was implemented using a fast split-Bregman algorithm, and numerical experiments show improvements in segmentation results over competing models, especially over the well-known Mumford–Shah type methods for low-contrast images.

Author Contributions

Conceptualisation, A.T. and K.C.; methodology, A.T.; software, L.B.; validation, L.B.; formal analysis, A.T.; investigation, L.B.; writing—original draft preparation, L.B., A.T. and K.C.; writing—review and editing, L.B., A.T. and K.C.; visualisation, L.B.; supervision, K.C.; project administration, K.C.; funding acquisition, K.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by EPSRC grant number EP/N014499/1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Digital Retinal Images for Vessel Extraction (DRIVE) dataset can be found at https://drive.grand-challenge.org/. Data accessed: 25 October 2021.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MSMumford–Shah
BZBlake–Zisserman
ADMMAlternating direction method of multipliers

References

  1. Kass, M.; Witkin, A.; Terzopoulos, D. Snakes: Active contour models. Int. J. Comput. Vis. 1988, 1, 321–331. [Google Scholar] [CrossRef]
  2. Caselles, V.; Kimmel, R.; Sapiro, G. Geodesic active contours. Int. J. Comput. Vis. 1997, 22, 61–79. [Google Scholar] [CrossRef]
  3. Mumford, D.B.; Shah, J. Optimal approximations by piecewise smooth functions and associated variational problems. Commun. Pure Appl. Math. 1989, 42, 577–685. [Google Scholar] [CrossRef] [Green Version]
  4. Chan, T.F.; Vese, L.A. Active contours without edges. IEEE Trans. Image Process. 2001, 10, 266–277. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Chan, T.F.; Esedoglu, S.; Nikolova, M. Algorithms for finding global minimizers of image segmentation and denoising models. SIAM J. Appl. Math. 2006, 66, 1632–1648. [Google Scholar] [CrossRef]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  7. He, K.; Zhang, X.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  8. Hesamian, M.H.; Jia, W.J.; He, X.J.; Kennedy, P. Deep Learning Techniques for Medical Image Segmentation: Achievements and Challenges. J. Digit. Imaging 2019, 32, 582–596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Burrows, L.; Chen, K.; Torella, F. On New Convolutional Neural Network Based Algorithms for Selective Segmentation of Images. In Proceedings of the MIUA 2020 Proceedings, Oxford, UK, 15–17 July 2020; Communications in Computer and Information Science book Series (CCIS). Springer: Berlin/Heidelberg, Germany, 2020; Volume 1048, pp. 93–104. [Google Scholar]
  10. Blake, A.; Zisserman, A. Visual Reconstruction; MIT Press: Cambridge, MA, USA, 1987. [Google Scholar]
  11. Ambrosio, L.; Tortorelli, V.M. Approximation of functional depending on jumps by elliptic functional via t-convergence. Commun. Pure Appl. Math. 1990, 43, 999–1036. [Google Scholar] [CrossRef]
  12. Cai, X.; Chan, R.; Zeng, T. A two-stage image segmentation method using a convex variant of the Mumford–Shah model and thresholding. SIAM J. Imaging Sci. 2013, 6, 368–390. [Google Scholar] [CrossRef]
  13. Burrows, L.; Guo, W.; Chen, K.; Torella, F. Reproducible kernel Hilbert space based global and local image segmentation. Inverse Probl. Imaging 2021, 15, 1. [Google Scholar] [CrossRef]
  14. Theljani, A.; Belhachmi, Z. A discrete approximation of Blake and Zisserman energy in image denoising with optimal choice of regularization parameters. Math. Methods Appl. Sci. 2021, 44, 3857–3871. [Google Scholar] [CrossRef]
  15. Zanetti, M.; Ruggiero, V.; Miranda, M., Jr. Numerical minimization of a second-order functional for image segmentation. Commun. Nonlinear Sci. Numer. Simul. 2016, 36, 528–548. [Google Scholar] [CrossRef]
  16. Theljani, A.; Habbal, A.; Kallel, M.; Chen, K. Game Theory and Its Applications in Imaging and Vision; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  17. Huang, Y.; Ng, M.K.; Wen, Y.W. A fast total variation minimization method for image restoration. Multiscale Modeling Simul. 2008, 7, 774–795. [Google Scholar] [CrossRef]
  18. Habbal, A.; Petersson, J.; Thellner, M. Multidisciplinary topology optimization solved as a Nash game. Int. J. Numer. Methods Eng. 2004, 61, 949–963. [Google Scholar] [CrossRef]
  19. Kallel, M.; Aboulaich, R.; Habbal, A.; Moakher, M. A Nash-game approach to joint image restoration and segmentation. Appl. Math. Model. 2014, 38, 3038–3053. [Google Scholar] [CrossRef]
  20. Kallel, M.; Moakher, M.; Theljani, A. The Cauchy problem for a nonlinear elliptic equation: Nash-game approach and application to image inpainting. Inverse Probl. Imaging 2015, 9, 853. [Google Scholar]
  21. Roy, S.; Borzì, A.; Habbal, A. Pedestrian motion modelled by Fokker–Planck Nash games. R. Soc. Open Sci. 2017, 4, 170648. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Nash, J. Non-Cooperative Games. Ann. Math. 1951, 54, 286–295. [Google Scholar] [CrossRef]
  23. Goldstein, T.; Osher, S. The split Bregman method for L1-regularized problems. SIAM J. Imaging Sci. 2009, 2, 323–343. [Google Scholar] [CrossRef]
  24. Chan, R.; Lanza, A.; Morigi, S.; Sgallari, F. Convex non-convex image segmentation. Numer. Math. 2018, 138, 635–680. [Google Scholar] [CrossRef]
  25. Cai, X.; Chan, R.; Schönlieb, C.B.; Steidl, G.; Zeng, T. Linkage Between Piecewise Constant Mumford–Shah Model and Rudin–Osher–Fatemi Model and Its Virtue in Image Segmentation. SIAM J. Sci. Comput. 2019, 41, B1310–B1340. [Google Scholar] [CrossRef]
Figure 1. Results from an ultrasound image: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CRCV contour. (e) Output of CCZ. (f) CCZ after thresholding. (g) CCZ contour. (h) Output of CNC. (i) CNC after thresholding. (j) CNC contour. (k) Output of T-ROF. (l) T-ROF after thresholding. (m) T-ROF contour. (n) Output g of our model. (o) Output G x of our model. (p) Output G y of our model. (q) Ours after thresholding. (r) Our contour.
Figure 1. Results from an ultrasound image: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CRCV contour. (e) Output of CCZ. (f) CCZ after thresholding. (g) CCZ contour. (h) Output of CNC. (i) CNC after thresholding. (j) CNC contour. (k) Output of T-ROF. (l) T-ROF after thresholding. (m) T-ROF contour. (n) Output g of our model. (o) Output G x of our model. (p) Output G y of our model. (q) Ours after thresholding. (r) Our contour.
Jimaging 07 00228 g001
Figure 2. Results from a blood vessel image: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CRCV contour. (e) Output of CCZ. (f) CCZ after thresholding. (g) CCZ contour. (h) Output of CNC. (i) CNC after thresholding. (j) CNC contour. (k) Output of T-ROF. (l) T-ROF after thresholding. (m) T-ROF contour. (n) Output g of our model. (o) Output G x of our model. (p) Output G y of our model. (q) Ours after thresholding. (r) Our contour.
Figure 2. Results from a blood vessel image: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CRCV contour. (e) Output of CCZ. (f) CCZ after thresholding. (g) CCZ contour. (h) Output of CNC. (i) CNC after thresholding. (j) CNC contour. (k) Output of T-ROF. (l) T-ROF after thresholding. (m) T-ROF contour. (n) Output g of our model. (o) Output G x of our model. (p) Output G y of our model. (q) Ours after thresholding. (r) Our contour.
Jimaging 07 00228 g002
Figure 3. MRI segmentation: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of CCZ. (d) CCZ after thresholding. (e) Output of CNC. (f) CNC after thresholding. (g) Output of T-ROF. (h) T-ROF after thresholding. (i) Output g of our model. (j) Output G x of our model. (k) Output G y of our model. (l) Ours after thresholding.
Figure 3. MRI segmentation: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of CCZ. (d) CCZ after thresholding. (e) Output of CNC. (f) CNC after thresholding. (g) Output of T-ROF. (h) T-ROF after thresholding. (i) Output g of our model. (j) Output G x of our model. (k) Output G y of our model. (l) Ours after thresholding.
Jimaging 07 00228 g003
Figure 4. MRI segmentation: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of CCZ. (d) CCZ after thresholding. (e) Output of CNC. (f) CNC after thresholding. (g) Output of T-ROF. (h) T-ROF after thresholding. (i) Output g of our model. (j) Output G x of our model. (k) Output G y of our model. (l) Ours after thresholding.
Figure 4. MRI segmentation: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of CCZ. (d) CCZ after thresholding. (e) Output of CNC. (f) CNC after thresholding. (g) Output of T-ROF. (h) T-ROF after thresholding. (i) Output g of our model. (j) Output G x of our model. (k) Output G y of our model. (l) Ours after thresholding.
Jimaging 07 00228 g004aJimaging 07 00228 g004b
Figure 5. MRI segmentation: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of CCZ. (d) CCZ after thresholding. (e) Output of CNC. (f) CNC after thresholding. (g) Output of T-ROF. (h) T-ROF after thresholding. (i) Output g of our model. (j) Output G x of our model. (k) Output G y of our model. (l) Ours after thresholding.
Figure 5. MRI segmentation: (a) Clean image. (b) Noisy image used as input to the models. (c) Output of CCZ. (d) CCZ after thresholding. (e) Output of CNC. (f) CNC after thresholding. (g) Output of T-ROF. (h) T-ROF after thresholding. (i) Output g of our model. (j) Output G x of our model. (k) Output G y of our model. (l) Ours after thresholding.
Jimaging 07 00228 g005aJimaging 07 00228 g005b
Figure 6. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Figure 6. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Jimaging 07 00228 g006
Figure 7. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Figure 7. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Jimaging 07 00228 g007
Figure 8. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Figure 8. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Jimaging 07 00228 g008
Figure 9. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Figure 9. (a) Clean image. (b) Noisy image used as input to the models. (c) Output of the CRCV model. (d) CCZ after thresholding. (e) CNC after thresholding. (f) T-ROF after thresholding. (g) DL Output. (h) Output g of our model. (i) Output G x of our model. (j) Output G y of our model. (k) Ours after thresholding. (l) Our contour.
Jimaging 07 00228 g009
Figure 10. Comparison of six methods (CRCV, CCZ, CNCS, TROF, DL, Ours): box plots of the quantitative results for DICE (a) and JACCARD (b) scores.
Figure 10. Comparison of six methods (CRCV, CCZ, CNCS, TROF, DL, Ours): box plots of the quantitative results for DICE (a) and JACCARD (b) scores.
Jimaging 07 00228 g010
Table 1. Quantitative results from images in the DRIVE dataset. Here, we show the two methods evaluated on 20 images and display the mean and standard deviations of both the DICE coefficient and JACCARD score. Note that the DL method was trained on 15 of these 20 images.
Table 1. Quantitative results from images in the DRIVE dataset. Here, we show the two methods evaluated on 20 images and display the mean and standard deviations of both the DICE coefficient and JACCARD score. Note that the DL method was trained on 15 of these 20 images.
DICEJACCARD
μ σ μ σ
CRCV0.7270.02910.5730.0358
CCZ0.9140.01390.8430.0236
CNC0.9390.01310.8840.0233
T-ROF0.9320.00830.8720.0145
DL0.9460.00910.8980.0163
Ours0.9500.00730.9050.0133
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Burrows, L.; Theljani, A.; Chen, K. On a Variational and Convex Model of the Blake–Zisserman Type for Segmentation of Low-Contrast and Piecewise Smooth Images. J. Imaging 2021, 7, 228. https://doi.org/10.3390/jimaging7110228

AMA Style

Burrows L, Theljani A, Chen K. On a Variational and Convex Model of the Blake–Zisserman Type for Segmentation of Low-Contrast and Piecewise Smooth Images. Journal of Imaging. 2021; 7(11):228. https://doi.org/10.3390/jimaging7110228

Chicago/Turabian Style

Burrows, Liam, Anis Theljani, and Ke Chen. 2021. "On a Variational and Convex Model of the Blake–Zisserman Type for Segmentation of Low-Contrast and Piecewise Smooth Images" Journal of Imaging 7, no. 11: 228. https://doi.org/10.3390/jimaging7110228

APA Style

Burrows, L., Theljani, A., & Chen, K. (2021). On a Variational and Convex Model of the Blake–Zisserman Type for Segmentation of Low-Contrast and Piecewise Smooth Images. Journal of Imaging, 7(11), 228. https://doi.org/10.3390/jimaging7110228

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop