Next Article in Journal
An Image Hashing Algorithm for Authentication with Multi-Attack Reference Generation and Adaptive Thresholding
Next Article in Special Issue
CYK Parsing over Distributed Representations
Previous Article in Journal
A Linear-Time Algorithm for the Isometric Reconciliation of Unrooted Trees
Previous Article in Special Issue
A Brief Survey of Fixed-Parameter Parallelism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatially Adaptive Regularization in Image Segmentation

by
Laura Antonelli
1,†,
Valentina De Simone
2,† and
Daniela di Serafino
2,*,†,‡
1
Institute for High Performance Computing and Networking (ICAR), Italian National Research Council (CNR), via P. Castellino 111, I-80131 Naples, Italy
2
Department of Mathematics and Physics, University of Campania “Luigi Vanvitelli”, viale A. Lincoln 5, I-81100 Caserta, Italy
*
Author to whom correspondence should be addressed.
All the authors contributed equally to this work.
Current address: Department of Mathematics and Physics, University of Campania “Luigi Vanvitelli”, viale A. Lincoln 5, I-81100 Caserta, Italy.
Algorithms 2020, 13(9), 226; https://doi.org/10.3390/a13090226
Submission received: 31 July 2020 / Revised: 26 August 2020 / Accepted: 6 September 2020 / Published: 8 September 2020
(This article belongs to the Special Issue 2020 Selected Papers from Algorithms Editorial Board Members)

Abstract

:
We present a total-variation-regularized image segmentation model that uses local regularization parameters to take into account spatial image information. We propose some techniques for defining those parameters, based on the cartoon-texture decomposition of the given image, on the mean and median filters, and on a thresholding technique, with the aim of preventing excessive regularization in piecewise-constant or smooth regions and preserving spatial features in nonsmooth regions. Our model is obtained by modifying a well-known image segmentation model that was developed by T. Chan, S. Esedoḡlu, and M. Nikolova. We solve the modified model by an alternating minimization method using split Bregman iterations. Numerical experiments show the effectiveness of our approach.

1. Introduction

Image segmentation is a fundamental problem in image processing and computer vision. Its goal is to divide the given image into regions that represent different objects in the scene, which can be identified taking into account different features. The research on image segmentation has made several advances in the last decades and various approaches have been developed to satisfy the requirements coming from different applications. In this work, we focus on variational models for image segmentation, which have been widely investigated, proving to be very effective on different images—see, e.g., [1,2,3,4]. Other recent approaches to image segmentation include learning-based methods, which often exploit deep-learning techniques [5,6,7,8]. However, in this case, a large amount of data must be available to train learning networks, thus making those approaches impractical in some  applications.
Roughly speaking, when a variational approach is used, the segmentation may be obtained by minimizing a cost functional that linearly combines regularization and data fidelity terms:
min u Ω E ( u ) F ( u ) + λ G ( u ) ,
where u : Ω R 2 R provides a representation of the segmentation and λ > 0 is a parameter that controls the weight of the fidelity term G versus the regularization term F. A widely-used segmentation model is the two-phase partitioning model that was introduced by Chan, Esedoḡlu, and Nikolova [3], which we refer to as the CEN model:
minimize u , c 1 , c 2 Ω | u | d x + λ Ω ( c 1 u ¯ ( x ) ) 2 u ( x ) + ( c 2 u ¯ ( x ) ) 2 ( 1 u ( x ) ) d x , s . t . 0 u 1 , c 1 , c 2 > 0 .
Here, u ¯ denotes the image to be segmented, which is assumed to take its values in [ 0 , 1 ] . The CEN model allows us to obtain one of the two domains defining the segmentation, Σ and Ω \ Σ , by setting
Σ = x : u ( x ) > α for   a . e . α ( 0 , 1 ) ,
where u is the solution of problem (2). Note that (2) is the result of a suitable relaxation of the Chan–Vese model [2] leading to a convex formulation of that problem for any given  ( c 1 , c 2 ) .
In this work, we start from a discrete version of the CEN model. Let
Θ = ( i , j ) : 0 i m 1 , 0 j n 1
be a discretization of Ω consisting of an m × n grid of pixels and let
| x u | i , j = | δ x + u | i , j , | y u | i , j = | δ y + u | i , j
where δ x + and δ y + are the forward finite-difference operators in the x- and y-directions, with unit spacing, and the values u i , j with indices outside  Θ are defined by replication. We consider the following discrete version of (2) with anisotropic discrete total variation (TV):
minimize u , c 1 , c 2 i , j ( | x u | i , j + | y u | i , j ) + λ i , j c 1 u ¯ i , j 2 u i , j + c 2 u ¯ i , j 2 ( 1 u i , j ) , s . t . 0 u i , j 1 , c 1 , c 2 > 0 .
After the minimization problem (3) is solved, the segmentation is obtained by taking
Σ = ( i , j ) Θ : u i , j > α ,
for some α ( 0 , 1 ) . Problem (3) is usually solved by alternating the minimization with respect to u and the minimization with respect to c 1 and c 2 . In the sequel, we denote E ( u , c 1 , c 2 ) the objective function in (3), and F ( u ) and G ( u , c 1 , c 2 ) its discrete regularization and fidelity terms, respectively.
The selection of a parameter λ able to balance F ( u ) and G ( u , c 1 , c 2 ) and to produce a meaningful solution is a critical issue. Too large values of λ may produce oversegmentation, while too small values of λ may produce undersegmentation [9]. Furthermore, a constant value of λ may not be suitable for the whole image, i.e., different regions of the image may need different values. In recent years, spatially adaptive techniques have been proposed, focusing on local information, such as gradient, curvature, texture, and noise estimation cues—see, e.g., [10]. Space-variant regularization has been also widely investigated in the context of image restoration, using TV and its generalizations—see, e.g., [11,12,13,14,15].
In this work, we propose some techniques for setting the parameter λ in an adaptive way based on spatial information, in order to prevent excessive regularization of smooth regions while preserving spatial features in nonsmooth areas. Our techniques are based on the so-called cartoon-texture decomposition of the given image, on the mean and median filters, and on a thresholding technique. The resulting locally adaptive segmentation model can be solved either by smoothing the discrete TV term—see, e.g., [16,17]—and applying optimization solvers for differentiable problems, such as spectral gradient methods [18,19,20,21], or by using directly optimization solvers for nondifferentiable problems, such as Bregman, proximal and ADMM methods [22,23,24,25,26,27,28]. In this work, we use an alternating minimization procedure exploiting the split Bregman (SB) method proposed in [24]. The results of numerical experiments on images with different characteristics show the effectiveness of our approach and the advantages coming from using local regularization.
The rest of this paper is organized, as follows. In Section 2 we propose our spatially adaptive techniques. In Section 3, we describe the solution of the segmentation model using those techniques by the aforementioned SB-based alternating minimization method. In Section 4 we discuss the results obtained with our approaches on several test images, also performing a comparison with the CEN model and a segmentation model developed for images with texture. The results show the effectiveness of our approach and the advantages coming from the use of local regularization. Some conclusions are provided in Section 5.

2. Defining Local Regularization by Exploiting Spatial Information

The regularization parameter λ plays an important role in the segmentation process, because it controls the tradeoff between data fidelity and regularization. In general, the smaller the parameter in (3) the smoother the image content, i.e., image details are flattened or blurred. Conversely, the larger the parameter the more the enhancement of image details and, hence, noise may be retained or amplified. Therefore, λ should be selected according to local spatial information. A small value of λ should be used in the smooth regions of the image to suppress noise, while a large value of λ should be considered to preserve spatial features in the nonsmooth regions. In other words, a matrix Λ = ( λ i , j ) should be associated with the image, where λ i , j weighs pixel ( i , j ) , as follows:
minimize u , c 1 , c 2 i , j ( | x u | i , j + | y u | i , j ) + i , j λ i , j c 1 u ¯ i , j 2 u i , j + c 2 u ¯ i , j 2 ( 1 u i , j ) , s . t . 0 u i , j 1 , c 1 , c 2 > 0 .
Furthermore, in order to avoid oversegmentation or undersegmentation, it is convenient to fix a minimum and a maximum value for the entries of Λ , so as to drive the level of regularization in a reasonable range, depending on the image to be segmented.
We define Λ as a function of the image u ¯ to be segmented:
f : u ¯ i , j λ i , j [ λ m i n , λ m a x ] ,
where 0 < λ m i n < λ m a x < . We propose three choices of f, detailed in the next subsections.
We note that problem (5) still has a unique solution for any fixed ( c 1 , c 2 ) , as a consequence of the next proposition.
Proposition 1.
For any fixed ( c 1 , c 2 ) , problem (5) is a convex problem.
Proof. 
Because the CEN model is convex for any fixed ( c 1 , c 2 ) , the thesis immediately follows from the fact that the parameters λ i , j are constant with respect to u.  □

2.1. Regularization Based on the Cartoon-Texture Decomposition

We define f ( u ¯ i , j ) by using the Cartoon-Texture Decomposition (CTD) of the image discussed in [29]. CTD splits any image u into the sum of two images, w and v, such that w represents the cartoon or geometric (piecewise-smooth) component of u, while v represents the oscillatory or textured component, i.e., v contains essentially textures, oscillating patterns, fine details, and noise. The algorithm for computing CTD acts as described next. For each image pixel, we decide whether it belongs to the cartoon or the textural part by computing a local indicator associated with an image window around the pixel. The main feature of a textured region is its high TV, which decreases very fast under low-pass filtering. This leads to the following definition of local total variation (LTV) at a pixel ( i , j ) :
L T V σ ( u ) i , j = ( L σ | u | i , j ) ,
where L σ is a low-pass filter, σ is a scale parameter, | u | i , j = ( x u ) i , j 2 + ( y u ) i , j 2 , and ∗ denotes the convolution product. The relative reduction rate of LTV,
( ρ σ ) i , j = L T V σ ( u ) i , j L T V σ ( L σ u ) i , j L T V σ ( u ) i , j ,
gives the local oscillatory behavior of the image. A value of ( ρ σ ) i , j close to 0 means that there is little relative reduction of LTV by the low-pass filter, thus the pixel ( i , j ) belongs to the cartoon. Conversely, ( ρ σ ) i , j close to 1 indicates that the relative reduction is large and, hence, the pixel belongs to a textured  region.
We use (8) for defining the weights λ i , j . The basic idea is that a large regularization parameter is needed if the pixel ( i , j ) belongs to the cartoon, while the parameter must be reduced in texture regions. Therefore, we set the function f in (6) as
f ( u ¯ i , j ) f CTD ( u ¯ i , j ) = max λ m i n λ m a x , 1 ( ρ σ ) i , j λ m a x
and ( ρ σ ) i , j is defined by using the given image u ¯ . Following [30], we set L σ equal to the Gaussian filter.

2.2. Regularization Based on the Mean and Median Filters

We define a technique that is based on spatial filters that are commonly used to enhance low-frequency details or to preserve edges [31,32]. More precisely, we combine the mean and median filters; the former aims at identifying smooth regions, where the regularization parameter can take small values, the latter aims at identifying edges, where the parameter must remain large. Mean filtering is a simple and easy-to-implement method for smoothing images, i.e., for reducing the intensity variation between a pixel and its neighbors. It also removes high-frequencies components due to the noise and the edges in the image, so the mean filter is a low-pass filter. The median filter preserves edges and useful details in the image.
Based on these considerations, we define a weight function, as follows:
ω i , j = | u ¯ i , j ( L h 1 u ¯ ) i , j | if   | ( L h 1 u ¯ ) i , j ( M h 2 u ¯ ) i , j | < t , 1 otherwise .
where h 1 is the window size of the mean filter L h 1 , h 2 is the window size of the median filter M h 2 , and t is a threshold value acting as a cutoff between the two filters. Note that 0 ω i , j 1 and the pixels in homogeneous regions have ω i , j close to 1. The function f in (6) is set as
f ( u ¯ i , j ) f MM ( u ¯ i , j ) = max λ m i n λ m a x , 1 ω i , j λ m a x ,
where MM stands for “Mean and Median”.

2.3. Regularization Based on Thresholding

This approach implicitly exploits the definition of Σ in (4) in order to set λ i , j . The idea is to use large values of λ i , j when u i , j is close to 1 and small values when u i , j is close to 0. Therefore, the parameters λ i , j are not defined in terms of the given image u ¯ only. If the function u identifying the segmentation were known a priori, we could define λ i , j = f ( u i , j ) as follows:
f ( u i , j ) f THR ( u i , j ) = 10 η i , j [ λ m i n , λ m a x ] ,
where
η i , j = e m a x ( 1 u i , j ) ( e m a x e m i n ) ,
e m a x = log 10 λ m a x and e m i n = log 10 λ m i n .
Because the function u must be computed by minimizing (3) and this is done by using an iterative procedure, we decided to update λ i , j at each iteration, while using the current value of u i , j . On the other hand, in this case evaluating f is computationally cheaper than in the previous approaches, which apply two-dimensional convolution operators; thus, the cost of the iterative update of λ i , j is practically negligible.

3. Solution by Split Bregman Iterations

As mentioned in Section 1, we use an alternating minimization method to solve problem (5). Given u i , j k 1 , by imposing the first-order optimality conditions with respect to c 1 and c 2 , we get
c 1 k = i , j λ i , j u ¯ i , j u i , j k 1 i , j λ i , j u i , j k 1 , c 2 k = i , j λ i , j u ¯ i , j ( 1 u i , j k 1 ) i , j λ i , j ( 1 u i , j k 1 ) .
For the solution of (5) with respect to u, we use the split Bregman (SB) method [24]. Let
G l o c ( u , c 1 , c 2 ) = i , j λ i , j c 1 u ¯ i , j 2 c 2 u ¯ i , j 2 u i , j = i , j r i , j u i , j ,
where
r i , j = λ i , j c 1 u ¯ i , j 2 c 2 u ¯ i , j 2 .
Following [33], we reformulate the minimization problem, as follows:
minimize u , d x , d y d x 1 + d y 1 + G l o c ( u , c 1 , c 2 ) , s . t . 0 u i , j 1 , d x = x u , d y = y u .
Given c 1 k and c 2 k , the SB method applied to (15) reads:
u k = argmin 0 u i , j 1 G l o c ( u , c 1 k , c 2 k ) + μ 2 d x k 1 x u b x k 1 2 2 + μ 2 d y k 1 y u b y k 1 2 2 , d x k = argmin d x d x 1 + μ 2 d x x u k b x k 1 2 2 , d x k = argmin d y d y 1 + μ 2 d y y u k b y k 1 2 2 , b x k = b x k 1 + μ ( x u k d x k ) , b y k = b y k 1 + μ ( x u k d y k ) ,
where μ > 0 .
Closed-form solutions of the minimization problems with respect to d x and d y can be computed using the soft-thresholding operator:
d x k = S x u k + b x k , 1 μ , d y k = S y u k + b y k , 1 μ ,
where, for any v = ( v i , j ) and any scalar γ > 0 ,
S ( v , γ ) = z = ( z i , j ) , z i , j = z i , j | z i , j | max | z i , j | γ , 0 .
Finally, an approximate solution to the minimization problem with respect to u can be obtained by applying Gauss–Seidel (GS) iterations to the following system, as explained in [33]:
Δ u i , j = r i , j μ + ( x ( b x k d x k ) i , j ) + ( y ( b y k d y k ) i , j ) ,
where Δ is the classical finite-difference discretization of the Laplacian. If the solution to (17) lies outside [ 0 , 1 ] m × n , then it is projected onto that set. We denote P [ 0 , 1 ] the corresponding projection  operator.
The overall solution method is outlined in Algorithm 1. Note that, when the approach in Section 2.3 is used, the values λ i , j must be updated at each iteration k, using (12) with u = u k .
Algorithm 1 Split Bregman (SB)-based method for spatially adaptive segmentation
  Input:  u ¯ , λ m i n , λ m a x , f , μ , α (with f defined in (9) or (11) or (12))
  Output:  u , c 1 , c 2
  Set u 0 = u ¯ , d x 0 = 0 , d y 0 = 0 , b x 0 = 0 , b y 0 = 0
  Compute Λ = f ( u 0 )
  for k = 1 , 2 , do
    Algorithms 13 00226 i001Compute c 1 k and c 2 k by (13)
Compute u k by applying GS iterations to system (17)
u k = P [ 0 , 1 ] ( u k )
d x k = S ( x u k + b x k , 1 / μ )
d y k = S ( y u k + b y k , 1 / μ )
Update b x k and b y k according to (16)
  end
  Set Σ = ( i , j ) Θ : u i , j k > α

4. Results and Comparisons

The three spatially adaptive regularization techniques were implemented in MATLAB, while using the Image Processing Toolbox. Algorithm 1 was implemented by modifying the Fast Global Minimization Algorithm for Active Contour Models by X. Bresson, available from http://htmlpreview.github.io/?https://github.com/xbresson/old_codes/blob/master/codes.html. This is a C code with a MATLAB MEX  interface.
L σ in (7) is defined as a rotationally symmetric Gaussian filter with size 3 and standard deviation σ = 2 . The mean and median filters use windows of size 3 and 7, respectively, and the parameter t in (10) is set as t = 0.5 . The parameter α = 0.5 is used to identify the domain Σ according to (4).
In the original and modified codes, the SB iterations are stopped, as follows:
| diff k diff k 1 | tol and k > maxit ,
where
diff l = sd ( u l ) sd ( u l ) · sd ( u l 1 ) , sd ( u l ) = i , j ( u i , j l u i , j l ) 2 ,
tol is a given tolerance and maxit denote the maximum number of outer iterations. The stopping criterion for the GS iterations is
E ( u k ) = 1 | msd k msd 1 | msd 1 tol GS and k > maxit GS
where
msd l = 1 m n i , j ( u i , j l u i , j l 1 ) 2 ,
and tol GS and maxit GS are the tolerance and the maximum number of iterations for the GS method, respectively. In our experiments we set tol = 10 6 , maxit = 30 , tol GS = 10 2 and maxit GS = 50 .
The adaptive models are compared with the original CEN model on different images widely used in image processing tests, as listed in Table 1 and shown in Figure 1, Figure 2, Figure 3 and Figure 4. In particular, the images bacteria, bacteria2, brain, cameraman, squirrel, and tiger are included in Bresson’s code distribution, flowerbed has been downloaded from the Berkeley segmentation dataset [34] available from https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/resources.html (image #86016), and ninetyeight is available from https://tineye.com/query/2817cf0d186fbfe263a188952829a3b5e699d839. We note that tiger is included in Bresson’s code as a test problem for a segmentation model specifically developed for images with texture [35], which is also implemented in the code. This model uses the well-known Kullback–Leibler divergence function regularized with a TV term. The model is solved with the SB method. We perform the segmentation of tiger with the CEN model, the textural segmentation model, and our approaches, in order to investigate whether our locally adaptive model can be also suitable for textural images. The cameraman image is perturbed with additive Gaussian noise, with zero mean and two values of standard deviation, σ = 15 , 25 , with the aim of evaluating the behavior of our adaptive approaches on noisy images. The noisy images are called cameraman15 and cameraman25.
For the images provided with Bresson’s code, the values of λ and μ associated with the original CEN model are set as in that code. For the remaining images, the values of λ and μ are set by trial and error, following the empirical rule reported in Bresson’s code. The values of λ m i n and λ m a x used in the spatially adaptive approaches are chosen, so that the corresponding λ in the original CEN model is in [ λ m i n , λ m a x ] , with few exceptions. The associated values of μ are set as in the non-adaptive case. The values of λ , λ m i n , λ m a x and μ used for each image are specified in Table 2, Table 3 and Table 4. The values of λ i , j in the adaptive models are initialized, as specified in (9), (11), and (12). As described in Section 2.3, in the strategy based on thresholding those values change at each iteration k of Algorithm 1. It is worth noting that our adaptive approaches also simplify the choice of the regularization parameter, which may be a time-consuming task.
The initial approximation u 0 is set equal to the given image u ¯ i , j , which takes its values in [ 0 , 1 ] , as specified in Section 1. This is used to compute the starting values of c 1 and  c 2 , and s i , j = c 1 u ¯ i , j 2 c 2 u ¯ i , j 2 for all ( i , j ) . In the original non-adaptive code, the value of λ is scaled using the difference between the maximum and the minimum value of s i , j ; the same scaling is applied to λ m i n and λ m a x in the implementations of the adaptive approaches.
We run the tests on an Intel Core i7 processor with clock frequency of 2.6 GHz, 8 GB of RAM, and a 64-bit Linux system.
For each of the six images bacteria, bacteria2, brain, flowerbed, ninetyeight, and squirrel, we show the results that were obtained with the spatially adaptive strategy yielding the best segmentation for that image. The corresponding (unscaled) values of λ , λ m i n , λ m a x , the value of μ , as well as the number of outer iterations and the mean number of GS iterations per outer iteration are reported in Table 2. Note that, for squirrel, we use two values of λ m i n , one equal to the value of λ in the CEN model and the other greater than that, obtained by trial and error. Both values of λ m i n produce the same segmentation, but the larger value of λ m a x reduces the number of outer and GS iterations. The segmentations corresponding to the data in Table 2 are shown in Figure 1, Figure 2 and Figure 3. For squirrel, we display the CTD segmentation computed by using the larger value of λ m i n .
We see that, on selected problems, CTD reduces the number of outer and GS iterations with respect to the CEN model; on the other hand, the setup of the regularization parameters is computationally more expensive. In terms of iterations, there is no clear winner between CEN and MM and between CEN and THR. The models based on the spatially adaptive techniques are slightly more expensive than the CEN model in this case too. A significant result is that the segmentations obtained with the adaptive techniques appear better than those obtained with the non-adaptive model, i.e., the spatially adaptive models can better outline boundaries between objects and foreground. This is clearly visible by looking at the segmentations of brain, bacteria2 (see the upper right corner), ninetyeight, bacteria (see the shape of the bacteria). It is also worth noting that the adaptive model based on THR removes textural details that do not belong to the flowerbed in the homonymous image.
The latter observation is confirmed by the experiments performed on tiger. Table 3 reports the corresponding model and algorithmic details, while the segmentations are shown in Figure 4 along with (visual) information on quantities used to define λ i , j (see (8) and (10)). We see that the CTD and MM strategies produce satisfactory results, although they have been obtained by generalizing a non-textural model.
Finally, we show the results obtained on cameraman and its noisy versions by using CEN, CTD, MM, and THR. The methods perform comparable numbers of inner and GS iterations (see Table 4), but the spatially adaptive model THR yields some improvement over the CEN model in the segmentation of the images that are affected by Gaussian noise (Figure 5).

5. Conclusions

We introduced spatially adaptive regularization in a well-established variational segmentation model with the aim of improving the segmentation of images by suitably taking into account their smooth and nonsmooth regions. To this aim, we introduced three techniques, based on the application of suitable spatial filters or thresholding. The locally adaptive models, solved via an alternating minimization method using split Bregman iterations, showed the effectiveness of our approaches on several images, also including textural and noisy images. We also believe that the proposed models may simplify the setup of the regularization parameter.
Future work can include the extension of our spatially adaptive strategies to other segmentation models and the development of further adaptive techniques.

Author Contributions

All the authors contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by Istituto Nazionale di Alta Matematica - Gruppo Nazionale per il Calcolo Scientifico (INdAM-GNCS), Italy. L. Antonelli was also supported by the Italian Ministry of University and Research under grant no. PON03PE_00060_5.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Mumford, D.; Shah, J. Optimal approximations by piecewise smooth functions and associated variational problems. Comm. Pure Appl. Math. 1989, 42, 577–685. [Google Scholar] [CrossRef] [Green Version]
  2. Chan, T.F.; Vese, L.A. Active contours without edges. IEEE Trans. Image Process. 2001, 10, 266–277. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Chan, T.F.; Esedoḡlu, S.; Nikolova, M. Algorithms for Finding Global Minimizers of Image Segmentation and Denoising Models. SIAM J. Appl. Math. 2006, 66, 1632–1648. [Google Scholar] [CrossRef]
  4. Wang, F.; Zhao, C.; Liu, J.; Huang, H. A variational image segmentation model based on normalized cut with adaptive similarity and spatial regularization. SIAM J. Imaging Sci. 2020, 13, 651–684. [Google Scholar] [CrossRef]
  5. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Martinez-Gonzalez, P.; Garcia-Rodriguez, J. A survey on deep learning techniques for image and video semantic segmentation. Appl. Soft Comput. 2018, 70, 41–65. [Google Scholar] [CrossRef]
  6. Guo, Y.; Liu, Y.; Georgiou, T.; Lew, M.S. A review of semantic segmentation using deep neural networks. Int. J. Multimed Inf. Retr. 2018, 7, 87–93. [Google Scholar] [CrossRef] [Green Version]
  7. Borji, A.; Cheng, M.M.; Hou, Q.; Jiang, H.; Li, J. Salient object detection: A survey. Comput. Vis. Media (Beijing) 2019, 5, 117–150. [Google Scholar] [CrossRef] [Green Version]
  8. Lateef, F.; Ruichek, Y. Survey on semantic segmentation using deep learning techniques. Neurocomputing 2019, 338, 321–348. [Google Scholar] [CrossRef]
  9. Candemir, S.; Akgül, Y.S. Adaptive Regularization Parameter for Graph Cut Segmentation. In Image Analysis and Recognition. ICIAR 2010; Campilho, A., Kamel, M., Eds.; Lect. Notes Comput. Sci. Springer: Berlin/Heidelberg, Germany, 2010; Volume 6111, pp. 117–126. [Google Scholar] [CrossRef]
  10. Rao, J.; Abugharbieh, R.; Hamarneh, G. Adaptive Regularization for Image Segmentation Using Local Image Curvature Cues. In Computer Vision—ECCV 2010; Daniilidis, K., Maragos, P., Paragios, N., Eds.; Lect. Notes Comput. Sci. Springer: Berlin/Heidelberg, Germany, 2010; Volume 6314, pp. 651–665. [Google Scholar] [CrossRef] [Green Version]
  11. Fong, W.L.; Ng, M.K. On selection of spatial-varying regularization parameters in total variation image restoration. In Proceedings of the 2011 International Workshop on Multidimensional (nD) Systems, Poitiers, France, 5–7 September 2011; pp. 1–5. [Google Scholar] [CrossRef]
  12. Bredies, K.; Dong, Y.; Hintermüller, M. Spatially dependent regularization parameter selection in total generalized variation models for image restoration. Int. J. Comput. Math. 2013, 90, 109–123. [Google Scholar] [CrossRef]
  13. Chan, R.H.; Lanza, A.; Morigi, S.; Sgallari, F. An adaptive strategy for the restoration of textured images using fractional order regularization. Numer. Math. Theory Methods Appl. 2013, 6, 276–296. [Google Scholar] [CrossRef]
  14. Ma, T.H.; Huang, T.Z.; Zhao, X.L. Spatially dependent regularization parameter selection for total generalized variation-based image denoising. Comput. Appl. Math. 2018, 37, 277–296. [Google Scholar] [CrossRef]
  15. Calatroni, L.; Lanza, A.; Pragliola, M.; Sgallari, F. A flexible space-variant anisotropic regularization for image restoration with automated parameter selection. SIAM J. Imaging Sci. 2019, 12, 1001–1037. [Google Scholar] [CrossRef]
  16. Antonelli, L.; De Simone, V. Comparison of minimization methods for nonsmooth image segmentation. Commun. Appl. Ind. Math. 2018, 9, 68–86. [Google Scholar] [CrossRef] [Green Version]
  17. Di Serafino, D.; Landi, G.; Viola, M. ACQUIRE: An inexact iteratively reweighted norm approach for TV-based Poisson image restoration. Appl. Math. Comput. 2020, 364, 124678. [Google Scholar] [CrossRef]
  18. Birgin, E.G.; Martínez, J.M.; Raydan, M. Nonmonotone spectral projected gradient methods on convex sets. SIAM J. Optim. 2000, 10, 1196–1211. [Google Scholar] [CrossRef]
  19. Antonelli, L.; De Simone, V.; di Serafino, D. On the Application of the Spectral Projected Gradient Method in Image Segmentation. J. Math. Imaging Vis. 2016, 54, 106–116. [Google Scholar] [CrossRef]
  20. Di Serafino, D.; Ruggiero, V.; Toraldo, G.; Zanni, L. On the steplength selection in gradient methods for unconstrained optimization. Appl. Math. Comput. 2018, 318, 176–195. [Google Scholar] [CrossRef] [Green Version]
  21. Crisci, S.; Porta, F.; Ruggiero, V.; Zanni, L. Spectral properties of Barzilai-Borwein rules in solving singly linearly constrained optimization problems subject to lower and upper bounds. SIAM J. Optim. 2020, 30, 1300–1326. [Google Scholar] [CrossRef]
  22. Bregman, L.M. The relaxation method of finding the common points of convex sets and its application to the solution of problems in convex programming. U.S.S.R. Comput. Math. Math. Phys. 1967, 7, 200–217. [Google Scholar] [CrossRef]
  23. Osher, S.; Burger, M.; Goldfarb, D.; Xu, J.; Yin, W. An iterative regularization method for total variation-based image restoration. Multiscale Model. Simul. 2005, 4, 460–489. [Google Scholar] [CrossRef]
  24. Goldstein, T.; Osher, S. The split Bregman method for L1-regularized problems. SIAM J. Imaging Sci. 2009, 2, 323–343. [Google Scholar] [CrossRef]
  25. Campagna, R.; Crisci, S.; Cuomo, S.; Marcellino, L.; Toraldo, G. Modification of TV-ROF denoising model based on split Bregman iterations. Appl. Math. Comput. 2017, 315, 45–467. [Google Scholar] [CrossRef] [Green Version]
  26. Parikh, N.; Boyd, S. Proximal Algorithms. Found. Trends Optim. 2014, 1, 127–239. [Google Scholar] [CrossRef]
  27. Beck, A.; Teboulle, M. Fast Gradient-Based Algorithms for Constrained Total Variation Image Denoising and Deblurring Problems. IEEE Trans. Image Process. 2009, 18, 2419–2434. [Google Scholar] [CrossRef] [Green Version]
  28. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Found. Trends Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
  29. Buades, A.; Le, T.M.; Morel, J.; Vese, L.A. Fast Cartoon + Texture Image Filters. IEEE Trans. Image Process. 2010, 19, 1978–1986. [Google Scholar] [CrossRef]
  30. Buades, A.; Le, T.M.; Morel, J.; Vese, L.A. Cartoon + Texture Image Decomposition. Image Process. Line 2011, 1, 200–207. [Google Scholar] [CrossRef] [Green Version]
  31. Jain, A.K. Fundamentals of Digital Image Processing; Prentice-Hall: Englewood Cliffs, NJ, USA, 1989. [Google Scholar]
  32. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 2nd ed.; Addison-Wesley Longman Publishing Co., Inc.: Boston, MA, USA, 2001. [Google Scholar]
  33. Goldstein, T.; Bresson, X.; Osher, S. Geometric applications of the split Bregman method: Segmentation and surface reconstruction. J. Sci. Comput. 2010, 45, 272–293. [Google Scholar] [CrossRef] [Green Version]
  34. Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour Detection and Hierarchical Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 898–916. [Google Scholar] [CrossRef] [Green Version]
  35. Houhou, N.; Thiran, J.P.; Bresson, X. Fast Texture Segmentation Based on Semi-Local Region Descriptor and Active Contour. Numer. Math. Theory Methods Appl. 2009, 2, 445–468. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Segmentations of brain and squirrel by using CEN and CTD.
Figure 1. Segmentations of brain and squirrel by using CEN and CTD.
Algorithms 13 00226 g001
Figure 2. Segmentations of bacteria2 and ninetyeight by using CEN and MM.
Figure 2. Segmentations of bacteria2 and ninetyeight by using CEN and MM.
Algorithms 13 00226 g002
Figure 3. Segmentations of bacteria and flowerbed by using CEN and THR.
Figure 3. Segmentations of bacteria and flowerbed by using CEN and THR.
Algorithms 13 00226 g003
Figure 4. Segmentations of tiger obtained by using the textural, CEN, CTD and MM models. A representation of quantities used to define λ i , j (see (8) and (10)) is also provided.
Figure 4. Segmentations of tiger obtained by using the textural, CEN, CTD and MM models. A representation of quantities used to define λ i , j (see (8) and (10)) is also provided.
Algorithms 13 00226 g004
Figure 5. Segmentations of cameraman and its noisy versions by using the CEN, CTD, MM, and THR  models.
Figure 5. Segmentations of cameraman and its noisy versions by using the CEN, CTD, MM, and THR  models.
Algorithms 13 00226 g005aAlgorithms 13 00226 g005b
Table 1. Test images and their sizes.
Table 1. Test images and their sizes.
ImageSize (pixels)
bacteria 233 × 256
bacteria2 380 × 391
brain 210 × 210
cameraman 204 × 204
flowerbed 321 × 481
ninetyeight 300 × 225
squirrel 167 × 230
tiger 321 × 481
Table 2. Segmentation of six test images by the CEN and spatially adaptive models: parameters and iterations. The value of μ is the same for all the models, thus it is reported only once per image.
Table 2. Segmentation of six test images by the CEN and spatially adaptive models: parameters and iterations. The value of μ is the same for all the models, thus it is reported only once per image.
λ μ it it GS λ min λ max it it GS
ImageCENCTD
brain 0.7 × 10 3 0.1 × 10 4 519.2 0.9 × 10 3 0.4 × 10 5 514.8
squirrel 0.118 × 10 3 0.1 × 10 4 650 0.118 × 10 3 0.1 × 10 4 1550
0.2 × 10 3 0.1 × 10 4 545
CENMM
bacteria2 0.1 × 10 3 0.1 × 10 4 250 0.1 × 10 3 0.1 × 10 4 1050
ninetyeight 0.2 × 10 2 0.1 × 10 3 438.5 0.1 × 10 2 0.5 × 10 2 350
CENTHR
bacteria 0.5 × 10 4 0.1 × 10 5 627.5 0.4 × 10 4 0.8 × 10 4 730.7
flowerbed 0.1 × 10 2 0.1 × 10 4 227 0.3 × 10 2 0.1 × 10 3 216.5
Table 3. Segmentation of tiger by the textural [35], CEN, CTD, and MM models: parameters and  iterations.
Table 3. Segmentation of tiger by the textural [35], CEN, CTD, and MM models: parameters and  iterations.
Image λ μ it it GS λ min λ max μ it it GS
Textural ModelCTD
tiger 0.3 × 10 1 0.1 × 10 3 512 0.1 × 10 1 0.1 × 10 2 0.1 × 10 2 850
CENMM
tiger 0.3 × 10 2 0.1 × 10 3 749.6 0.1 × 10 1 0.1 × 10 2 0.1 × 10 2 850
Table 4. Segmentations of cameraman and its noisy versions by using CEN and the spatially adaptive models: parameters and iterations. The value of μ is the same for all of the models, thus it is reported only once per image.
Table 4. Segmentations of cameraman and its noisy versions by using CEN and the spatially adaptive models: parameters and iterations. The value of μ is the same for all of the models, thus it is reported only once per image.
CENCTD
Image λ μ it it GS λ m i n λ m a x it it GS
cameraman 0.8 × 10 3 0.1 × 10 3 42.8 0.5 × 10 3 0.1 × 10 4 43
cameraman15 0.3 × 10 3 0.1 × 10 3 54.2 0.3 × 10 3 0.1 × 10 4 54.2
cameraman25 0.17 × 10 3 0.1 × 10 3 67 0.17 × 10 3 0.8 × 10 3 67
MMTHR
Image λ m i n λ m a x it it GS λ m i n λ m a x it it GS
cameraman 0.5 × 10 3 0.1 × 10 4 43 0.5 × 10 3 0.1 × 10 4 42.8
cameraman15 0.3 × 10 3 0.1 × 10 4 44.2 0.3 × 10 3 0.1 × 10 4 53.6
cameraman25 0.17 × 10 3 0.5 × 10 3 67 0.17 × 10 3 0.8 × 10 3 75.7

Share and Cite

MDPI and ACS Style

Antonelli, L.; De Simone, V.; di Serafino, D. Spatially Adaptive Regularization in Image Segmentation. Algorithms 2020, 13, 226. https://doi.org/10.3390/a13090226

AMA Style

Antonelli L, De Simone V, di Serafino D. Spatially Adaptive Regularization in Image Segmentation. Algorithms. 2020; 13(9):226. https://doi.org/10.3390/a13090226

Chicago/Turabian Style

Antonelli, Laura, Valentina De Simone, and Daniela di Serafino. 2020. "Spatially Adaptive Regularization in Image Segmentation" Algorithms 13, no. 9: 226. https://doi.org/10.3390/a13090226

APA Style

Antonelli, L., De Simone, V., & di Serafino, D. (2020). Spatially Adaptive Regularization in Image Segmentation. Algorithms, 13(9), 226. https://doi.org/10.3390/a13090226

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop