Next Article in Journal
An Adaptive Learning Model for Multiscale Texture Features in Polyp Classification via Computed Tomographic Colonography
Next Article in Special Issue
DFusion: Denoised TSDF Fusion of Multiple Depth Maps with Sensor Pose Noises
Previous Article in Journal
Kinetic Electromagnetic Energy Harvester for Railway Applications—Development and Test with Wireless Sensor
Previous Article in Special Issue
Deep-Learning-Based Adaptive Advertising with Augmented Reality
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Content-Aware SLIC Super-Pixels for Semi-Dark Images (SLIC++)

1
High Performance Cloud Computing Center (HPC3), Department of Computer and Information Sciences, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Malaysia
2
Faculty of Engineering Science and Technology, Iqra University, Karachi 75600, Pakistan
3
Department of Computer Science, Shaheed Zulfiqar Ali Bhutto Institute of Science and Technology, Karachi 75600, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 906; https://doi.org/10.3390/s22030906
Submission received: 27 September 2021 / Revised: 16 November 2021 / Accepted: 17 November 2021 / Published: 25 January 2022
(This article belongs to the Special Issue Computer Vision and Machine Learning for Intelligent Sensing Systems)

Abstract

:
Super-pixels represent perceptually similar visual feature vectors of the image. Super-pixels are the meaningful group of pixels of the image, bunched together based on the color and proximity of singular pixel. Computation of super-pixels is highly affected in terms of accuracy if the image has high pixel intensities, i.e., a semi-dark image is observed. For computation of super-pixels, a widely used method is SLIC (Simple Linear Iterative Clustering), due to its simplistic approach. The SLIC is considerably faster than other state-of-the-art methods. However, it lacks in functionality to retain the content-aware information of the image due to constrained underlying clustering technique. Moreover, the efficiency of SLIC on semi-dark images is lower than bright images. We extend the functionality of SLIC to several computational distance measures to identify potential substitutes resulting in regular and accurate image segments. We propose a novel SLIC extension, namely, SLIC++ based on hybrid distance measure to retain content-aware information (lacking in SLIC). This makes SLIC++ more efficient than SLIC. The proposed SLIC++ does not only hold efficiency for normal images but also for semi-dark images. The hybrid content-aware distance measure effectively integrates the Euclidean super-pixel calculation features with Geodesic distance calculations to retain the angular movements of the components present in the visual image exclusively targeting semi-dark images. The proposed method is quantitively and qualitatively analyzed using the Berkeley dataset. We not only visually illustrate the benchmarking results, but also report on the associated accuracies against the ground-truth image segments in terms of boundary precision. SLIC++ attains high accuracy and creates content-aware super-pixels even if the images are semi-dark in nature. Our findings show that SLIC++ achieves precision of 39.7%, outperforming the precision of SLIC by a substantial margin of up to 8.1%.

1. Introduction

Image segmentation has potential to reduce the image complexities associated with processing of singular image primitives. Low-level segmentation of an image in non-overlapping set of regions called super-pixels helps in pre-processing and speeding up further high-level computational tasks related to visual images. The coherence feature of super-pixels allows faster architectural functionalities of many visual applications including object localization [1], tracking [2], posture estimation [3], recognition [4,5], semantic segmentation [6], instance segmentation [7], and segmentation of medical imagery [8,9]. These applications will be aided by super-pixels in terms of boosted performances, as the super-pixels put forward only the discriminating visual information [10].
Low-level segmentation tends to result in incorrect segmentations if the visual image has high pixel intensities; these high pixeled values are usually the biproduct of visual scenes captured in low lighting conditions, i.e., semi-dark images and dark images. The obtained incorrect super-pixels are attributed to the underlying approach used for the final segment creation. The existing super-pixel methods fail due to incorrect pixel manipulations for base operational functionality. The currently employed pixel manipulation relies on straight line differences for super-pixel creation. These straight-line difference manipulations fail to retain the content-aware information of the image. The results are further degraded if the image has low contrasted values, which result in no clear discrimination among the objects present. The existing super-pixel creation methods are divided into two categories based on the implemented workflow. The two categories are graph-based and gradient ascent based [10,11]. The former, focuses on minimization of cost function by grouping and treating each pixel of the image as a graph node. The later, iteratively processes each image pixel using clustering techniques until convergence [12]. One of the typical features observed for identification of super-pixel accuracy is regularity, i.e., to what extent the super-pixel is close to the actual object boundary of the image. All the graph-based methods for super-pixel calculations suffer from poorly adhered super-pixels which result in irregularity of segments presented by super-pixels [11]. Additionally, graph-based methods are constrained by the excessive computational complexity and other initialization parameters. Whereas, the gradient-ascent methods are simplistic in nature and are recommended in the literature due to resultant high performance and accuracy [13,14]. However, there are some issues associated with content irrelevant manipulation of singular pixels to form resultant super-pixels. Some of the key features of using super-pixel segmentations are:
  • Super-pixels abstraction potentially decreases the overhead of processing each pixel at a time.
  • Gray-level or color images can be processed by a single algorithm implementation.
  • Integrated user control provided for performance tuning.
With these advantages of super-pixels, they are highly preferred. However, super-pixel abstraction methods backed by gradient-ascent workflows are also limited in their working functionality to retain the contextual information of the given image [15]. The contextual information retainment is required to achieve the richer details of the visual image. This loss of contextual information is caused by flawed pixel clusters created based on Euclidean distance measure [16]. As the Euclidean distance measures calculates straight line differences among pixels which ends up in irregular and lousy super-pixels. Moreover, further degradation can be expected to process the semi-dark images where high pixel intensities along with no clear boundaries are observed. In such scenarios, the propagation of inaccurate super-pixels will affect the overall functionality of automated solutions [11]. To overcome this problem of information loss for creation of compact and uniform super-pixels, we propose content-aware distance measures for image pixel cluster creation. The content-aware distance measure as the core foundational component of gradient-ascent methods for super-pixel creation will not only help in alleviation of information loss, but it will also help in preserving less observant/perceptually visible information of semi-dark images. The state-of-the-art methods for super-pixels creation have not been analyzed exclusively on the semi-dark images which further raises the concerns related to segmentation accuracy. In nutshell, the problems in existing segmentation algorithms are:
  • Absence of classification of state-of-the-art methods based on low level manipulations.
  • Inherited discontinuity in super-pixel segmentation due to inconsistent manipulations.
  • Unknown pixel grouping criteria in terms of distance measure to retain fine grained details.
  • Unknown effect of semi-dark images on final super-pixel segmentations.
To resolve these issues, the presented research exclusively presents multifaceted study offering following features:
  • Classification of Literary Studies w.r.t Singular Pixel Manipulation Strategies:
The categorization of the existing studies is based on entire image taken as one entity. The image entity can represent either graph or a feature space to be clustered, i.e., graph-based or gradient-ascent based methodology for pixel grouping. To the best of our knowledge, there has been no study that categorizes existing studies based on the manipulation strategy performed over each pixel. We present the detailed comparative analysis of existing research highlighting their core functionality as the basis for classification.
2.
Investigation of Appropriate Pixel Grouping Scheme:
The grouping scheme backing the image segmentation module being crucial component can highly affect the accuracy of the entire model. For this reason, to propose a novel extension as a generalized solution of all types of images including semi-dark images, we present a detailed qualitative investigation of up to seven distance measures for grouping pixels to create super-pixels. The investigation resulted in shortlisted pixel grouping measures to retain fine grained details of the visual image.
3.
Novel Hybrid Content-Aware Extension of SLIC—SLIC++:
SLIC, being the simplest and fastest solution for the pixel grouping, remains the inspiration, and we enhance the performance of SLIC by adding content-aware feature in its discourse. The proposed extension holds the fundamental functionality with improved features to preserve content-aware information. The enhancement results in better segmentation accuracy by extracting regular and continuous super-pixels for all scenarios including semi-dark scenarios.
4.
Comprehensive Perceptual Results focusing Semi-dark Images:
To assess the performance of the proposed extension SLIC++ for extraction of the richer information of the visual scene, we conduct experiments over semi-dark images. The experimental analysis is benchmarked against the standard super-pixel creation methods to verify that incorporating content-aware hybrid distance measure leads to improved performance. The perceptual results further conform better performance, the scalability, and generalizability of results produced by SLIC++.

Paper Organization

The remainder of paper is organized as follows: Section 2 presents the prior super-pixel creation research and its relevance to semi-dark technology. In this section we also present critical analysis of studies proposed over period of two decades and their possible applicability for semi-dark imagery. We also critically analyze two closely related studies and highlight the difference among them w.r.t SLIC++. Section 3 describes the extension hypothesis and the final detailed proposal for super-pixel segmentation of semi-dark images. Section 4 presents the detailed quantitative and qualitative analysis of SLIC extension against state-of-the-art algorithms validating the proposal. Section 5 discusses the applicability of the proposed algorithm in the domain of computer vision. Finally, Section 6 concludes the presented research and points out some future directions of research.

2. Literature Review

2.1. Limited Semi-Dark Image Centric Research Focusing Gradient-Ascent Methods

The gradient ascent methods are also called clustering-based methods. These methods take the input image and rasterize the image. Then, based on the local image cues such as color and spatial information the pixels are clustered iteratively. After each iteration, gradients are calculated to refine the new clusters from the previously created clusters [14]. The iterative process continues till the algorithm converges after the gradients stop changing, thus named gradient-ascent methods. A lot of research has been already done in the domain of gradient-ascent methods; the list of these methods is presented in Table 1.
The gradient ascent methods for super-pixel creation seems to be a promising solution due to their simplicity of implementation, speed of processing and easy adaptation for handling the latest demands of complex visual image scenarios. However, the concerns associated with underlying proper extraction strategies is one of the challenging aspects to cater the dynamic featural requirements imposed by complex visual image scenarios such as semi-dark images.

2.2. Critical Analysis of Gradient-Ascent Super-Pixel Creation Algorithms Based on Manipulation Strategy

For the critical analysis we have considered gradient ascent based super-pixel algorithms presented over period of two decades ranging from 2001 through 2021. The studies are retrieved from Google Scholar’s repository with keywords including super-pixel segmentation, pixel abstraction, content sensitive super-pixel creation, content-aware super-pixel segmentation. The search resulted in a lot of segmentation related studies in domain of image processing including the basic image transformations along with related super-pixel segmentation studies. For the critical analysis, the studies mentioning clustering based super-pixel creations were shortlisted due to their relevance with proposed algorithm. The key features of these studies are critically analyzed and comprehensively presented in Table 1, along with the critiques for respective handling concerns associated with semi-dark imagery.

Key Takeaways

The critical analysis presented in Table 1 uncovers the fact that recent research explicitly points towards the need of segmentation algorithm which considers the content relevant super-pixel segmentations. To accomplish this task several techniques are proposed with incorporation of prior transformations of image via deep learning methods, simple image processing, and probabilistic methods. Mostly the research uses and conforms the achievements of Simple Linear Iterative Clustering (SLIC). Furthermore, most of the research studies are using SLIC algorithm for super-pixel creation as base mechanisms with added features. Generally, the algorithms proposed in last decade have computational complexity of O N , whereas if neural networks are employed for automation of required parameter initialization, then the complexity becomes O N N o .   o f   l a y e r s . All the proposed algorithms use two distance measures for final super-pixel creation, i.e., Euclidean, or geodesic distance measure. However, all these studies have not mentioned the occurrence of semi-dark images and their impact on the overall performance. It is estimated that huge margin of the existing image dataset already includes the problem centric image data. The Berkeley dataset that is substantially used for performance analysis of super-pixel algorithms contains up to 63% semi-dark images. The proposed study uses the semi-dark images of Berkeley dataset for benchmarking analysis of the SLIC++ algorithm.

2.3. Exclusiveness of SLIC++ w.r.t Recent Developments

The recent studies substantially focus on super-pixels with the induced key features of content sensitivity and adherence of the final segmentations; consequently several related research studies have been proposed. Generally, the desired features are good boundary adherence, compact super-pixel size and low complexity. The same features are required for super-pixel segmentation of semi-dark images. In this section we briefly review the recent developments which are closely related to our proposed method for creating super-pixel in semi-dark images.
BASS (Boundary-Aware Super-pixel Segmentation) [28], is closely related to the methodology that we have chosen, i.e., incorporation of content relevant information in the final pixel labeling which ends up with the creation of super-pixels. However, the major difference resides in the initialization of the super-pixel seeds/centers. BASS recommends the usage of forest classification method prior to the super-pixel creation. This forest classification of image space results with the creation of a binary image with highlighted boundary information over the image space. This boundary information is then utilized to aid the initialization process of the seed/ cluster centers. Theoretically, the problem with this entire configuration is additional complexity of boundary map creation which raise the complexity from O N to O N l o g N . This boundary map creation and its associated condition of addition and deletion of seeds is expected to further introduce undesired super-pixel feature under-segmentation. The under-segmentation might take place due to easy seed deletion condition and difficult addition condition which means more seeds would be deleted and less would be added. This aspect of the study is not desired for the super-pixel creation in semi-dark scenarios. On the contrary we propose regularly distributed seeds along with usage of both the recommended distance measures without prior image transformation which further reduces the overall complexity. Finally, we also propose the usage of geodesic distance for color components of the pixel rather than only using it for spatial component.
Intrinsic manifold SLIC [30], is an extension of manifold SLIC which proposes usage of manifolds to map high dimensional image data on the manifolds resembling Euclidean space near each point. IMSLIC uses Geodesic Centroidal Voronoi Tessellations (GCVT) this allows flexibility of skipping the post-processing heuristics. For computation of geodesic distance on image manifold weighted image graph is overlayed with the same graph theory of edges, nodes, and weights. For this mapping 8-connected neighbors of pixel are considered. This entire process of mapping and calculation of geodesic distances seems complex. The theoretical complexity is O N , however with the incorporation of image graph the computational complexity will increase. Moreover, the conducted study computes only geodesic distance between the pixels leaving behind the Euclidean counterpart. With substantially less complexity we propose to implement both the distance measures for all the crucial pixel components.

2.4. Summary and Critiques

The comprehensive literature survey is conducted to benefit the readers and provide a kickstart review of advancements of super-pixel segmentation over the period of two decades. Moreover, the survey resulted in critical analysis of existing segmentation techniques which steered the attention to studies conducted for adverse image scenarios such as semi-dark images. Arguably, with the increased automated solutions the incoming image data will be of dynamic nature (including lighting conditions). To deal with this dynamic image data, there is a critical need of a super-pixel segmentation technique that takes into account the aspect of semi-dark imagery and results in regular and content-aware super-pixel segmentation in semi-dark scenarios. The super-pixel segmentation techniques currently employed for the task suffer from two major issues, i.e., high complexity, and information loss. The information loss associated with the gradient-ascent methods is attributed to restrictions imposed due to usage of Euclidean image space which totally loses the context of the information present in the image by calculating straight line differences. Many attempts have been made to incorporate CNN probabilistic methods in super-pixel creation methods to optimize and aid the final segmentation results. However, to the best of our knowledge there has been no method proposed exclusively for semi-dark images scenarios keeping the simplicity and optimal performance intact.
In following sections, we describe the preliminaries which are the base for the proposed extension of SLIC namely SLIC++. We also present several distance measures incorporated in base SLIC algorithm namely SLIC+ to analyze the performance for semi-dark images.

3. Materials and Methods

3.1. The Semi-Dark Dataset

For the analysis of the content-aware super-pixel segmentation algorithm, we have used the state-of-the-art dataset which has been used in the literature for years now. The Berkeley image dataset [39] has been used for the comprehensive analysis and benchmarking of the proposed SLIC++ algorithm with the state-of-the-art algorithms. The Berkeley image dataset namely BSDS 500 has got five hundred images overall, whereas the problem under consideration is of semi-dark images. For this purpose, we have initially extracted semi-dark images using RPLC (Relative Perceived Luminance Classification) algorithm. The labels are created based on the manipulation of color model information, i.e., Hue, Saturation, Lightness (HSL) [40]. The final semi-dark images extracted from the BSDS-500 dataset turned out to be 316 images. Each image has resolution of either 321 × 481 or 481 × 321 dimensions. The BSDS-500 image dataset provides the basis for empirical analysis of segmentation algorithms. For the performance analysis and boundary detection, the BSDS-500 dataset provides ground-truth labels by at least five human annotators on average. This raises questions about the selection of annotation provided by the subjects. To deal with this problem, we have performed a simple logic over the image ground truth labels. All the image labels are iterated with ‘OR’ operation to generate singular ground truth image label. The ‘OR’ operation is performed to make sure that the final ground truth is suggested by most of the human annotators. Finally, every image is segmented and benchmarked against this single ground truth labeled image.

3.2. Desiderata of Accurate Super-Pixels

Generally, for super-pixel algorithms there are no definite features for super-pixels to be accurate. The literary studies refer accurate super-pixels in terms of boundary adherence, connectivity of super-pixels, super-pixel partitioning, compactness, regularity, efficiency, controllable number of super-pixels and so on [13,41,42]. As the proposed study is research focused on semi-dark images, we take into account features that are desired for conformation of accurate boundary extraction in semi-dark images.
1. 
Boundary Adherence
The boundary adherence is the measure to compute the accuracy to which the boundary has been extracted by super-pixels against boundary image or ground-truth images. The idea is to preserve information as much as possible by creating super-pixels over the image. The boundary adherence feature is basically a measure that results how accurately the super-pixels have followed the ground-truth boundaries. This can be easily calculated by segmentation quality metrics precision-recall.
2. 
Efficiency with Less Complexity
As super-pixel segmentation algorithms are now widely used as preprocessing step for further visually intelligent tasks. The second desired feature is efficiency with less complexity. The focus should be creation of memory-efficient and optimal usage of processing resources so that more memory and computational resources can be used by subsequent process. We take into account this feature and propose an algorithm that uses exactly same resources as of Basic SLIC with added distance measures in its discourse.
3. 
Controllable Number of Super-Pixels
The controllable number is super-pixels is a desired feature to ensure the optimal boundary is extracted using the computational resources ideally. The super-pixel algorithms are susceptible to this feature that is number of super-pixels. The number of super-pixels to be created can directly impact the overall algorithm performance. The performance is degraded in terms of under-segmentation or over-segmentation error. In the former one, the respective algorithm fails to retrieve most of the boundaries due to the smaller number of super-pixels to be created, whereas the latter one retrieves maximum boundary portions of the ground-truth images but there is surplus of computational resources.
Nevertheless, as mentioned earlier there is a huge list of accuracy measures and all those measures refer to different segmentation aspects and features. The required features and subsequent accuracy measures to be reported depend on the application of algorithm. For semi-dark image segmentation, it is mandatory to ensure that most of the optimal boundary is extracted and this requirement can be related to precision-recall metrics.

3.3. SLIC Preliminaries

Before presenting the SLIC++, we first introduce base functionality of SLIC. The overall functionality is based on creation of restricted windows in which the user defined seeds are placed, and clustering of image point is performed in this restricted window. This restricted window is called Voronoi Tessellations [43]. Voronoi Tessellations is all about partitioning the image plane into convex polygons. This polygon is square in case of SLIC initialization windows. The Voronoi tessellations are made such that each partition has one generating point and all the point in the partition are close to the generating point or the mass center of that partition. As the generating point lies in the center these partitions are also called Centroidal Voronoi Tessellation (CVT). The SLIC algorithm considers CIELAB color space for the processing, where every pixel p on image I is presented by color components and spatial components as c p = l p ,   a p ,   b p being colour components and p u , v being spatial components. For any two pixels SLIC measures straight line difference or Euclidean distance between the two pixels for the entire image space 5 .
The spatial distance between two pixels is given by d s and color component distance d c are given in Equations (1) and (2).
d s = u 1 u 2 2 + v 1 v 2 2
And,
d c = l 1 l 2 2 + a 1 a 2 2 + b 1 b 2 2
Here d s and d c represent Euclidean distance between pixel p 1 and p 2 . Instead of simple Euclidean, SLIC uses distance term infused with Euclidean norm given by Equation (3).
D s = d c + m S d s
The final distance term is normalized using interval S and m provides control over the super-pixel compactness which results in perceptually meaningful distance with balanced aspect of spatial and color components. Provided the number of super-pixels K seeds ( s i ) i = 1 K are evenly distributed in over the image I clusters are created in restricted regions of Voronoi Tessellations. The initialization seed are placed in image space within a window of 2 S × 2 S having center s i . After that simple K-means is performed over the pixels residing in the window to its center. SLIC computes the distance between pixels using Equation (3) and iteratively processes the pixels until convergence.

3.4. The Extension Hypothesis—Fusion Similarity Measure

The super-pixels created by the SLIC algorithm basically uses the Euclidean distance measure to create pixel clusters or the super-pixels based on the seed or cluster centers. The Euclidean distance measure takes into account the similarity among pixels using straight line differences among cluster centers and the image pixels. This property of distance measure results in distortion of extracted boundaries of image. The reason is measure remains same no matter if there is a path along the pixels. The path along the pixels will result in smoother and content relevant pixels [16,36]. The Euclidean distance overlays a segmentation map over the image without having relevance to the actual content present in the image. Moreover, large diversity in the image (light conditions/high density portions) result in unavoidable distortion. Therefore, we hypothesize to use accurate distance measure which presents content relevant information of the visual scene. For this reason, we extend the functionality of SLIC by replacing the Euclidean distance measure with four potential similarity measures including chessboard, cosine, Minkowski, and geodesic and named it as SLIC+. These distance measures have been used in the literature integrated in clustering algorithms for synthetic textual data clustering where studies mentioned to render reasonable results for focused problem solving [44]. However, we use these similarity measures to investigate the effects on visual images using SLIC approach. Prior to implementation, a brief introductory discussion will help understand the overall integration and foundation for choosing these similarity measures. The distance measures are basically the distance transforms applied on different images, specifying the distance from each pixel to the desired pixel. For uniformity and easy understanding, let pixel p 1 and p 2 have the coordinates ( x 1 ,   y 1 ) and ( x 2 ,   y 2 ) , respectively.
  • Chessboard: This measure calculates the maximum distance between vectors. This can be referred to measuring path between the pixels based on eight connected neighborhood whose edges are one unit apart. The chessboard distance along any co-ordinate is given by identifying maximum, as presented in Equation (4).
    D c h e s s = max x 2 x 1 , y 2 y 1
    Rationale of Consideration: Since the problem with existing similarity measures is loss of information, chessboard is one of the alternate to be incorporated in super-pixel creation base. This measure is considered as it takes into account information of eight connected neighbors of pixels under consideration. However, it might add computational overhead due to the same.
  • Cosine: This measure calculates distance based on the angle between two vectors. The cosine angular dimensions counteract the problem of high dimensionality. The inner angular product between the vectors turns out to be one if vectors were previously normalized. Cosine distance is based on cosine similarity which is then plugged-in distance equation. Equations (5) and (6) shows calculation of cosine distance between pixels.
    c o s i n e   s i m i l a r i t y = p 1 . p 2 p 1 2 p 2 2
    D c o s i n e = 1 c o s i n e   s i m i l a r i t y  
    Rationale of Consideration: One of the aspects of content aware similarity measure is to retain the angular information thus we attempted to incorporate this measure. The resulting super-pixels are expected to retain the content relevant boundaries. However, this measure does not consider magnitude of the vectors/pixels due to which boundary performance might fall.
  • Minkowski: This measure is a bit more intricate. It can be used for normed vector spaces, where distance is represented as vector having some length. The measure multiplies a positive weight value which changes the length whilst keeping its direction. Equation (7) presents distance formulation of Minkowski similarity measure.
    D m i n = p 2 p 1 µ 1 / µ
    Here µ is the weight, if its value is set to 1 the resultant measure corresponds to Manhattan distance measure. µ = 2 , refers to euclidean and µ = , refers to chessboard or Chebyshev distance measure.
    Rationale of Consideration: As user-control in respective application is desired, Minkowski similarity provides the functionality by replacing merely one parameter which changes the entire operationality without changing the core equations. However, here we still have problems relating to the retainment of angular information.
  • Geodesic: This measure considers geometric movements along the pixel path in image space. This distance presents locally shortest path in the image plane. Geodesic distance computes distance between two pixels which results in surface segmentation with minimum distortion. Efficient numerical implementation of geodesic distance is achieved using first order approximation. For approximation parametric surfaces are considered with n number of points on the surface. Given an image mask, geodesic distance for image pixels can be calculated using Equation (8).
    D g e o = min P x i , x j 0 1 D ( P x i , x j t P ˙ x i , x j t d t
    where P x i , x j t is connected path between pixel x i , x j , provided t = 0,1. The density function D x increments the distance and can be computed using Equation (9).
    D x = e E x υ ,   E x = I G σ I + γ
    where υ is scaling factor, E x is edge measurement also provides normalization of gradient magnitude of image I . G σ is the Gaussian function with its standard deviation being σ . γ minimizes the effect of weak intensity boundaries over density function. D(x) always produces constant distance, for homogeneous appearing regions if E(x) is zero D(x) becomes one.
    Rationale of Consideration: For shape analysis by computing distances geodesic has been the natural choice. However, computing geodesic distance is computationally expensive and is susceptible to noise [44]. Therefore, to overcome effect of noise geodesic distance should be used in amalgamation of Euclidean properties to retain maximum possible information in terms of minimum distance among pixels and their relevant angles.
The mentioned distance measures for identification of similarity among pixels based on pixel proximity provides different functionality features including extraction of information based on the 4-connected and 8-connected pixel neighborhood, and incorporation of geometric flows to keep track of angular movements of image pixels. However, none of these similarity measures provide balanced equation with integrated features of optimal boundary extraction based on connected neighbors and their angular movements. Thus, we hypothesize boundary extraction to be more accurate and intricate in presence of a similarity measure which provides greater information of spatial component provided by neighborhood pixels along with geometric flows.

3.5. SLIC++ Proposal

3.5.1. Euclidean Geodesic—Content-Aware Similarity Measure

Considering the simplicity and fast computation as critical components for segmentation, the proposed algorithm uses fusion of Euclidean and geodesic distance measures. The depiction of Euclidean and geodesic similarity is presented in Figure 1, where straight line shows Euclidean similarity while curved line shows geodesic similarity. Since using only Euclidean similarity loses the context information due to usage of straight-line distance and geodesic similarity focuses more on the actual possible path along the pixels. We propose the fusion of both the similarities to extract accurate information of image pixels and their associations.
Using the same logic as SLIC, we propose a normalized similarity measure. The normalization is based on the interval S between the pixel cluster centers. To provide the control over super-pixel same variable m is also used. Beforehand the contribution of Euclidean and geodesic distance in final similarity measure cannot be determined in terms of optimized performance. Hence, we have introduced two weight parameters for proposal of final similarity measure which based on weighted combination of Euclidean and geodesic distance. The proposed similarity measure is presented in Equation (10).
D c a = w 1 d 1 + m S d 2 + w 2 d 3 + m S d 4
where D c a is content-aware distance measure, d 1 and d 2 are same as d s and d c (Equations (5) and (6)) calculating the Euclidean distance for spatial and color component of image pixels. Variables d 3 and d 4 presents color and spatial component distance calculation using geodesic distance equation 8. Specifically, d 3 represents geodesic calculation of color components of image pixel and d 4 represents geodesic calculation of spatial components of pixel. Here, again we introduce similar normalization as of SLIC using variable S and m to provide control super-pixel compactness using geometric flows. The weights w 1 and w 2 further provides user control to choose the contribution of Euclidean and geodesic distance in final segmentation. These weights provide user flexibility, and these values can be changed based on the application. Moreover, these weights can be further tuned in future studies.

3.5.2. Proposal of Content-Aware Feature Infusion in SLIC

The SLIC++ is proposed to extract the optimal information from a visual scene captured in semi-dark scenarios. Nevertheless, the same algorithm holds for any type of image if the objective is to retrieve maximum information from the image space. The steps involved in computing super-pixels are written in SLIC++ algorithm (refer Algorithm 1). Basically, super-pixels are perceptual cluster computed based on pixel proximity and color intensities. Some of the parameters include: K being the number of super-pixels, N total number of pixels, A approximate number of pixels also called area of super-pixel, and S length of super-pixel.
Algorithm 1. SLIC++ Algorithm
1: Initialize K cluster center with seed ( s i ) i = 1 K defined at regular intervals S
2: Move cluster centers in n × n pixel neighborhood to lowest gradient
3: Repeat
4:      F o r each cluster center s i do
5:     Assign the pixel from 2 S × 2 S in a square window or CVT using distance measure given by Equation (3).
6:     Assign the pixel from 2 S × 2 S in a square window or CVT using distance measure given by Equation (10).
7:   End for
8: Compute new cluster centers and residual error ε r r   (distance between previous centers and recomputed centers).
9:  U n t i l   ε r r <= threshold
10: Enforce connectivity.
Keeping simplicity and fast computation intact we present SLIC++ algorithm, here only one of the steps mentioned on step 5 or 6 will be used. If step 5 is implemented, i.e., distance measure given by Equation (3) is used entire functionality of SLIC algorithm is implemented. Whereas, if step 6 is implemented, i.e., distance measure given by Equation (10) is used entire functionality of SLIC algorithm is implemented.
  • Initialization and Termination
For initialization a grid of initial point is created separated by distance S in each direction as seen in Figure 2. The number of initial centers is given as parameter K . Placement of initial center in restricted squared grids can result in error if the initial center is placed on the edge of image content. This initial center is termed as confused center. To overcome this error gradient of the image is computed and the cluster center is moved in the direction of minimum gradient. The gradient is computed with 4-neighboring pixels and the centroid is moved. To solve this mathematically L2 Norm distance is computed among four connected neighbors of center pixel. The gradient calculation is given by Equation (11).
G x , y = x + 1 , y x 1 , y 2 + x , y + 1 x , y 1 2
G x , y is the gradient of center pixel under consideration.
The gradient of the image pixels is calculated until stability where pixels stop changing the clusters based on the initialized clusters. Overall, the termination and optimization is controlled by parameter ‘n’ which represents number of iterations the overall SLIC algorithm goes through before finally resulting in super-pixel creation of the image. To keep the uniformity in presented research we have selected ‘n’ as 10 which has been a common practice [11,14,29,32].
How it works?
The incoming image is converted to CIELAB space. The user provides information of all the initialization parameters including ‘ K ’, ‘ m ’, ‘ n ’. Referring to the algorithm steps presented in SLIC++ algorithm. Step 1, places K number of super-pixels provided by user on an equidistant grid. This grid is created separated by S , where S is given by N K , N is total number of image pixels. Step 2 performs reallocation of initial seed takes places subjected to the gradient condition to overcome the effect of initial centers placed over the edge pixels in image. Step 3 through step 7, are iteratively executed till the image pixels stop changing the clusters based on the cluster centers/seeds. Steps 5 or 6 are chosen based for respective implementation of SLIC or SLIC++ vice versa. Step 5 and 6 basically performs clustering over the image pixels based on different distance measures. If user opts for SLIC then Euclidean distance measure is used (base functionality). If user opts for SLIC++ then proposed hybrid distance measure is used. Step 8 checks if the new cluster center after every iteration of clustering is different than the previous center (distance between previous centers and recomputed centers). Step 9 keeps track of the threshold value for iterations as specified by the parameter ‘n’. Step 10 enforces connectivity among the created super-pixel/clusters of image pixels.
The simple difference in the implementation of SLIC and SLIC++ lies in the usage of distance measure being used for the computation of image super-pixels. The presented research shows merely changing the distance measure to content-aware computational distance measure leads to better accuracy of results against the ground-truth for semi-dark images.
b.
Algorithm Complexity
The proposed algorithm follows the same steps as of Basic SLIC by introducing a new content-aware distance equation, thus the complexity of the proposed SLIC++ remains the same without any addition of new parameters, except the weights associated to the Euclidean and geodesic distance. These weights are merely scaler values to be taken into account in the core implementation of content-aware variant of SLIC, i.e., SLIC++. Hence, the complexity for the pixel manipulation is up to O N where N is the total number of image pixels. With the minimum possible imposed requirements of computation SLIC++ manages to find accurate balance of implementation with infused functionality of Euclidean and geodesic distance. This fusion results in optimal boundary detection verified in terms of precision-recall in Section 4.

4. Validation of the Proposed Algorithm

4.1. Experimental Setup and Implementation Details

Following the proposed algorithm and details of implementation scheme, SLIC++ is implemented in MATLAB. The benchmarking analysis and experiments are conducted in MATLAB workspace version R2020a using the core computer vision and machine learning toolboxes. For experiments, the semi-dark images of Berkeley dataset have been used. The reported experiments are conducted on processor with specs core i7 10750H CPU, 16 GB RAM and 64-bit operating system.
The images are extracted form a folder using Fullfile method then incoming RGB images are converted in CIELAB space. After that parameter initialization takes place to get the algorithm started. Based on the number of K seed are initialized on the CIELAB image space and the condition relating to the gradient is checked using several different built-in methods. After that each pixel is processed using the proposed similarity measure and super-pixels are created until the threshold specified by user is reached. Similarly, the performance of reported state-of-the-arts is checked using the same environmental setup using the relevant parameters. Finally, the reported boundary performance is reported in form of precision recall measure to check the boundary adherence of super-pixel methods including Meanshift, SLIC and SLIC++. For analysis in terms of precision recall bfscore method is used which takes in the segmented image, ground-truth image and compares the extracted boundary with the ground-truth boundary by returning parameters precision, recall and score.

4.2. Parameter Selection

In this section we introduce the parameter associated with Meanshift, SLIC and SLIC++. Starting off with the proposed algorithm, SLIC++ uses several parameters as of Basic SLIC. Scaling factor m is set to 10, threshold on the iteration is set to value 10 represented by variable n and parameter S is computed based on N number of image pixels divided by user defined number of super-pixels in terms of variable K. The variable K provides user control for the number of super-pixels. Compact super-pixels are created as the value of K is increased but it increases the computational overhead. We have reported the performance using four different set of values of K, i.e., 500, 1000, 1500 and 2000. All these parameters including m, n and K are kept same as for the basic SLIC experiments. However, there are some additional parameters associated with SLIC++ which are w 1 and w 2 and their values are set to 0.3175 and 0.6825, respectively. The weights are cautiously picked based on trial-and-error experimentation procedure. The images were tested for a range of different weights. The weight values were varied to have weight ratios including 10:90, 30:70, 50:50, 70:30, and 90:10 for Euclidean and Geodesic distance, respectively. The ratio of 30:70 retains empirically maximum and perceptually meaningful super-pixels resulting in the optimal performance against the ground-truth. For Meanshift implementation the bandwidth parameter is set to 16 and 32, keeping rest of the implementation parameter default. Table 2 shows the averaged performance of the proposed SLIC++ algorithm acquired by varying different values of weights for random test cases.
Empirically optimized performance of SLIC++ over 30:70 weight ratio for Euclidean and Geodesic distance hybridization has been tabulated in Table 2 row number 4, 9, 14 (formatted bold and italics). Moreover, the parameter values have been set as K = 500, m = 10 and n = 10 for the conducted experiments.

4.3. Performance Analysis

For performance analysis we considered two different experimental setups including qualitative analysis and quantitative analysis. Initially, we extended and analyzed the performance of SLIC with different distance measures to propose the most relevant distance measure for optimal boundary extension in semi-dark images. Then we compare the proposed algorithm with state-of-the-art super-pixel segmentation algorithms. The detail of the analysis is presented in following sub-sections.

4.3.1. Numeric Analysis of SLIC Extension with Different Distance Measures

For the detailed analysis of the proposed algorithm, we first compare the performance of basic SLIC with the variants of SLIC+ proposed in this study. The evaluation is presented in form of precision recall. For the optimal boundary detection greater values of precision are required. High precision rates relate to low number of false positives eventually resulting in high chance of accurate boundary retrieval, whereas high recall rates are relevant to matching of ground-truth boundaries to segmented boundary. Mathematically, precision is probability of valid results and recall is probability of detected ground-truths data [42]. For analysis of image segmentation modules, both high precision and recall are required to ensure maximum information retrieval [45].
Table 3 shows performance analysis of basic SLIC and its variants over randomly picked semi-dark images.
Table 2 depicts all the extension of SLIC perform better in terms of precision-recall. The parameters are kept uniform for all the experiments specifically parameter m and n as in SLIC [11]. Moreover, there is up to 3–9% gain in precision percentage using SLIC++ as compared to the basic SLIC algorithm. The relevant scores based on precision and recall also shoot up by margin of 5–9% using SLIC++ (row 1 vs. 6 and row 7 vs. 12). However, the performance of other variants of SLIC is subjective to dimensions of incoming data, magnitudes, and memory overload. There usually is no defined consensus regarding best generalized performer in terms of similarity measure so far [44]. Thus, we propose an integration of two similarity measures which takes into account minimal processing resources and still provides optimal boundary detection.
For further detailed qualitative analysis using the same test cases by changing the number of super-pixels we extend the analysis of SLIC versus SLIC++. The precision recall and score graphs are shown in Figure 3.
In Figure 3, solid lines represents performance of SLIC and SLIC++ for Test case 1 and dashed lines represents performance for Test case 2. Figure 3a shows precision curves of SLIC++ are substantially better than the SLIC presented by brown (dark and light) lines for test case 1 and 2, respectively. Figure 3b shows the SLIC++ recall is less the resulting recall of SLIC for the same images. Subsequently, based on the precision, recall and the final scores SLIC++ outperforms basic SLIC on semi-dark images. For number of pixels set to 1000 there is a drop observed in precision and recall of SLIC++, this behavior can be attributed to accuracy measure’s intolerance, i.e., even mutual refinements may result in low precision and recall values [45]. Nevertheless, performance for retrieval increases with increasing number of super-pixels and SLIC++ outperforms SLIC up to margin of 10%.

4.3.2. Comparative Analysis with State-of-the-Art

For the benchmarking of SLIC++ two different algorithms, i.e., SLIC and Meanshift are considered. To investigate the performance of SLIC and SLIC++ for the analysis over entire Berkeley dataset (semi-dark images), we set the number of super-pixels to 1500. The number of super-pixels is chosen 1500 because the peak performance of both the algorithms in experiment for test case 1 and 2 (refer Figure 3) is achieved by setting this parameter to value 1500. For the comparative analysis we also used Meanshift algorithm with input parameter, i.e., bandwidth set to 32. The bandwidth of Meanshift decides the complexity of the algorithm as this value is decreased the segmentation becomes more intricate with the overhead of computational complexity. To maintain computational resources throughout the experiment and keeping it uniform the parameters are chosen. The summary statistics of the obtained super-pixel segmentation results are shown in Table 4. The numerals presented in table are averaged values of precision, recall, and scores obtained for 316 images separately. The average precision, recall, and scores are presented in Table 4.
Table 4 shows SLIC++ achieves average percentage score up to 54%, whereas SLIC maintains a score of 47%. Finally, Meanshift achieves a score of 55%, which is greater than SLIC++ but as stated earlier for segmentation application greater values of precision and recall are required. So, comparing the recall of SLIC++ versus Meanshift a huge difference is observed. This difference is in terms of low recall of Meanshift which means algorithm fails to capture salient image structure [45] which is not desired for semi-dark image segmentation.

4.3.3. Boundary Precision Visualization against Ground-Truth

To validate the point-of-view relating to high precision and high recall we present perceptual results of Meanshift, SLIC and SLIC++. Notice that, the high precision means the algorithm has retrieved most of the boundary as presented by the ground-truth, whereas high recall means most of the salient structural information is retrieved from the visual scene. Meanshift resulted in a minimum recall, which hypothetically means the structural information was lost. Table 5 presents how Meanshift, SLIC, and SLIC++ performed in terms of perceptual results for visual information retrieval. The reported results are for parameters K = 1500 for SLIC and SLIC++ and bandwidth = 32 for Meanshift.
As super-pixels are not just about the boundary detection, resulting applications also expect the structural information present in the visual scene. Consequently, we are not just interested in the object boundaries but also the small structural information present in the visual image specifically semi-dark images. Table 5 shows SLIC and SLIC++ not only retrieves boundaries correctly with minimal computational power consumed but also retrieves the structural information. Column 4 shows the fact by mapping prediction over ground-truth image. For test case 1, in column 4 row id 3 Meanshift fails to extract the structural information as few green lines are observed. Whereas, for the same image SLIC and SLIC++ perform better as a lot of green textured lines are observed (refer column 4 row id 1 and 2). Meanwhile for test case 2, all three algorithms perform equally likely. Similar performance is observed with test case 3, SLIC and SLIC++ retains structural information better than Meanshift. Since Meanshift resulted in minimum recall over the entire semi-dark Berkeley dataset (refer Table 4) it does not qualify to be a good fit for super-pixel segmentation. The reason is less reliability of structural information fetching and its performance is highly subjective to the incoming input images.

4.3.4. Visualizing Super-Pixels on Images

For one more layer of subjective analysis of super-pixel performance we present super-pixel masks in this section. Initially, we present the input image in Figure 4 with the highlighted boxes to look closely for retrieval of structural information from the image. Here, the red box shows the texture information present on the hill whereas the green box shows water flowing in a very dark region of the semi-dark image.
Using the input image presented in Figure 4, we conducted experiments by changing the initialization parameters of all three algorithms. Table 5 shows the perceptual analysis visualizing the retrieval of salient structural information.
Table 6 shows that Meanshift extracts the boundaries correctly, whereas it loses all the contextual information when the bandwidth parameter is set to 32. This loss of information is attributed to low recall scores, whereas decreasing the value of the bandwidth increases the computational complexity and at the cost of additional complexity Meanshift now retrieves contextual information. SLIC and SLIC++ with minimal computational power retains structural information as seen in the red and green boxes in rows 1 and 2 of Table 5. Moreover, as the number of super-pixels ‘K’ increases, better and greater structural information retrieval is observed.
Figure 5 shows a zoomed in view of the super-pixels created by SLIC and SLIC++, residing in the red box. Here, we can see that SLIC++ retrieves content-aware information and SLIC ends up creating circular super-pixels (Figure 5a) due to the content irrelevant distance measure being used in its operational discourse.

Key Takeaways

The benchmarking analysis shows that the proposed algorithm SLIC++ achieves robust performance over different cases. The results of SLIC++ are more predictable as compared to the state-of-the-art methods Meanshift and SLIC. The performance of Meanshift is highly subjective as the recall keeps changing. Less recall values eventually result in less scores at the cost of information loss. Whereas, SLIC achieves 7% less scores and 8% less precision values in terms of boundary retrieval. The results of SLIC++ indicate that the proposed content-aware distance measure integrated in base SLIC demonstrates superior results. The significant improvement to the existing knowledge of super-pixel creation research is hybridization of proximity measures. Based on the comprehensive research it is seen that the hybrid measure performs better than the singular proximity measure counterparts of the same algorithm. These measures substantially control the end results of super-pixel segmentation in terms of accuracy. The proposed hybrid proximity measure carefully finds a balance between the two existing distance measure and performs clustering over image pixels making sure to retain content-aware information.

5. Limitations of Content-Aware SLIC++ Super-Pixels

The super-pixel segmentation algorithms are considered pre-processing step for wide range of computer vision applications. To obtain the optimal performance of sophisticated applications, the base super-pixel algorithm SLIC uses set of input parameters. These parameters allow the user control over different aspects of image segmentation. The idea is to extract uniform super-pixels throughout the image grid to maintain reliable learning statistics throughout the process. To make this possible the SLIC initially allows user to choose number of pixels ‘K’ (values ranging from 500–2000), parameter ‘m’ (where m = 10) which decides the extent of enforcement of color similarity over spatial similarity, number of iterations ‘N’ (where N = 10) which decides the convergence of the algorithm, neighborhood window ‘w’ (where w = 3) for gradient calculation to relocate cluster centers (if placed on edge pixel). This makes four input parameters for the base SLIC, whereas the proposed extension SLIC++ introduces two more weight parameters, w 1 and w 2 (0.3175 and 0.6825, respectively), to decide the impact of each distance measure in the hybrid distance measure. All these parameters significantly control the accuracy of segmentation results. Incorrect selection of these parameters leads to overall poor performance. Hence, for diverse applications, initial parameter search is necessary, which in turn requires several runs. For the reported research, using the state-of-the-art segmentation dataset, i.e., Berkeley dataset we chose the parameters as selected by the base SLIC. These parameters offer good performance over the image size of 321 × 481 or 481 × 321, whereas, as we increased the resolution of images during the extended research we observed that a higher value of ‘K’ is required for better segmentation accuracy.
For the existing research, we conducted experiments focused to identify the gains associated with usage of the proposed content-aware distance measure over the straight line distance measure. For the extended research, the input parameters shall be considered for optimization.

6. Emerging Successes and Practical Implications

Several decades of research in computer vision for boosted implementations resulting in fast accurate decisions, super-pixels have been a topic of research for long time now. The super-pixel segmentation is taken as entry stage providing pre-processing functionality for sophisticated intelligent workflows such as semantic image segmentation. To speed up the overall process of training and testing of these intelligent systems super-pixels are probable to provide remedies. As the intelligent automated vision systems have critical applications in medicine [46,47], manufacturing [48], surveillance [49], tracking [2] and so on. For this reason, fast and accurate visual decision are required. As the environmental conditions in form of visual dynamicity is challenging task to tackle by pre-processing modules. These modules are required to provide reliable visual results. Many super-pixel creation algorithms have been proposed over time to solve focused problems of image content sparsity [30], initialization optimization [28], and accurate edge detection [38]. However, the topic of the lightning condition in this domain remains untouched and needs attention. The dynamic lightning condition is a key component in autonomous vehicles, autonomous robots, surgical robots. The Berkeley dataset is comprised of images of different objects, ranging from humans, flowers, mountains, animals and so on. The conducted research holds for applications of autonomous robots and autonomous vehicles. However, the proposed algorithm is backed by the core concepts of image segmentation. For this reason, the presented work can be extended for variety of applications. Depending on the nature of application, the ranges of input parameters would be changed based on the required sensitivity of the end results, such as for the segmentation application in the medical domain compact where content-aware information is required. Consequently, the input values including the number of super-pixels to be created will be carefully selected. To handle the pre-processing problems associated with dynamic lightning conditions focusing autonomous robotics, the proposed extension of SLIC is a good fit. SLIC++ imposing minimum prerequisite conditions provides direct control over the working functionality and still manages to provide optimal information retrieval from the visual scenes for not only normal images but rather inclusive of semi-dark images.

7. Conclusions and Future Work

7.1. Conclusions

In this paper, we introduced a content-aware similarity measure which not only solved the problem of boundary retrieval in semi-dark images but is also applicable to other image types such as bright and dark. The content-aware measure is integrated in SLIC to create content-aware super-pixels which can then be used by other automated applications for fast implementations of high-level vision task. We also report results of integration of SLIC with existing similarity measures and describe their limitations of applicability for visual image data. To validate out proposed algorithm along with the proposed similarity measure, we conduct qualitative and quantitative evaluations on semi-dark images extracted from Berkeley dataset. We also compare SLIC++ with state-of-the-art super-pixel algorithms. Our comparisons show that the SLIC++ outperforms the existing super-pixel algorithms in terms of precision and score values by a margin of 8% and 7%, respectively. Perceptual experimental results also confirm that the proposed extension of SLIC, i.e., SLIC++ backed by content-aware distance measure outperforms the existing super-pixel creation methods. Moreover, SLIC++ results in consistent and reliable performance for different test image cases characterizing a generic workflow for super-pixel creation.

7.2. Future Work

For the extended research, density-based similarity measures integrated with content-aware nature may lead the future analysis. The density-based feature is expected to aid the overall all working functionality with noise resistant properties against the noisy incoming image data. Moreover, another aspect is the creation of accurate super-pixels in the presence of non-linearly separable data properties. Finally, the input parameter selection for the initialization of SLIC variants depending on the application domain and incoming image size remains an open area of research.

Author Contributions

Project administration, supervision, conceptualization, formal analysis, investigation, review, M.A.H.; conceptualization, methodology, analysis, visualization, investigation, writing—original draft preparation, data preparation and curation, M.M.M.; project administration, resources, review, K.R.; funding acquisition, conceptualization, review, S.H.A.; methodology, software, analysis, visualization, S.S.R.; validation, writing—review and editing, M.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research study is a part of the funded project under a matching grant scheme supported by Iqra University, Pakistan, and Universiti Teknologi PETRONAS (UTP), Malaysia. Grant number: 015MEO-227.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Following dataset is used. Berkeley: https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/ (accessed on 27 September 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Al-Huda, Z.; Peng, B.; Yang, Y.; Algburi, R.N. Object Scale Selection of Hierarchical Image Segmentation with Deep Seeds. Int. J. Pattern Recogn. Artif. Intell. 2021, 35, 2154026. [Google Scholar] [CrossRef]
  2. Riera, L.; Ozcan, K.; Merickel, J.; Rizzo, M.; Sarkar, S.; Sharma, A. Detecting and Tracking Unsafe Lane Departure Events for Predicting Driver Safety in Challenging Naturalistic Driving Data. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 238–245. [Google Scholar]
  3. Nadeem, A.; Jalal, A.; Kim, K. Automatic human posture estimation for sport activity recognition with robust body parts detection and entropy markov model. Multimed. Tools Appl. 2021, 80, 21465–21498. [Google Scholar] [CrossRef]
  4. Rashid, M.; Khan, M.A.; Alhaisoni, M.; Wang, S.-H.; Naqvi, S.R.; Rehman, A.; Saba, T. A sustainable deep learning framework for object recognition using multi-layers deep features fusion and selection. Sustainability 2020, 12, 5037. [Google Scholar] [CrossRef]
  5. Jaffari, R.; Hashmani, M.A.; Reyes-Aldasoro, C.C. A Novel Focal Phi Loss for Power Line Segmentation with Auxiliary Classifier U-Net. Sensors 2021, 21, 2803. [Google Scholar] [CrossRef] [PubMed]
  6. Memon, M.M.; Hashmani, M.A.; Rizvi, S.S.H. Novel Content Aware Pixel Abstraction for Image Semantic Segmentation. J. Hunan Univ. Nat. Sci. 2020, 47, 9. [Google Scholar]
  7. Tian, Z.; Shen, C.; Chen, H. Conditional convolutions for instance segmentation. In Proceedings of the Computer Vision–ECCV 2020 16th European Conference, Glasgow, UK, 23–28 August 2020; Part I 16; pp. 282–298. [Google Scholar]
  8. Junejo, A.Z.; Memon, S.A.; Memon, I.Z.; Talpur, S. Brain Tumor Segmentation Using 3D Magnetic Resonance Imaging Scans. In Proceedings of the 2018 1st International Conference on Advanced Research in Engineering Sciences (ARES), Dubai, United Arab Emirates, 15 June 2018; pp. 1–6. [Google Scholar]
  9. Hamadi, A.B. Interactive Automation of COVID-19 Classification through X-Ray Images using Machine Learning. J. Indep. Stud. Res. Comput. 2021, 18. [Google Scholar] [CrossRef]
  10. Zhao, L.; Li, Z.; Men, C.; Liu, Y. Superpixels extracted via region fusion with boundary constraint. J. Vis. Commun. Image Represent. 2020, 66, 102743. [Google Scholar] [CrossRef]
  11. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. Slic Superpixels; EPFL Technical Report 149300; School of Computer and Communication Sciences, Ecole Polytechnique Fedrale de Lausanne: Lausanne, Switzerland, 2010; pp. 1–15. [Google Scholar]
  12. Kim, K.-S.; Zhang, D.; Kang, M.-C.; Ko, S.-J. Improved simple linear iterative clustering superpixels. In Proceedings of the 2013 IEEE International Symposium on Consumer Electronics (ISCE), Hsinchu, Taiwan, 3–6 June 2013; pp. 259–260. [Google Scholar]
  13. Wang, M.; Liu, X.; Gao, Y.; Ma, X.; Soomro, N.Q. Superpixel segmentation: A benchmark. Signal Process. Image Commun. 2017, 56, 28–39. [Google Scholar] [CrossRef]
  14. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [Green Version]
  15. Liu, Y.-J.; Yu, C.-C.; Yu, M.-J.; He, Y. Manifold SLIC: A fast method to compute content-sensitive superpixels. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 651–659. [Google Scholar]
  16. Wang, P.; Zeng, G.; Gan, R.; Wang, J.; Zha, H. Structure-sensitive superpixels via geodesic distance. Int. J. Comput. Vis. 2013, 103, 1–21. [Google Scholar] [CrossRef] [Green Version]
  17. Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef] [Green Version]
  18. Sheikh, Y.A.; Khan, E.A.; Kanade, T. Mode-seeking by medoidshifts. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2017; pp. 1–8. [Google Scholar]
  19. Vedaldi, A.; Soatto, S. Quick shift and kernel methods for mode seeking. In Proceedings of the European Conference on Computer Vision; ECCV 2008; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5305. [Google Scholar] [CrossRef] [Green Version]
  20. Levinshtein, A.; Stere, A.; Kutulakos, K.N.; Fleet, D.J.; Dickinson, S.J.; Siddiqi, K. Turbopixels: Fast superpixels using geometric flows. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2290–2297. [Google Scholar] [CrossRef] [Green Version]
  21. Moore, A.P.; Prince, S.J.; Warrell, J.; Mohammed, U.; Jones, G. Scene shape priors for superpixel segmentation. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 771–778. [Google Scholar]
  22. Veksler, O.; Boykov, Y.; Mehrani, P. Superpixels and supervoxels in an energy optimization framework. In Proceedings of the European Conference on Computer Vision; ECCV 2010; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6315. [Google Scholar] [CrossRef] [Green Version]
  23. Van den Bergh, M.; Boix, X.; Roig, G.; de Capitani, B.; Van Gool, L. Seeds: Superpixels extracted via energy-driven sampling. Int. J. Comput. Vis. 2012, 111, 298–314. [Google Scholar] [CrossRef]
  24. Weikersdorfer, D.; Schick, A.; Cremers, D. Depth-adaptive supervoxels for RGB-D video segmentation. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 2708–2712. [Google Scholar]
  25. Conrad, C.; Mertz, M.; Mester, R. Contour-relaxed superpixels. In Proceedings of the International Workshop on Energy Minimization Methods in Computer Vision and Pattern Recognition; EMMCVPR 2013; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany; Germany, 2013; Volume 8081. [Google Scholar] [CrossRef]
  26. Xu, L.; Zeng, L.; Wang, Z. Saliency-based superpixels. Signal Image Video Process. 2014, 8, 181–190. [Google Scholar] [CrossRef]
  27. Li, Z.; Chen, J. Superpixel segmentation using linear spectral clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 12 June 2015; pp. 1356–1363. [Google Scholar]
  28. Rubio, A.; Yu, L.; Simo-Serra, E.; Moreno-Noguer, F. BASS: Boundary-aware superpixel segmentation. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2824–2829. [Google Scholar]
  29. Wang, H.; Peng, X.; Xiao, X.; Liu, Y. BSLIC: Slic superpixels based on boundary term. Symmetry 2017, 9, 31. [Google Scholar] [CrossRef] [Green Version]
  30. Liu, Y.-J.; Yu, M.; Li, B.-J.; He, Y. Intrinsic manifold SLIC: A simple and efficient method for computing content-sensitive superpixels. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 653–666. [Google Scholar] [CrossRef]
  31. Akyilmaz, E.; Leloglu, U.M. Similarity ratio based adaptive Mahalanobis distance algorithm to generate SAR superpixels. Can. J. Remote Sens. 2017, 43, 569–581. [Google Scholar] [CrossRef]
  32. Lowekamp, B.C.; Chen, D.T.; Yaniv, Z.; Yoo, T.S. Scalable Simple Linear Iterative Clustering (SSLIC) Using a Generic and Parallel Approach. arXiv 2018, arXiv:1806.08741. [Google Scholar]
  33. Chuchvara, A.; Gotchev, A. Content-Adaptive Superpixel Segmentation Via Image Transformation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1505–1509. [Google Scholar]
  34. Uziel, R.; Ronen, M.; Freifeld, O. Bayesian adaptive superpixel segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8470–8479. [Google Scholar] [CrossRef]
  35. Yang, F.; Sun, Q.; Jin, H.; Zhou, Z. Superpixel Segmentation With Fully Convolutional Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13961–13970. [Google Scholar]
  36. Wu, J.; Liu, C.; Li, B. Texture-aware and structure-preserving superpixel segmentation. Comput. Graph. 2021, 94, 152–163. [Google Scholar] [CrossRef]
  37. Chuchvara, A.; Gotchev, A. Efficient Image-Warping Framework for Content-Adaptive Superpixels Generation. IEEE Signal Process. Lett. 2021, 28, 1948–1952. [Google Scholar] [CrossRef]
  38. Yu, Y.; Yang, Y.; Liu, K. Edge-Aware Superpixel Segmentation with Unsupervised Convolutional Neural Networks. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1504–1508. [Google Scholar]
  39. Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 898–916. [Google Scholar] [CrossRef] [Green Version]
  40. Memon, M.M.; Hashmani, M.A.; Junejo, A.Z.; Rizvi, S.S.; Arain, A.A. A Novel Luminance-Based Algorithm for Classification of Semi-Dark Images. Appl. Sci. 2021, 11, 8694. [Google Scholar] [CrossRef]
  41. Stutz, D.; Hermans, A.; Leibe, B. Superpixels: An evaluation of the state-of-the-art. Comput. Vis. Image Underst. 2018, 166, 1–27. [Google Scholar] [CrossRef] [Green Version]
  42. Neubert, P.; Protzel, P. Superpixel benchmark and comparison. J. Comput. Phys. 1983, 51, 191–207. [Google Scholar]
  43. Tanemura, M.; Ogawa, T.; Ogita, N. A new algorithm for three-dimensional Voronoi tessellation. J. Comput. Phys. 1983, 51, 191–207. [Google Scholar] [CrossRef]
  44. Yu, J.; Kim, S.B. Density-based geodesic distance for identifying the noisy and nonlinear clusters. Inf. Sci. 2016, 360, 231–243. [Google Scholar] [CrossRef]
  45. Monteiro, F.C.; Campilho, A.C. Performance evaluation of image segmentation. In Image Analysis and Recognition. ICIAR 2006; Lecture Notes in Computer Science; Campilho, A., Kamel, M.S., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4141. [Google Scholar] [CrossRef] [Green Version]
  46. Gauriau, R. Shape-Based Approaches for Fast Multi-Organ Localization and Segmentation in 3D Medical Images; Telecom ParisTech: Paris, France, 2015. [Google Scholar]
  47. Dai, W.; Dong, N.; Wang, Z.; Liang, X.; Zhang, H.; Xing, E.P. Scan: Structure correcting adversarial network for organ segmentation in chest x-rays. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 263–273. [Google Scholar]
  48. Khan, A.; Mineo, C.; Dobie, G.; Macleod, C.; Pierce, G. Vision guided robotic inspection for parts in manufacturing and remanufacturing industry. J. Remanuf. 2021, 11, 49–70. [Google Scholar] [CrossRef]
  49. Fedorov, A.; Nikolskaia, K.; Ivanov, S.; Shepelev, V.; Minbaleev, A. Traffic flow estimation with data from a video surveillance camera. J. Big Data 2019, 6, 1–15. [Google Scholar] [CrossRef]
Figure 1. Irrelevance of Euclidean distance measure for super-pixel creation relating to image content.
Figure 1. Irrelevance of Euclidean distance measure for super-pixel creation relating to image content.
Sensors 22 00906 g001
Figure 2. Restricted Image search area for super-pixel creation specified by input argument for image window under consideration [31].
Figure 2. Restricted Image search area for super-pixel creation specified by input argument for image window under consideration [31].
Sensors 22 00906 g002
Figure 3. SLIC v/s SLIC++ performance over different number of pixels: (a) precision values; (b) recall value; (c) score values.
Figure 3. SLIC v/s SLIC++ performance over different number of pixels: (a) precision values; (b) recall value; (c) score values.
Sensors 22 00906 g003
Figure 4. Input image with highlighted regions for detailed analysis.
Figure 4. Input image with highlighted regions for detailed analysis.
Sensors 22 00906 g004
Figure 5. Zoomed in view of test case image for content-aware super-pixel analysis created by SLIC++ (b) against SLIC (a).
Figure 5. Zoomed in view of test case image for content-aware super-pixel analysis created by SLIC++ (b) against SLIC (a).
Sensors 22 00906 g005
Table 1. Critical analysis of gradient-ascent super-pixel creation algorithms.
Table 1. Critical analysis of gradient-ascent super-pixel creation algorithms.
MethodComplexityPixel Manipulation StrategyDistance MeasureDatasetSemi-Dark Image MentionsYearRef.
Meanshift O N 2 Mode seeking to locate local maximaEuclideanNot mentioned2002[17]
Medoidshift O N 2 Approximates local gradient using weighted estimates of medoids.EuclideanNot mentioned2007[18]
Quickshift O d N 2 Parzen’s density estimation for pixel valuesNon-EuclideanCaltech-42008[19]
TurboPixel O N Geometric flows for limited pixelsGradient calculation for boundary pixels onlyBerkeley Dataset2009[20]
Scene Shape Super-pixel (SSP) O N 3 2   l o g N Shortest path manipulation with prior information of boundary.Probabilistic modeling plus Euclidean space manipulationDynamic road scenes. No explicit mentions of semi-dark images but we suspect presence of semi2009[21]
Compact Super-pixels (CS)-Approximation of distance between pixels and further optimization with graph cut methodsEuclidean3D images2010[22]
Compact Intensity Super-pixels-Same as CS, With added color constant information.Euclidean3D images2010[22]
SLIC O N Gradient optimization after every iterationEuclideanBerkeley Dataset2012[11]
SEEDS-Energy optimization for super-pixel is based on hill-climbing.EuclideanBerkeley Dataset2012[23]
Structure Sensitive Super-pixels O N Super-pixel densities are checked, and energy minimization is conformed.GeodesicBerkeley Dataset2013[16]
Depth-adaptive super-pixels O N Super-pixel density identification, followed by sampling and finally k-means to create final clustersEuclideanRGB-D dataset consisting of 11 images2013[24]
Contour Relaxed Super-pixels O N Uses pre-segmentation technique to create homogeneity constraintNot mentionedNot mentioned2013[25]
Saliency-based super-pixel O N Super-pixel creation followed by merging operator based on saliency.EuclideanNot mentioned2014[26]
Linear Spectral Clustering O N Two-fold pixel manipulation strategy of optimization based on graph and clustering based algorithms.EuclideanBerkeley Dataset2015[27]
Manifold SLIC O N Same as SLIC but with mapping over manifold.EuclideanBerkeley Dataset
(Random)
2016[15]
BASS (Boundary-Aware Super-pixel Segmentation) O N l o g N Extension of SLIC initially creates boundary then uses SLIC with different distance measures, along with optimization of initialization parameters.Euclidean + GeodesicFashionista,
Berkeley Segmentation Dataset (BSD), HorseSeg,
DogSeg, MSRA Salient Object Database, Complex Scene Saliency Dataset (CSSD) and Extended CSSD
(ECSSD)
2016[28]
BSLIC O 6 N Extension of SLIC initializes seed within a hexagonal space rather than squareEuclideanBerkeley Dataset2017[29]
Intrinsic Manifold SLIC O N Extension of Manifold SLIC with geodesic distance measureGeodesicBerkeley Dataset
(Random)
2017[30]
Similarity Ratio based Super-pixels O N Extension of SLIC. Proposes automatic scaling of coordinate axes.MahanlanobisSAR Image dataset2017[31]
Scalable SLIC O N Optimized initialization parameters such as ‘n’ number of super-pixels, focused research to parallelization of sequential implementation.Euclideancyrosection Visible Human Male
dataset
2018[32]
Content adaptive super-pixel segmentation O N 2 Work on prior transformation of image with highlighted edges created by edge filtersEuclidean (with graph-based transformation)Berkeley Dataset2019[33]
BASS (Bayesian Adaptive Super-Pixel Segmentation) O N Uses probabilistic methods to intelligently initialize the super-pixel seeds.EuclideanBerkeley Dataset 2019[34]
Super-pixel segmentation with fully convolutional networks O N N o .   o f   l a y e r s Attempts to use neural networks for automatic seed initialization over grid.EuclideanBerkeley Dataset, SceneFlow Dataset2020[35]
Texture-aware and structure preserving Super-pixels O N The seed initialization takes place in circular grid.Three different distance measure (without explicit details)Berkeley Dataset2021[36]
Efficient Image-Warping Framework for Content-Adaptive Super-pixels Generation O N Warping transform is used along with SLIC for creation of adaptive super-pixels.EuclideanBerkeley Dataset2021[37]
Edge aware super-pixel segmentation with unsupervised CNN O N N o .   o f   l a y e r s Edges are detected using unsupervised convolutional neural networks then passed to super-pixel segmentation algorithmsEntropy based clusteringBerkeley Dataset2021[38]
Table 2. Summary statistics of average performance of SLIC++ for varying weights.
Table 2. Summary statistics of average performance of SLIC++ for varying weights.
RowRatio w 1 (Euclidean) w 2 (Geodesic)PrecisionRecallScore
Test Case 1 (Image ID = 14037):
110:900.11230.88770.478820.889300.62248
270:300.68250.31750.388500.922100.54660
350:500.48630.51370.377800.930400.53740
430:700.31750.68250.488540.891240.63113
590:100.88770.11230.388400.873400.53770
Test Case 2 (Image ID = 26031):
610:900.11230.88770.216230.788080.33935
770:300.68250.31750.183700.827900.30070
850:500.48630.51370.189100.850000.31000
930:700.31750.68250.226610.865200.35920
1090:100.88770.11230.186500.790000.30220
Test Case 3 (Image ID = 108082):
1110:900.11230.88770.270230.898320.41548
1270:300.68250.31750.218400.821600.34510
1350:500.48630.51370.226400.868000.35920
1430:700.31750.68250.285470.916290.43532
1590:100.88770.11230.223600.794700.34900
Table 3. Performance analysis of SLIC extensions.
Table 3. Performance analysis of SLIC extensions.
RowKmnParametersScorePrecisionRecallDistance Measure
Test Case 1 (Image ID = 14037):
15001010 0.544300.391200.89390Euclidean—SLIC
25001010 0.612340.465630.89406Chessboard—SLIC+
35001010 0.597130.444070.91118Cosine—SLIC+
45001010 µ = 4 0.627920.473450.93199Min4—SLIC+
55001010 0.561280.437770.78186Geodesic—SLIC+
65001010 w 1 = 0.3175 w 2 = 0.6825 0.631130.488540.89124Euclidean Geodesic—SLIC++
Test Case 2 (Image ID = 26031):
75001010 0.304200.186900.81740Euclidean—SLIC
85001010 0.354540.220980.89623Chessboard—SLIC+
95001010 0.356980.223290.88957Cosine—SLIC+
105001010 µ = 4 0.340570.209590.90798Min4—SLIC+
115001010 0.337150.213690.79842Geodesic—SLIC+
125001010 w 1 = 0.3175 w 2 = 0.6825 0.35920.226610.86525Euclidean Geodesic—SLIC++
Test Case 3 (Image ID = 108082):
135001010 0.354100.227200.80260Euclidean—SLIC
145001010 0.420990.277200.87476Chessboard—SLIC+
155001010 0.383680.242510.91811Cosine—SLIC+
165001010 µ = 4 0.424650.276940.91004Min4—SLIC+
175001010 0.403820.267640.82216Geodesic—SLIC+
185001010 w 1 = 0.3175 w 2 = 0.6825 0.435320.285470.91629Euclidean Geodesic—SLIC++
Table 4. Summary statistics of average performance for Berkeley dataset.
Table 4. Summary statistics of average performance for Berkeley dataset.
AlgorithmScorePrecisionRecall
SLIC0.470200.316040.97719
SLIC++0.547990.397760.93470
Meanshift-320.557050.575730.68416
Table 5. Semi-dark perceptual results conforming boundary retrieval.
Table 5. Semi-dark perceptual results conforming boundary retrieval.
Row IDImageGroundtruth ImagePredictionPrediction Map Compared with Groundtruth
Test Case 1:
1 Sensors 22 00906 i001 Sensors 22 00906 i002 Sensors 22 00906 i003 Sensors 22 00906 i004
2 Sensors 22 00906 i005 Sensors 22 00906 i006 Sensors 22 00906 i007 Sensors 22 00906 i008
3 Sensors 22 00906 i009 Sensors 22 00906 i010 Sensors 22 00906 i011 Sensors 22 00906 i012
Test Case 2:
4 Sensors 22 00906 i013 Sensors 22 00906 i014 Sensors 22 00906 i015 Sensors 22 00906 i016
5 Sensors 22 00906 i017 Sensors 22 00906 i018 Sensors 22 00906 i019 Sensors 22 00906 i020
6 Sensors 22 00906 i021 Sensors 22 00906 i022 Sensors 22 00906 i023 Sensors 22 00906 i024
Test Case 3:
7 Sensors 22 00906 i025 Sensors 22 00906 i026 Sensors 22 00906 i027 Sensors 22 00906 i028
8 Sensors 22 00906 i029 Sensors 22 00906 i030 Sensors 22 00906 i031 Sensors 22 00906 i032
9 Sensors 22 00906 i033 Sensors 22 00906 i034 Sensors 22 00906 i035 Sensors 22 00906 i036
Table 6. Detailed perceptual analysis with increasing parameters.
Table 6. Detailed perceptual analysis with increasing parameters.
Number of Super-Pixels/Algorithm500100015002000
SLIC Sensors 22 00906 i037 Sensors 22 00906 i038 Sensors 22 00906 i039 Sensors 22 00906 i040
SLIC++ Sensors 22 00906 i041 Sensors 22 00906 i042 Sensors 22 00906 i043 Sensors 22 00906 i044
Bandwidth/
Meanshift
1632
Sensors 22 00906 i045 Sensors 22 00906 i046
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hashmani, M.A.; Memon, M.M.; Raza, K.; Adil, S.H.; Rizvi, S.S.; Umair, M. Content-Aware SLIC Super-Pixels for Semi-Dark Images (SLIC++). Sensors 2022, 22, 906. https://doi.org/10.3390/s22030906

AMA Style

Hashmani MA, Memon MM, Raza K, Adil SH, Rizvi SS, Umair M. Content-Aware SLIC Super-Pixels for Semi-Dark Images (SLIC++). Sensors. 2022; 22(3):906. https://doi.org/10.3390/s22030906

Chicago/Turabian Style

Hashmani, Manzoor Ahmed, Mehak Maqbool Memon, Kamran Raza, Syed Hasan Adil, Syed Sajjad Rizvi, and Muhammad Umair. 2022. "Content-Aware SLIC Super-Pixels for Semi-Dark Images (SLIC++)" Sensors 22, no. 3: 906. https://doi.org/10.3390/s22030906

APA Style

Hashmani, M. A., Memon, M. M., Raza, K., Adil, S. H., Rizvi, S. S., & Umair, M. (2022). Content-Aware SLIC Super-Pixels for Semi-Dark Images (SLIC++). Sensors, 22(3), 906. https://doi.org/10.3390/s22030906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop