Next Article in Journal
LiveDrive AI: A Pilot Study of a Machine Learning-Powered Diagnostic System for Real-Time, Non-Invasive Detection of Mild Cognitive Impairment
Previous Article in Journal
Analysis of Connectivity in Electromyography Signals to Examine Neural Correlations in the Activation of Lower Leg Muscles for Postural Stability: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Weakly Supervised Nuclei Segmentation with Point-Guided Attention and Self-Supervised Pseudo-Labeling

Institute of Electronic Information Engineering, Beihang University, 37 Xueyuan Road, Haidian District, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(1), 85; https://doi.org/10.3390/bioengineering12010085
Submission received: 20 December 2024 / Revised: 12 January 2025 / Accepted: 15 January 2025 / Published: 17 January 2025
(This article belongs to the Section Biosignal Processing)

Abstract

:
Due to the labor-intensive manual annotations for nuclei segmentation, point-supervised segmentation based on nuclei coordinate supervision has gained recognition in recent years. Despite great progress, two challenges hinder the performance of weakly supervised nuclei segmentation methods: (1) The stable and effective segmentation of adjacent cell nuclei remains an unresolved challenge. (2) Existing approaches rely solely on initial pseudo-labels generated from point annotations for training, and inaccurate labels may lead the model to assimilate a considerable amount of noise information, thereby diminishing performance. To address these issues, we propose a method based on center-point prediction and pseudo-label updating for precise nuclei segmentation. First, we devise a Gaussian kernel mechanism that employs multi-scale Gaussian masks for multi-branch center-point prediction. The generated center points are utilized by the segmentation module to facilitate the effective separation of adjacent nuclei. Next, we introduce a point-guided attention mechanism that concentrates the segmentation module’s attention around authentic point labels, reducing the noise impact caused by pseudo-labels. Finally, a label updating mechanism based on the exponential moving average (EMA) and k-means clustering is introduced to enhance the quality of pseudo-labels. The experimental results on three public datasets demonstrate that our approach has achieved state-of-the-art performance across multiple metrics. This method can significantly reduce annotation costs and reliance on clinical experts, facilitating large-scale dataset training and promoting the adoption of automated analysis in clinical applications.

1. Introduction

Histopathological images play a crucial role in pathological diagnosis [1]. By analyzing these images, pathologists can devise appropriate treatment plans and prognosis evaluations for patients [2,3]. As a computer-aided technique, nuclei segmentation plays a crucial role in pathological image analysis, such as revealing tumor differentiation status [4] and cellular heterogeneity [5] and evaluating the tumor microenvironment [6], including immune cell infiltration [7] and angiogenesis [8]. Additionally, compared to manual methods, it significantly reduces the costs of pathological image analysis, shortens the analysis time, and enhances the efficiency and accuracy of medical image processing. It also aids pathologists in better analyzing nuclear morphology and structure, enabling early disease detection, quantitative assessment, and scientific research [9,10,11].
Traditional nuclei segmentation methods primarily include level set methods [12], watershed algorithms [13,14], edge detection, clustering [15], and thresholding techniques. These methods typically require extensive manual pre-processing and post-processing steps, such as noise removal and morphological operations. Additionally, they have a strong dependency on parameter settings, making them difficult to adapt to complex nuclear shapes and variations, resulting in highly unstable predictions.
In recent years, deep learning methods have been widely applied in medical image processing due to their ability to efficiently handle large-scale data and exhibit strong generalization capabilities. Semantic segmentation networks [16], like U-Net [17], are used to localize and segment nuclei in images, achieving success in medical image segmentation with their straightforward and effective architectures. Micro-Net [18] builds on U-Net by incorporating multi-scale feature extraction to enhance the detection capability for nuclei of varying sizes. The Vision Transformer (ViT) [19] is one of the most significant achievements in recent years for visual tasks. It applies the Transformer model [20], originally designed for natural language processing, to the field of image analysis, challenging the dominance of convolutional neural networks (CNNs). It enables the use of attention mechanisms in image segmentation. Subsequently, numerous ViT-CNN hybrid networks, such as Ds-TransUNet [21] and CellDETR [22], have been applied to the field of cell detection. With the advent of the era of large models, general segmentation models like SAM [23] and OMG-Seg [24] are becoming the next focus for researchers. However, these semantic segmentation methods cannot differentiate between different instances of the same category, such as multiple cell nuclei within one image. In contrast, instance segmentation networks [25,26,27] employ more rigorous labeling to precisely annotate the boundaries of each cell nucleus, distinguishing closely adjacent or overlapping nuclei. However, fully supervised methods for cell nuclei instance segmentation rely heavily on extensive pixel-level annotation data, which are time-consuming to acquire and annotate, often requiring expert guidance. Additionally, these approaches may encounter overfitting issues when confronted with the intricate morphological and structural variations inherent in cell nuclei.
Compared to fully supervised methods for nuclei segmentation, weakly supervised approaches have gained attention due to their simplicity and cost-effectiveness in acquiring data labels. However, these methods sacrifice label accuracy. Effectively leveraging ambiguous annotation information to enhance model performance remains a major challenge in weakly supervised nuclei segmentation. Weakly supervised annotations commonly include box, point, and image-level annotations:
  • Box annotations simplify the annotation process by drawing a rectangular box around each instance. BoxInst [28] replaces the original pixelwise mask loss with projection loss and pairwise affinity loss, achieving promising results using only box annotations. Building on BoxInst, BoxTeacher [29] introduces a mask-aware confidence score to estimate the quality of pseudo-labels, along with noise-aware pixel loss and denoising affinity loss to adaptively optimize the pseudo-labels for the student network. Wang et al. developed a polar-transformation-based MIL strategy to improve segmentation with loose bounding box supervision [30]. Their method emphasizes pixels closer to the polar origin, achieving robust segmentation performance even with imprecise bounding boxes. These methods are suitable for objects with clear boundaries, but it is generally less effective for objects with complex shapes or those that are densely packed. Moreover, this method still requires relatively complex annotation efforts.
  • Image-level annotations are the weakest form of annotation in weakly supervised learning. Unlike other methods that describe specific regions or object details within the image, this approach only uses labels that describe the overall content of the image, for example, classification labels, or even labels distinguishing between positive and negative samples in an image. MICRA-Net [31] leverages latent information from the trained model and uses gradient class activation maps to generate detailed feature maps, reducing the need for expert annotations. Zhou et al. employed cyclic learning [32] based on multi-task learning, alternating between classification and semi-supervised tasks. Their method achieves performance close to fully supervised approaches and is compatible with different backbones and segmentation architectures. While image-level label-based methods are efficient for annotation, they do not provide spatial information or the boundary details of instances, which leads to suboptimal performance in tasks like nuclei segmentation.
  • Point annotations [33,34,35,36,37,38] not only facilitate easy label generation but also provide spatial location information of the nuclei, thus presenting an excellent alternative. Given that point annotations do not provide information about the boundaries of nuclei, existing methodologies leverage Voronoi diagrams and k-means clustering algorithms [39] to generate pseudo-labels (as illustrated in Figure 1). Notably, Qu’s method [40,41] successfully reduced the annotation time on a custom dataset from 114 min to 15 min. Some methods [35,42] employ the Sobel filter to generate pseudo-boundary maps for nuclei boundary refinement. Liu et al. [43] directly processed the segmentation prediction map for more stringent instance segmentation boundary labels, feeding them into an instance segmentation network for learning. SC-Net [44] uses a co-training strategy and self-supervised learning to improve segmentation accuracy, achieving strong results on MoNuSeg and CPM datasets. BoNuS [45] learns nuclei interior and boundary information through a boundary mining loss and detects missing nuclei using a curriculum-learning-based detection module.
However, the pseudo-labels generated from point annotations have compromised boundary accuracy, introducing noise that can negatively affect model training. Directly utilizing the generated probability map for subsequent supervision may amplify this noise. Furthermore, effectively separating adjacent nuclei with similar color and shape features poses a difficult task in nuclei segmentation (as shown in Figure 2). Some methods [45,46] seek to leverage spatial information from point annotations for instance separation through center-point prediction. However, as the direct use of point annotations for training is not feasible, a common approach is to generate Gaussian masks around point coordinates for training. Nevertheless, determining an optimal Gaussian kernel radius for center-point predictions becomes challenging due to variations in nuclei morphologies and sizes across different organs in training images. Additionally, it is hard to achieve stable predictions in single-branch center-point predictions, often resulting in false positive and false negative predictions, which is harmful to instance separation.
To address these challenges, this paper introduces a new framework that integrates three key components: the multi-scale Gaussian kernel module, the point-guided attention module, and the pseudo-label updating module. These components serve specific purposes: the multi-scale Gaussian kernel module distinguishes instances, the point-guided attention module reduces the impact of pseudo-label noise, and the pseudo-label updating module enhances the quality of pseudo-labels. Our contributions are summarized as follows:
  • To tackle the difficulties in setting the Gaussian kernel radius and the instability of center-point predictions, we introduce a multi-scale Gaussian kernel mechanism, which utilizes multiple branches with different radii to accurately predict center points to separate adjacent nuclei.
  • To mitigate noise from inaccurate pseudo-labels, we introduce a point-guided attention module. This enables the segmentation module to focus on learning features from the center-point prediction module, concentrating attention near true point labels.
  • In addition, we introduce a module that optimizes pseudo-label boundaries using the exponential moving average (EMA) and k-means clustering, which leverages the model’s historical training information, point label information, and the color information of the training images to enhance the quality of pseudo-labels, aiming to improve segmentation performance.
  • Our point-supervised approach has achieved state-of-the-art performance across various metrics in experiments conducted on three public datasets. Ablation studies have further validated the effectiveness of our modules.

2. Materials and Methods

In this section, the study is divided into three main modules, including the multi-scale Gaussian kernel module, the point-guided attention module, and the pseudo-label updating module. Figure 3 illustrates the framework proposed in this work. The entire model consists of the following modules:
  • Two types of decoders: (1) the segmentation decoder branch for nuclei region detection; (2) the Gaussian decoder branch for center-point prediction. The labels used by the segmentation branch are generated from point labels via the Voronoi algorithm and k-means algorithm. Since the boundaries of these labels are predicted, they are referred to as pseudo-labels. The labels used by the Gaussian branches are generated by applying Gaussian expansion to each point in the point labels. All decoders share a common backbone.
  • Multi-scale Gaussian kernel module: Multiple Gaussian branches are designed, each guided by a Gaussian mask with a different Gaussian expansion radius for training, with multi-branch predictions for center points. The center prediction map assists the segmentation branch in performing more accurate instance segmentation.
  • Pseudo-label updating module: During training, the pseudo-labels used by the segmentation branch are iteratively optimized by this module in conjunction with historical training information to reduce the boundary noise caused by pseudo-labels.
  • Point-guided attention module: The feature maps of each layer of the segmentation branch decoder pass through this module to learn features from the corresponding Gaussian branches. By focusing attention on areas near the real point labels, this module further reduces boundary noise.
Each module is described in detail in the subsequent sections.

2.1. Multi-Scale Gaussian Kernel Module

A multi-scale Gaussian kernel mechanism is designed to stabilize the prediction of center points. The multi-scale Gaussian kernel module comprises k Gaussian branches. For each branch, we utilize Gaussian masks with different radii as labels for center-point prediction. The formula for generating a Gaussian mask M i ( x , y ) is as follows:
M i x , y = exp d i 2 x , y 2 σ 2 if d x , y < r i , 0 otherwise ,
where σ denotes the Gaussian bandwidth, and  r i represents the Gaussian kernel radius in the i-th Gaussian branch. M i ( x , y ) and d i ( x , y ) correspond to the Gaussian mask and the distance from point ( x , y ) to the nearest center point in the same branch, respectively.
The Gaussian branch is trained using weighted Mean Squared Error Loss. The specific loss function formula is as follows:
L g a u s s = 1 k j = 1 k 1 Ω i Ω w i p i j M i j 2 ,
where Ω is the set of non-ignored pixels, and the variables p i j and M i j represent the pixel value at pixel i in the Gaussian prediction map and the Gauss label corresponding to the j-th Gaussian branch, respectively. w i is the weight of pixel i. Considering the imbalance between annotated points and background pixels, to enhance the influence factor of the foreground area and encourage more foreground predictions by the Gaussian branches, w i is set to 10 for pixels with a mask value greater than 0 and to 1 for background pixels. The final loss for the entire multi-scale Gaussian kernel module is obtained by averaging the loss of each branch.
The reason for choosing multi-radius and multi-branch approaches is that the sizes and morphologies of cells vary across different images, making it challenging to set a single appropriate Gaussian kernel radius. By employing multiple radii, the nuclei can adaptively find the most suitable radius for center-point prediction. Additionally, multi-branch predictions are more stable compared to single-branch predictions.
After obtaining the nuclear foreground probability map from the segmentation branch and the Gaussian prediction map from the Gaussian branch, a corresponding model inference strategy is designed. This strategy allows the predicted center points to fully guide and assist the segmentation network in performing more precise instance segmentation, effectively separating adjacent cells that are difficult to distinguish.
The specific strategy is illustrated in the testing phase part of Figure 3. From the Gaussian branch, we obtain k Gaussian prediction maps. Only pixels predicted as foreground by at least k / 2 Gaussian branches are considered foreground. The final foreground map will contain several connected components. Noise regions with an area of fewer than 20 pixels are removed. For the remaining connected components, we compute their centroid to obtain the initial center-point prediction map C. For the generated segmentation prediction map, each connected component is assigned a unique instance label to generate the initial segmentation prediction instance map S.
Subsequently, the Voronoi algorithm is employed to generate the Voronoi map using the predicted center points. The Voronoi diagram aims to divide the plane into several polygonal regions based on a set of seed points. Each region contains one generating point, and all points within that region are closer to the generating point than to any other point in other regions.
Given a set of generating points in a plane, a Voronoi cell is defined as
V ( p i ) = { x R 2 j i , d ( x , p i ) < d ( x , p j ) } ,
where d ( x , p i ) denotes the Euclidean distance between point x and point p i . Pixels in different regions of the Voronoi diagram P are marked with unique instance ids. Multiplying this map by the segmentation prediction instance map S produces the coarse instance map I. However, due to cell morphology variations and model prediction errors, not all nuclei are perfectly assigned to their respective Voronoi regions. To address this, we propose a model inference strategy for instance refinement, aiming to better align nuclei with their corresponding Voronoi regions.
Algorithm 1 outlines our model’s inference strategy. To ensure that each cell instance in the segmentation prediction map corresponds to a predicted centroid, we proceed as follows:
  • For a given segmentation instance, if no centroids are found within it, we identify this as a false negative in the Gaussian prediction. Consequently, a centroid is added to the center of this instance. As shown in Figure 4a,b, assuming S i represents instances originally predicted correctly, they have been erroneously segmented into 4 parts due to the lack of corresponding centroids. This segmentation error is evident. However, after adding centroids, the Voronoi diagram generated from the new center-point map effectively segments it into the correct shape.
  • In addition, Gaussian prediction centroids that fall within the background area of the segmentation prediction map are considered false positives and are discarded. The updated C is then used to generate P using the Voronoi algorithm.
  • For each instance S i in S, the Voronoi map P divides the instance into several blocks. The regions occupying less than 10% of S i ’s total area typically represent errors caused by inaccurate Voronoi partition (e.g., S i 2 , S i 3 , S i 4 , S i 5 in Figure 4c). These are merged into the largest neighboring sub-region. Centroids not corresponding to any instance are considered false positives and are removed. The updated C is used to regenerate P and I, iterating until each centroid corresponds to one instance in I.
Algorithm 1: Inference Strategy
Bioengineering 12 00085 i001

2.2. Pseudo-Label Updating Module

Since the model is trained using point annotations only, the Voronoi algorithm and k-means clustering are initially employed to generate the initial pseudo-labels for the segmentation branch. However, the quality of the labels’ boundaries generated by this method remains suboptimal. The pseudo-label updating module is illustrated in Figure 5. The entire pseudo-label updating strategy is divided into the training phase and the inference phase.
In the training phase, inspired by temporal ensembling [47] and Lin’s approach [44], we utilize the weight map to retain historical training information. Since the model has not yet learned sufficient features in the early stages of training, for the first 60 training epochs, we exclusively utilize the initially generated cluster labels for training (Voronoi labels are excluded from the label update process). The processed cluster label is a 3-value mask, where 0, 1, and 2 represent background, foreground, and ignored pixels, respectively. At the 60th epoch, the probability map p predicted by the model is used. This map is used to fill in the ignored pixels in the cluster labels K , generating the initial weight map W 0 . The value of the pixel ( x , y ) in the map, W 0 ( x , y ) , is generated as follows:
W 0 x , y = K x , y if K x , y = 0 or 1 , p x , y if K x , y = 2 .
During these steps, the exponential moving average (EMA) is incorporated for weight map updates. The EMA is chosen for pseudo-label updating because it gives more weight to recent predictions while smoothing out noise from earlier iterations. Compared to simple averaging or sliding window methods, the EMA better mitigates short-term fluctuations during updates, adapting progressively to dynamic changes in training. Additionally, its cumulative updating property enhances the stability and quality of pseudo-labels, which is crucial for improving model performance and reducing noise propagation in weakly supervised tasks. The formula can be expressed as follows:
W n = α p n + 1 α W n 1 ,
where p n represents the probability map predicted by the model in the n-th epoch. The new weight map W n is generated by weighting the previous weight map W n 1 . After thresholding, the resulting weight map is employed to guide subsequent training. However, to prevent amplifying errors caused by prediction inaccuracies, pixels ignored by K continue to be ignored. The labels are updated at regular intervals.
After training, during the inference phase, image color information and point annotation details are integrated. For each pixel x i , we combine the corresponding predicted probability map value p i with the RGB values ( r i , g i , b i ) of the image. Additionally, we include d i , generated by the distance transform [45] from point annotations. These values together form the feature vector f x i = r ^ i , g ^ i , b ^ i , d ^ i , p ^ i , which is used for k-means clustering:
a r g min S i = 1 k x i S j f x i c j 2 .
In this equation, k represents the number of clusters, c j denotes the center of the j-th cluster, and S j is the set of pixels assigned to the j-th cluster. The objective is to minimize the sum of squared distances between each pixel’s feature vector f x i and the corresponding cluster center c j . The 5 different variables are normalized or clipped to similar ranges to enhance clustering performance. The newly generated cluster labels will be used for the 2nd training cycle. The model produces its final results after the second iteration.

2.3. Point-Guided Attention Module

Although a pseudo-label update strategy was designed to improve the quality of pseudo-labels, boundary regions may still introduce noise, impacting model training. Therefore, a point-guided attention module is designed to focus the model on true point annotations, reducing the noise from pseudo-labels.
The specific framework is shown in Figure 6. The decoders for both the segmentation branch and the Gaussian branches are structurally identical, forming a U-shaped architecture [17] with the shared encoder. At each level, the segmentation branch feature map undergoes point-guided attention with the corresponding Gaussian feature maps. Subsequently, these feature maps are concatenated with the encoder’s feature map at the same level and upsampled to produce the feature map for the next level.
Similar to the approach in the Vision Transformer [19], the input feature maps are flattened into vectors for attention calculations. However, due to the large size of the feature maps, directly flattening them may result in high computational complexity. Therefore, the module splits the feature maps into 7 × 7 patches and performs point-guided attention on the corresponding patches from each branch before concatenating them. This patch-based approach enables parallel computation, thereby significantly enhancing the computational efficiency of the model. The mechanism for generating the Query, Key, and Value in the segmentation branch is as follows:
Q i = L Q j = 1 k g j m i k , K i = L K j = 1 k g j m i k ,
V i = L V S m i ,
S m i = s o f t m a x Q i K i T d k V i ,
where Q i , K i , and V i represent the Query, Key, and Value vectors for the i-th input, respectively, and k denotes the number of Gaussian branches. d k is the dimensionality of the Key vectors. S m i indicates the i-th part of the segmentation feature, g j m i indicates the j-th Gaussian feature. L Q · , L K · , and L V · denote linear projection for the input. To ensure the independence of predictions from each Gaussian branch, we exclusively perform self-attention calculations within each Gaussian branch, without sharing features with other branches.

2.4. Loss Function

The nuclei segmentation branch is trained jointly using both Voronoi labels and cluster labels. The corresponding losses, denoted by L v o r and L c l u , both employ a combination of Dice Loss and Cross-Entropy (CE) Loss:
L v o r , L c l u = L D i c e + L C E ,
L D i c e = 1 2 × i = 1 N p i × t i + ϵ i = 1 N p i 2 + i = 1 N t i 2 + ϵ ,
L C E = 1 N i = 1 N c = 1 C G i C log S i C ,
where for L dice , p i represents the model’s prediction, t i is the ground-truth label, N is the number of pixels in the sample, and ε is a small regularization term to avoid division by zero. For L CE , C denotes the class corresponding to pixel i, and G i C and S i C represent the binary label and its corresponding predicted probability, respectively.
The overall loss is derived from the combined losses of the segmentation branch and the Gaussian branches:
L t o t a l = λ 1 L v o r + ( 2 λ 1 ) L c l u + λ 2 L g a u s s ,
where λ 1 and λ 2 represent loss weights.

3. Results

3.1. Datasets

Three datasets were employed to evaluate our method: the Monuseg dataset [48], the CPM17 dataset [49], and the CoNSeP dataset [25].
Monuseg: The Multi-Organ Nuclei Segmentation Challenge dataset is a public dataset downloaded from The Cancer Genome Atlas (TCGA), comprising histopathological images from nine organs. The dataset consists of 30 training images and 14 testing images, each with a resolution of 1000 × 1000 pixels, and a total of 28,846 nuclei.
CPM17: The Computational Precision Medicine dataset released in 2017 includes 64 H&E-stained images, each with the size of 500 × 500 or 600 × 600 pixels, containing a total of 7570 nuclei. The dataset is divided into 32 training and 32 testing images. The training and test sets both consist of images from four cancer types: glioblastoma multiforme (GBM), low-grade glioma (LGG), head and neck squamous cell carcinoma (HNSCC), and non-small-cell lung cancer (NSCLC), with eight images per cancer type.
CoNSeP: The Colorectal Nuclear Segmentation and Phenotypes dataset consists of 41 H&E-stained image tiles, each with a size of 1000 × 1000 pixels and a magnification of 40×. The images contain a total of 24,319 nuclei annotated by pathologists. The dataset is divided into 27 training images and 14 test images.

3.2. Evaluation Metrics

The model was evaluated using five metrics: object-level Dice coefficient Dice o b j [50], Aggregated Jaccard Index (AJI) [51], detection quality (DQ), segmentation quality (SQ), and panoptic quality (PQ) [25]. It should be mentioned that the Dice o b j measures the similarity between predicted and ground-truth objects in segmentation, focusing on the accurate detection of individual objects like nuclei or cells. These five evaluation metrics assess different aspects of model performance. Specifically, AJI and DQ mainly measure nuclei localization, while Dice o b j and SQ are focused on nuclei segmentation. PQ offers a comprehensive evaluation of the model’s overall detection quality.

3.3. Implementation Details

ResNet34 [52] was utilized as the backbone, and the Adam optimizer was employed for training, with a learning rate and weight decay set to 1 × 10−4. The batch size was set to 8, and the model was trained for 120 epochs. The same augmentation operations as described in Qu’s method [40] were performed. Each training image was divided into 16 overlapping patches of size 250 × 250 pixels. These patches were then transformed (cropped, rotated, etc.) to obtain image patches of size 224 × 224 pixels, which were used for model training. During testing, each image was cropped into 224 × 224 pixel patches with an overlap of 80 pixels. Empirically, we employed four Gaussian branches with radii set to 7, 9, 11, and 13. These parameters were chosen because when the Gaussian kernel radius is smaller than 7, most values in the Gaussian mask approach zero, making it difficult for the model to learn effective features. Conversely, when the radius exceeds 13, adjacent Gaussian kernels begin to overlap, leading to reduced accuracy in center-point prediction. The Gaussian bandwidth was defined as one-third of the corresponding Gaussian kernel radius. In the pseudo-label update strategy, the EMA weight α was set to 0.1. The weight map was updated every 10 epochs. In the loss function, the weights λ 1 and λ 2 were set to 1. The model was trained using an NVIDIA GeForce RTX 3090 GPU with PyTorch version 1.8.0 and CUDA version 11.2.

3.4. Comparative Results

Table 1 presents the comparative experimental results of our method against the most popular or highest-performing point-supervised and fully supervised methods, which we used as baselines. Our method surpasses the state-of-the-art (SOTA) by 3.4% in AJI and 5.1% in DQ metrics on Monuseg. On CPM17, it outperforms the SOTA by 2.4% in AJI and 5.0% in DQ. In the CoNSeP dataset, our method exceeds the SOTA by 2.8% in AJI and 0.9% in DQ. This indicates that our method effectively distinguishes instances. In terms of the Dice o b j and PQ metrics, our model surpasses the state-of-the-art by 2.4% and 3.2% on the MoNuSeg dataset, by 1.2% and 3.2% on the CPM17 dataset, and by 1.6% and 0.5% on the CoNSeP dataset, respectively.
However, due to the absence of nuclei boundary information in the labels used by our approach, there is still a gap between our method and fully supervised approaches in segmentation metrics such as Dice o b j and SQ. This is particularly evident in datasets with complex cell morphology and dense distributions, such as CoNSeP. The model focuses more attention on the areas near each nucleus’s center, which leads to a slight decrease in boundary precision and a minor drop in the SQ metric. However, this significantly improves the model’s ability to separate densely packed nuclei, which is essential for accurate cell counting, making this trade-off acceptable. Although there remains a gap between our model and fully supervised methods, our approach still outperforms most point-supervised methods across the majority of metrics.
The visual comparison results with some point-supervised methods are presented in Figure 7. In the yellow-highlighted regions, other methods fail to effectively differentiate closely adjacent cells, whereas our method demonstrates superior performance in distinguishing instances and boundaries, significantly improving the overall segmentation performance of the model.
We also evaluated the nuclei segmentation performance across different organs. As shown in Table 2, our model achieves high metrics for organs such as the bladder, brain, and lung. However, for the colon and breast, where the cells are typically elongated and densely packed, the performance is relatively poor and does not reach the average across the entire test set.

3.5. Ablation Studies

3.5.1. Ablation Study on Three Modules

An ablation study is presented to validate the effectiveness of each module in our method. From Figure 8, it is evident that models B, D, E, and F, with the inclusion of the multi-scale Gaussian kernel module, exhibit the enhanced separation of adjacent cells. This improvement is particularly manifested in object-level metrics such as AJI and DQ in Table 3. The comparison between models A and C reveals that the introduction of the pseudo-label updating module results in improvements of 1.1%, 0.5%, and 4.0% in Dice o b j metrics on Monuseg, CPM17, and CoNSeP, respectively. There are also improvements in all other metrics. From the comparison between models B and D in Table 3, the introduction of point-guided attention leads to improvements of 2.3%, 1.0%, and 1.2% in AJI on Monuseg, CPM17, and CoNSeP, respectively. Finally, the combination of the three modules results in the optimal performance across most metrics.

3.5.2. Ablation Study on the Loss Function

In the loss function, λ 1 and λ 2 are used to balance the weights between the segmentation branch and the Gaussian branch, respectively. For the segmentation branch, we directly adopted the loss function from [45]. An ablation study was conducted on these parameters to illustrate how their values influence segmentation performance. The best results are achieved when both λ 1 and λ 2 are set to 1.
The results of varying one parameter while keeping the other at its optimal value are shown in Figure 9. The first row illustrates the impact of λ 1 on performance when λ 2 is set to 1. When λ 1 is too small, the model performs poorly because the Voronoi labels contain center-point information, and using only cluster labels fails to separate close nuclei. Conversely, with a large λ 1 , the model neglects boundary information from the cluster labels. The best performance occurs when λ 1 is around 1, so we use this value in subsequent experiments. The second row shows the effect of λ 2 when λ 1 is fixed at 1, with the best results in the range of 1 to 1.5. For simplicity, we use λ 2 = 1 .

3.5.3. Ablation Study on the Gaussian Branches

To investigate the impact of the multi-scale Gaussian kernel module on model prediction, the model was trained using 1 to 6 Gaussian branches separately. To prevent the Gaussian mask from being too small for training and the Gaussian kernels from merging when the radius is too large, integer values between 6 and 16 were selected for the Gaussian kernel radius in our experiments. As shown in Table 4, the experimental data for each branch were taken from the best results obtained with different Gaussian kernel radii. The metrics progressively improve as the number of Gaussian branches increases from 1 to around 4. However, as the number of branches continues to increase, the improvement is marginal. Consequently, we opt for using four Gaussian branches. It is worth noting that the kernel radius also affects the experiment, and more branches do not necessarily lead to better results. Nevertheless, a multi-branch and multi-scale design helps the model adapt to varying kernel parameters, reducing errors caused by inaccurate single predictions and enhancing overall robustness.
Figure 10 shows the impact of the number of Gaussian branches on model complexity, training time, and inference time. With each additional Gaussian branch, the model parameters and training time increase by approximately a factor of 1.1. In the inference phase, the inference time also grows by about 1.1 times. After the input images are processed by the model to produce outputs, additional time is required for instance partition of the output segmentation maps. During post-processing, we use a consolidated center prediction map from multiple Gaussian branches. The post-processing time is determined by the number of nuclei in the coarse instance predictions, rather than the number of Gaussian branches. This post-processing time accounts for approximately 70% of the total inference time.

3.5.4. Ablation Study on Inference Strategy

The impact of the inference strategy on model performance was investigated across all three datasets. The results are presented in Table 5. It can be observed that, despite the longer post-processing time, the model achieved improvements of 0.7%, 0.1%, and 0.9% in Dice o b j scores on Monuseg, CPM17, and CoNSeP datasets, respectively. It should be noted that the inference strategy did not directly alter the foreground region predicted by the segmentation map, thus yielding no direct positive impact on the SQ metric.
Table 6 displays the inference time of different methods on the Monuseg dataset. Weak-Anno [40], WSPP [45], and SC-Net [44] directly use semantic segmentation maps to compute connected components for instance maps during inference; consequently, the inference time is short. However, these methods cannot effectively distinguish adjacent instances. SPN+IEN [43] performs better than previous methods; however, its post-processing operations are excessively time-consuming. Thanks to the Transformer’s superior parallel computing capabilities, our approach achieves an inference time of 0.513 s, comparable to existing point-supervised methods for generating a coarse instance map. Generating a refined instance map incurs a notable increase in time (2.044 s). However, our method maintains an inference time comparable to common instance segmentation networks like HoverNet (1.977 s) and Mask-RCNN (2.773 s). Additionally, it delivers improved performance across various metrics.

3.6. Impact of Pseudo-Label Updating Strategy

Figure 11 illustrates the impact of the pseudo-label updating strategy. Original clustering labels exhibit low accuracy in delineating cell nuclei boundaries, with significant areas of ignored pixels leading to wasted label information. While the updated pseudo-labels still exhibit discrepancies compared to the ground truth, their boundary precision has significantly improved relative to the initial pseudo-labels. In addition, the extent of ignored areas has been substantially reduced, positively influencing model training.

4. Discussion

In this study, we introduce a nuclei segmentation method based on point supervision, aiming to achieve a performance close to that of fully supervised instance segmentation with significantly lower annotation costs. Our method demonstrates a significant advantage in differentiating adjacent instances. In contrast to the semantic segmentation algorithms used for comparison [40,44,45], which rely solely on pseudo-labels, our approach utilizes Gaussian decoders to support the segmentation decoder through a point-guided attention mechanism and predicted center-point maps. This integration significantly improves the precision of differentiating adjacent instances. The effectiveness of this approach is corroborated by the results presented in Figure 7 and Table 3. Ablation experiments on the Gaussian decoder branches show that using multiple branches with varying radii for center-point prediction provides greater accuracy and stability compared to single-branch predictions. Moreover, the introduction of Gaussian decoders does not significantly increase the inference time, which remains comparable to that of common instance segmentation networks. This indicates that our method is practical for real-world clinical applications. Furthermore, as shown in Figure 11 and Table 3, our method effectively mitigates the impact of pseudo-label noise. The point-guided attention mechanism reduces dependency on pseudo-label boundary quality, while the pseudo-label updating strategy leverages historical training information to improve boundary accuracy. These enhancements not only enhance model performance but also accelerate convergence.
Our method also exhibits strong scalability and adaptability to other imaging modalities and datasets: (1) By requiring only simple point annotations for training, it significantly reduces annotation costs compared to fully supervised methods, making it inherently scalable and adaptable to other datasets or application scenarios. (2) The model does not rely on the characteristics of histopathological images. It is applicable to networks with dense instances and can be easily transferred to datasets from other imaging modalities, such as CT and ultrasound. (3) Due to the parallel computing advantages of the Transformer, the model can perform fast inference on large-scale datasets. This capability also enables the design of deeper and more complex network architectures, making it suitable for tackling more complex imaging tasks.
Our method can be integrated into clinical workflows in several ways. First, in clinical applications, our method reduces the annotation workload for doctors and specialists, allowing them to dedicate more time to diagnosis and treatment, which improves efficiency. Second, the method’s ability to work across different imaging modalities shows its potential for broader application in various medical environments. Third, the response time of our method is comparable to standard instance segmentation networks, offering timely feedback for medical diagnoses. Compared to other point-supervised segmentation algorithms, it provides more accurate nuclei localization to support diagnosis while requiring less training effort than fully supervised methods.
Although our method has shown promising results across multiple datasets, we acknowledge that there are still shortcomings in our approach. (1) First, our method demonstrates significant advantages in segmenting circular cells, as observed in test images of the stomach and prostate sections of the MonuSeg dataset. However, for elongated and densely packed cells, such as those in the liver and colon sections, the model performance is poorer. As shown in Figure 12, elongated cells are often predicted to be shorter than the ground truth, and overlapping regions in crowded areas are classified as background to ensure instance separation. The point-guided attention mechanism, together with the use of distance maps as feature vectors during pseudo-label clustering, enhances the model’s ability to differentiate instances. However, this approach may sacrifice some boundary accuracy. Therefore, improving the boundary quality of elongated cells will be a major focus of our future work. (2) Our aim is to maximize performance with weak labels. While computational cost reduction was not our primary goal, it remains a significant concern. We will further reduce model complexity and inference time to meet the requirements of practical clinical applications. (3) Although point annotations require significantly less effort compared to full annotations, marking the center point for every cell in larger datasets can still be a considerable burden. In future work, we will attempt to further reduce the amount of annotation needed, such as using partial points for incomplete annotation [45,54]. Additionally, we will design algorithms to minimize the model’s reliance on point annotations, alleviating the burden on pathologists and reducing application costs.

5. Conclusions

In this paper, we propose a nuclei segmentation algorithm based solely on point labels for annotation. Our approach uses a multi-scale Gaussian kernel mechanism and a point-guided attention module. These components effectively leverage spatial information from point labels, improving segmentation accuracy and reducing pseudo-label noise. Additionally, we introduce a pseudo-label updating strategy that integrates historical training data and image color information to enhance overall training performance. Our method achieves state-of-the-art performance among similar approaches on three public datasets. However, our method still has limitations, such as suboptimal performance in segmenting elongated cells, the need for further reduction in model complexity, and the potential to further reduce the amount of center-point annotation. We hope that our work, while maintaining good segmentation performance, can reduce the amount of data annotation to assist clinical practices.

Author Contributions

Conceptualization, Y.M. and L.C.; methodology, Y.M.; software, Y.M.; validation, Y.M., L.C., L.Z. and Q.Z.; formal analysis, Y.M.; investigation, Y.M.; resources, L.C. and Q.Z.; data curation, Y.M.; writing—original draft preparation, Y.M. and L.Z.; writing—review and editing, Y.M., L.C, L.Z and Q.Z.; visualization, Y.M and L.Z.; supervision, L.C. and Q.Z.; project administration, L.C. and Q.Z.; funding acquisition, L.C. and Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under 62072021 and the Capital’s Funds for Health Improvement and Research, No.2024-1-2084.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the nature of open-public dataset usage. The public dataset providers requested only the citation.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated or analyzed during the study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cai, L.; Huang, S.; Zhang, Y.; Lu, J.; Zhang, Y. Rethinking Attention-Based Multiple Instance Learning for Whole-Slide Pathological Image Classification: An Instance Attribute Viewpoint. arXiv 2024, arXiv:2404.00351. [Google Scholar]
  2. Campanella, G.; Hanna, M.G.; Geneslaw, L.; Miraflor, A.; Werneck Krauss Silva, V.; Busam, K.J.; Brogi, E.; Reuter, V.E.; Klimstra, D.S.; Fuchs, T.J. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nat. Med. 2019, 25, 1301–1309. [Google Scholar] [CrossRef] [PubMed]
  3. Elmore, J.G.; Longton, G.M.; Carney, P.A.; Geller, B.M.; Onega, T.L.; Tosteson, A.N.; Nelson, H.D.; Pepe, M.S.; Allison, K.H.; Schnitt, S.J.; et al. Diagnostic concordance among pathologists interpreting breast biopsy specimens. JAMA 2015, 313, 1122–1132. [Google Scholar] [CrossRef] [PubMed]
  4. Prokhnevska, N.; Cardenas, M.A.; Valanparambil, R.M.; Sobierajska, E.; Barwick, B.G.; Jansen, C.; Kissick, H. CD8+ T cell activation in cancer comprises an initial activation phase in lymph nodes followed by effector differentiation within the tumor. Immunity 2023, 56, 107–124.e5. [Google Scholar] [CrossRef] [PubMed]
  5. Zeng, Q.; Mousa, M.; Nadukkandy, A.S.; Franssens, L.; Alnaqbi, H.; Alshamsi, F.Y.; Safar, H.A.; Carmeliet, P. Understanding tumour endothelial cell heterogeneity and function from single-cell omics. Nat. Rev. Cancer 2023, 23, 544–564. [Google Scholar] [CrossRef]
  6. de Visser, K.E.; Joyce, J.A. The evolving tumor microenvironment: From cancer initiation to metastatic outgrowth. Cancer Cell 2023, 41, 374–403. [Google Scholar] [CrossRef]
  7. Melssen, M.M.; Sheybani, N.D.; Leick, K.M.; Slingluff, C.L. Barriers to immune cell infiltration in tumors. J. ImmunoTherapy Cancer 2023, 11, e006401. [Google Scholar] [CrossRef]
  8. Cao, Y.; Langer, R.; Ferrara, N. Targeting angiogenesis in oncology, ophthalmology and beyond. Nat. Rev. Drug Discov. 2023, 22, 476–495. [Google Scholar] [CrossRef]
  9. Alsubaie, N.; Sirinukunwattana, K.; Raza, S.E.A.; Snead, D.R.J.; Rajpoot, N.M. A bottom-up approach for tumour differentiation in whole slide images of lung adenocarcinoma. In Medical Imaging; SPIE: Bellingham, WA, USA, 2018. [Google Scholar]
  10. Lal, S.; Das, D.; Alabhya, K.; Kanfade, A.; Kumar, A.; Kini, J. NucleiSegNet: Robust deep learning architecture for the nuclei segmentation of liver cancer histopathology images. Comput. Biol. Med. 2021, 128, 104075. [Google Scholar] [CrossRef] [PubMed]
  11. Lu, C.; Romo-Bucheli, D.; Wang, X.; Janowczyk, A.; Ganesan, S.; Gilmore, H.; Rimm, D.; Madabhushi, A. Nuclear shape and orientation features from H&E images predict survival in early-stage estrogen receptor-positive breast cancers. Lab. Investig. 2018, 98, 1438–1448. [Google Scholar] [CrossRef]
  12. Qi, X.; Xing, F.; Foran, D.; Yang, L. Robust segmentation of overlapping cells in histopathology specimens using parallel seed detection and repulsive level set. IEEE Trans. Biomed. Eng. 2012, 59, 754–765. [Google Scholar] [CrossRef] [PubMed]
  13. Mouelhi, A.; Sayadi, M.; Fnaiech, F.; Mrad, K.; Romdhane, K.B. Automatic image segmentation of nuclear stained breast tissue sections using color active contour model and an improved watershed method. Biomed. Signal Process. Control 2013, 8, 421–436. [Google Scholar] [CrossRef]
  14. Shu, J.; Fu, H.; Qiu, G.; Kaye, P.; Ilyas, M. Segmenting overlapping cell nuclei in digital histopathology images. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; Volume 2013, pp. 5445–5448. [Google Scholar] [CrossRef]
  15. Plissiti, M.; Nikou, C.; Charchanti, A. Automated Detection of Cell Nuclei in Pap Smear Images Using Morphological Reconstruction and Clustering. IEEE Trans. Inf. Technol. Biomed. 2010, 15, 233–241. [Google Scholar] [CrossRef] [PubMed]
  16. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  17. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
  18. Raza, S.E.A.; Cheung, L.; Shaban, M.; Graham, S.; Epstein, D.; Pelengaris, S.; Khan, M.; Rajpoot, N.M. Micro-Net: A unified model for segmentation of various objects in microscopy images. Med Image Anal. 2019, 52, 160–173. [Google Scholar] [CrossRef] [PubMed]
  19. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  20. Vaswani, A. Attention is all you need. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2017. [Google Scholar]
  21. Lin, A.; Chen, B.; Xu, J.; Zhang, Z.; Lu, G.; Zhang, D. Ds-transunet: Dual swin transformer u-net for medical image segmentation. IEEE Trans. Instrum. Meas. 2022, 71, 1–15. [Google Scholar] [CrossRef]
  22. Prangemeier, T.; Reich, C.; Koeppl, H. Attention-based transformers for instance segmentation of cells in microstructures. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, South Korea, 16–19 December 2020; pp. 700–707. [Google Scholar]
  23. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Istanbul, Turkey, 5–8 December 2023; pp. 4015–4026. [Google Scholar]
  24. Li, X.; Yuan, H.; Li, W.; Ding, H.; Wu, S.; Zhang, W.; Li, Y.; Chen, K.; Loy, C.C. OMG-Seg: Is one model good enough for all segmentation? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Lisboa, Portugal, 3–6 December 2024; pp. 27948–27959. [Google Scholar]
  25. Graham, S.; Vu, Q.D.; Raza, S.E.A.; Azam, A.; Tsang, Y.W.; Kwak, J.T.; Rajpoot, N. Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images. Med. Image Anal. 2019, 58, 101563. [Google Scholar] [CrossRef] [PubMed]
  26. He, H.; Huang, Z.; Ding, Y.; Song, G.; Wang, L.; Ren, Q.; Chen, J. CDNet: Centripetal Direction Network for Nuclear Instance Segmentation. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 4006–4015. [Google Scholar] [CrossRef]
  27. He, Z.; Unberath, M.; Ke, J.; Shen, Y. TransNuSeg: A Lightweight Multi-task Transformer for Nuclei Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; pp. 206–215. [Google Scholar] [CrossRef]
  28. Tian, Z.; Shen, C.; Wang, X.; Chen, H. BoxInst: High-Performance Instance Segmentation with Box Annotations. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 5439–5448. [Google Scholar] [CrossRef]
  29. Cheng, T.; Wang, X.; Chen, S.; Zhang, Q.; Liu, W. BoxTeacher: Exploring High-Quality Pseudo Labels for Weakly Supervised Instance Segmentation. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 3145–3154. [Google Scholar]
  30. Wang, J.; Xia, B. Weakly supervised image segmentation beyond tight bounding box annotations. Comput. Biol. Med. 2024, 169, 107913. [Google Scholar] [CrossRef] [PubMed]
  31. Bilodeau, A.; Delmas, C.V.L.; Parent, M.; Koninck, P.D.; Durand, A.; Lavoie-Cardinal, F. Microscopy analysis neural network to solve detection, enumeration and segmentation from image-level annotations. Nat. Mach. Intell. 2022, 4, 455–466. [Google Scholar] [CrossRef]
  32. Zhou, Y.; Wu, Y.; Wang, Z.; Wei, B.; Lai, M.; Shou, J.; Fan, Y.; Xu, Y. Cyclic Learning: Bridging Image-Level Labels and Nuclei Instance Segmentation. IEEE Trans. Med. Imaging 2023, 42, 3104–3116. [Google Scholar] [CrossRef]
  33. Kim, B.; Jeong, J.; Han, D.; Hwang, S.J. The Devil is in the Points: Weakly Semi-Supervised Instance Segmentation via Point-Guided Mask Representation. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 11360–11370. [Google Scholar] [CrossRef]
  34. Nishimura, K.; Bise, R. Weakly Supervised Cell-Instance Segmentation with Two Types of Weak Labels by Single Instance Pasting. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 3184–3193. [Google Scholar] [CrossRef]
  35. Tian, K.; Zhang, J.; Shen, H.; Yan, K.; Dong, P.; Yao, J.; Che, S.; Luo, P.; Han, X. Weakly-Supervised Nucleus Segmentation Based on Point Annotations: A Coarse-to-Fine Self-Stimulated Learning Strategy. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2020, Lima, Peru, 4–8 October 2020; pp. 299–308. [Google Scholar] [CrossRef]
  36. Wang, Z.; Zhang, Y.; Wang, Y.; Cai, L.; Zhang, Y. Dynamic Pseudo Label Optimization in Point-Supervised Nuclei Segmentation. arXiv 2024, arXiv:2406.16427. [Google Scholar]
  37. Xia, L.; Qu, Z.; An, J.; Gao, Z. A Weakly Supervised Method With Colorization for Nuclei Segmentation Using Point Annotations. IEEE Trans. Instrum. Meas. 2023, 72, 1–11. [Google Scholar] [CrossRef]
  38. Zhang, S.; Yu, Z.; Liu, L.; Wang, X.; Zhou, A.; Chen, K. Group R-CNN for Weakly Semi-supervised Object Detection with Points. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 9407–9416. [Google Scholar] [CrossRef]
  39. Hartigan, J.A.; Wong, M.A. A K-Means Clustering Algorithm. J. R. Stat. Soc. Ser. C Appl. Stat. 2018, 28, 100–108. [Google Scholar] [CrossRef]
  40. Qu, H.; Wu, P.; Huang, Q.; Yi, J.; Riedlinger, G.M.; De, S.; Metaxas, D.N. Weakly Supervised Deep Nuclei Segmentation using Points Annotation in Histopathology Images. In Proceedings of the International Conference on Medical Imaging with Deep Learning, London, UK, 8–10 July 2019. [Google Scholar]
  41. Qu, H.; Yi, J.; Huang, Q.; Wu, P.; Metaxas, D. Nuclei Segmentation Using Mixed Points and Masks Selected From Uncertainty. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 973–976. [Google Scholar] [CrossRef]
  42. Yoo, I.; Yoo, D.; Paeng, K. PseudoEdgeNet: Nuclei Segmentation only with Point Annotations. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; pp. 731–739. [Google Scholar]
  43. Liu, W.; He, Q.; He, X. Weakly Supervised Nuclei Segmentation Via Instance Learning. In Proceedings of the 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–5. [Google Scholar]
  44. Lin, Y.; Qu, Z.; Chen, H.; Gao, Z.; Li, Y.; Xia, L.; Ma, K.; Zheng, Y.; Cheng, K.T. Nuclei segmentation with point annotations from pathology images via self-supervised learning and co-training. Med. Image Anal. 2023, 89, 102933. [Google Scholar] [CrossRef]
  45. Qu, H.; Wu, P.; Huang, Q.; Yi, J.; Yan, Z.; Li, K.; Metaxas, D.N. Weakly Supervised Deep Nuclei Segmentation Using Partial Points Annotation in Histopathology Images. IEEE Trans. Med. Imaging 2020, 39, 3655–3666. [Google Scholar] [CrossRef]
  46. Nam, S.; Jeong, J.; Luna, M.; Chikontwe, P.; Park, S.H. PROnet: Point Refinement Using Shape-Guided Offset Map for Nuclei Instance Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; pp. 528–538. [Google Scholar] [CrossRef]
  47. Laine, S.; Aila, T. Temporal Ensembling for Semi-Supervised Learning. arXiv 2017, arXiv:1610.02242. [Google Scholar]
  48. Kumar, N.; Verma, R.; An, D.; Zhou, Y.; Onder, O.F.; Tsougenis, E.; Sethi, A. A Multi-Organ Nucleus Segmentation Challenge. IEEE Trans. Med. Imaging 2020, 39, 1380–1391. [Google Scholar] [CrossRef]
  49. Vu, Q.D.; Graham, S.; Kurc, T.; To, M.N.N.; Shaban, M.; Qaiser, T.; Farahani, K. Methods for Segmentation and Classification of Digital Microscopy Tissue Images. Front. Bioeng. Biotechnol. 2019, 7, 433738. [Google Scholar] [CrossRef]
  50. Sirinukunwattana, K.; Snead, D.R.J.; Rajpoot, N.M. A Stochastic Polygons Model for Glandular Structures in Colon Histology Images. IEEE Trans. Med. Imaging 2015, 34, 2366–2378. [Google Scholar] [CrossRef]
  51. Kumar, N.; Verma, R.; Sharma, S.; Bhargava, S.; Vahadane, A.; Sethi, A. A Dataset and a Technique for Generalized Nuclear Segmentation for Computational Pathology. IEEE Trans. Med. Imaging 2017, 36, 1550–1560. [Google Scholar] [CrossRef] [PubMed]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  53. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. arXiv 2018, arXiv:1703.06870. [Google Scholar]
  54. Lin, Y.; Wang, Z.; Zhang, D.; Cheng, K.T.; Chen, H. BoNuS: Boundary Mining for Nuclei Segmentation With Partial Point Labels. IEEE Trans. Med. Imaging 2024, 43, 2137–2147. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Different labels for the nuclei segmentation task. (a) Input image. (b) Pixel-level instance label. (c) Point annotation. (d) Binary mask. (e) Voronoi label. (f) Cluster label. (g) Distance map. In (b), different instances are marked with specific colors. In (c,d), white and black pixels, respectively, represent foreground and background regions, while in (e,f), green, red, and black pixels indicate foreground, background, and ignored areas, respectively. In (g), each pixel value represents the distance from that pixel to the nearest centroid, depicted as a grayscale image.
Figure 1. Different labels for the nuclei segmentation task. (a) Input image. (b) Pixel-level instance label. (c) Point annotation. (d) Binary mask. (e) Voronoi label. (f) Cluster label. (g) Distance map. In (b), different instances are marked with specific colors. In (c,d), white and black pixels, respectively, represent foreground and background regions, while in (e,f), green, red, and black pixels indicate foreground, background, and ignored areas, respectively. In (g), each pixel value represents the distance from that pixel to the nearest centroid, depicted as a grayscale image.
Bioengineering 12 00085 g001
Figure 2. A bad case of segmenting adjacent nuclei. (a) Input image. (b) Ground truth. (c) Predicted instance map. As indicated within the yellow box, multiple closely adjacent nuclei are predicted as a single nucleus.
Figure 2. A bad case of segmenting adjacent nuclei. (a) Input image. (b) Ground truth. (c) Predicted instance map. As indicated within the yellow box, multiple closely adjacent nuclei are predicted as a single nucleus.
Bioengineering 12 00085 g002
Figure 3. Overview of our proposed method. The Voronoi label ν , cluster label K , and Gaussian masks in the figure are all generated from point annotations. The model outputs two maps: (1) the center-point prediction from the Gaussian branch and (2) the segmentation prediction from the segmentation branch. During post-processing, the center-point map refines the segmentation map for more precise instance segmentation.
Figure 3. Overview of our proposed method. The Voronoi label ν , cluster label K , and Gaussian masks in the figure are all generated from point annotations. The model outputs two maps: (1) the center-point prediction from the Gaussian branch and (2) the segmentation prediction from the segmentation branch. During post-processing, the center-point map refines the segmentation map for more precise instance segmentation.
Bioengineering 12 00085 g003
Figure 4. The issues arising in inference. The images in (a,b), respectively, depict the comparison result of segmentation before and after adding a centroid to an instance that does not correspond to a center point. The images in (c,d), respectively, illustrate the comparison result of nuclei segmentation before and after merging instances.
Figure 4. The issues arising in inference. The images in (a,b), respectively, depict the comparison result of segmentation before and after adding a centroid to an instance that does not correspond to a center point. The images in (c,d), respectively, illustrate the comparison result of nuclei segmentation before and after merging instances.
Bioengineering 12 00085 g004
Figure 5. Pseudo-label update strategy. Black arrows denote label updates during a cycle of training, while green arrows represent updates after completing a training cycle.
Figure 5. Pseudo-label update strategy. Black arrows denote label updates during a cycle of training, while green arrows represent updates after completing a training cycle.
Bioengineering 12 00085 g005
Figure 6. The structure of the point-guided attention module. (a) The interaction of the point-guided blocks within the model. (b) The processing of feature maps by the point-guided block in the segmentation branch and the Gaussian branches. (c) The main part of point-guided attention. Gray, blue, and yellow blocks indicate feature maps from the encoder, segmentation decoder, and Gaussian decoder, respectively. The abbreviations “Norm”, “Attn”, “PG”, and “FFN” refer to layer normalization, attention, point-guided block, and feed-forward network.
Figure 6. The structure of the point-guided attention module. (a) The interaction of the point-guided blocks within the model. (b) The processing of feature maps by the point-guided block in the segmentation branch and the Gaussian branches. (c) The main part of point-guided attention. Gray, blue, and yellow blocks indicate feature maps from the encoder, segmentation decoder, and Gaussian decoder, respectively. The abbreviations “Norm”, “Attn”, “PG”, and “FFN” refer to layer normalization, attention, point-guided block, and feed-forward network.
Bioengineering 12 00085 g006
Figure 7. Segmentation results of different methods. The regions marked with yellow borders highlight the advantages of our approach. The test images are cropped from the Monuseg (first and second rows), CPM17 (third and fourth rows), and CoNSeP (fifth and sixth rows) datasets. Different nuclei are represented in different colors.
Figure 7. Segmentation results of different methods. The regions marked with yellow borders highlight the advantages of our approach. The test images are cropped from the Monuseg (first and second rows), CPM17 (third and fourth rows), and CoNSeP (fifth and sixth rows) datasets. Different nuclei are represented in different colors.
Bioengineering 12 00085 g007
Figure 8. The results of ablation experiments. The letters below the figures correspond to the models listed in Table 3.
Figure 8. The results of ablation experiments. The letters below the figures correspond to the models listed in Table 3.
Bioengineering 12 00085 g008
Figure 9. The results of using different λ 1 and λ 2 values across three datasets.
Figure 9. The results of using different λ 1 and λ 2 values across three datasets.
Bioengineering 12 00085 g009
Figure 10. Quantitative results with different numbers of Gaussian branches on the Monuseg dataset. (a) Model complexity. (b) Training time. (c) Inference time.
Figure 10. Quantitative results with different numbers of Gaussian branches on the Monuseg dataset. (a) Model complexity. (b) Training time. (c) Inference time.
Bioengineering 12 00085 g010
Figure 11. Visualization results of the pseudo-label updating strategy. Rows 2, 3, and 4 depict the ground truth, initial clustering labels, and updated clustering labels obtained after one round of training, where green, red, and black represent nuclei regions, non-nuclei regions, and uncertain regions, respectively.
Figure 11. Visualization results of the pseudo-label updating strategy. Rows 2, 3, and 4 depict the ground truth, initial clustering labels, and updated clustering labels obtained after one round of training, where green, red, and black represent nuclei regions, non-nuclei regions, and uncertain regions, respectively.
Bioengineering 12 00085 g011
Figure 12. Some failure cases occur in our method when dealing with elongated and densely packed cells; a successful case is also shown. (a) Missed detection. (b) Failed segmentation of densely packed elongated cells. (c) A successful case.
Figure 12. Some failure cases occur in our method when dealing with elongated and densely packed cells; a successful case is also shown. (a) Missed detection. (b) Failed segmentation of densely packed elongated cells. (c) A successful case.
Bioengineering 12 00085 g012
Table 1. Comparative experiments on the Monuseg, CPM17, and CoNSeP datasets.
Table 1. Comparative experiments on the Monuseg, CPM17, and CoNSeP datasets.
MethodsMonusegCPM17CoNSeP
Dice obj AJIDQSQPQ Dice obj AJIDQSQPQ Dice obj AJIDQSQPQ
Fully Supervised
Mask-RCNN [53]0.8070.6230.8060.7610.6140.8370.6860.8470.8010.6800.7320.5050.6260.7110.446
Micro-Net [18]0.7930.6030.7600.7570.5760.8450.6740.8380.7820.6570.7560.5310.6090.7470.455
HoverNet [25]0.8170.6180.7700.7730.5970.8480.7050.8540.8140.6970.8070.5710.7020.7780.547
Point-Supervised
Weak-Anno [40]0.7290.5460.7140.7170.5130.7710.6070.7500.7160.5420.6050.3530.4580.6790.312
PseudoEdgeNet [42]0.7110.5060.6370.7120.4540.7280.5530.7330.6390.4690.5570.2700.3910.6840.269
C2FNet [35]0.7170.5390.7010.7150.5010.7350.5670.6570.6830.4480.5690.2590.4290.6830.293
WSPP [45]0.7330.5460.7240.6970.5060.7460.5610.7420.6890.5170.6090.3580.4600.6830.315
SPN+IEN [43]0.7480.5780.7460.7180.5370.7750.6120.7750.7260.5650.6350.4050.4840.6690.326
SC-Net [44]0.7380.5620.7300.7120.5210.7590.5980.7460.7060.5300.6090.3720.4360.6950.305
Ours0.7720.6120.7970.7140.5690.7870.6360.8250.7210.5970.6510.4330.4930.6670.331
Table 2. The segmentation results on the MonuSeg test set across different organs.
Table 2. The segmentation results on the MonuSeg test set across different organs.
Organ Dice obj AJIDQSQPQ
Bladder0.7870.6350.8190.7210.590
Brain0.7810.6200.7860.7070.556
Breast0.7540.5910.7620.6960.530
Colon0.7310.5440.7190.6980.502
Kidney0.7770.6230.8200.7230.593
Lung0.7760.6240.8340.7180.599
Prostate0.7740.6100.7850.7260.570
Table 3. Ablation study on the proposed methods, where “MG”, “PA”, and “LU” represent the multi-scale Gaussian kernel module, point-guided attention module, and pseudo-label updating module. Results are presented as mean ± standard deviation from 5 runs.
Table 3. Ablation study on the proposed methods, where “MG”, “PA”, and “LU” represent the multi-scale Gaussian kernel module, point-guided attention module, and pseudo-label updating module. Results are presented as mean ± standard deviation from 5 runs.
MGPALUMonusegCPM17CoNSeP
Dice obj AJIDQSQPQ Dice obj AJIDQSQPQ Dice obj AJIDQSQPQ
A0.726 ± 0.0120.534 ± 0.0220.695 ± 0.0210.714 ± 0.0030.496 ± 0.0180.764± 0.0070.597 ± 0.0100.742 ± 0.0080.712 ± 0.0040.531 ± 0.0110.582 ± 0.0230.337 ± 0.0160.440 ± 0.0180.672 ± 0.0070.296 ± 0.016
B🗸0.742 ± 0.0050.571 ± 0.0030.729 ± 0.0030.717 ± 0.0020.523 ± 0.0040.771 ± 0.0030.612 ± 0.0050.780 ± 0.0030.716 ± 0.0020.558 ± 0.0030.629 ± 0.0100.415 ± 0.0050.467 ± 0.0060.673 ± 0.0030.316 ± 0.006
C🗸0.739 ± 0.0070.559 ± 0.0070.733 ± 0.0060.719 ± 0.0040.528 ± 0.0070.772 ± 0.0040.613 ± 0.0020.799 ± 0.0040.715 ± 0.0030.573 ± 0.0060.633 ± 0.0120.353 ± 0.0140.446 ± 0.130.675 ± 0.0060.302 ± 0.011
D🗸🗸0.761 ± 0.0030.593 ± 0.0040.769 ± 0.0050.717 ± 0.0020.552 ± 0.0060.776 ± 0.0020.625 ± 0.0020.804 ± 0.0030.718 ± 0.0020.579 ± 0.0050.632 ± 0.0060.425 ± 0.0070.481 ± 0.0050.666 ± 0.0070.319 ± 0.002
E🗸🗸0.753 ± 0.0040.581 ± 0.0050.738 ± 0.0060.719 ± 0.0030.531 ± 0.0060.780 ± 0.0030.625 ± 0.0010.807 ± 0.0030.715 ± 0.0020.579 ± 0.0050.641 ± 0.0060.423 ± 0.0070.454 ± 0.0060.677 ± 0.0040.308 ± 0.006
F🗸🗸🗸0.770 ± 0.0020.609 ± 0.0030.793 ± 0.0040.715 ± 0.0030.565 ± 0.0040.785 ± 0.0020.634 ± 0.0020.821 ± 0.0040.718 ± 0.0030.592 ± 0.0050.647 ± 0.0040.428 ± 0.0050.488 ± 0.0050.667 ± 0.0040.327 ± 0.004
Table 4. Ablation study on the number of Gaussian branches.
Table 4. Ablation study on the number of Gaussian branches.
NumMonusegCPM17CoNSeP
Dice obj AJIDQSQPQ Dice obj AJIDQSQPQ Dice obj AJIDQSQPQ
10.7420.5610.7230.7190.5200.7720.5960.7580.7130.5410.6190.3760.4630.6740.313
20.7460.5720.7280.7180.5230.7730.6140.7810.7160.5590.6390.4190.4720.6760.320
30.7460.5730.7300.7180.5230.7740.6160.7810.7170.5600.6400.4210.4750.6760.322
40.7470.5740.7320.7190.5270.7740.6170.7830.7180.5610.6390.4200.4730.6760.322
50.7460.5720.7310.7190.5250.7740.6160.7820.7170.5610.6410.4220.4760.6750.323
60.7460.5710.7310.7190.5250.7750.6170.7820.7180.5620.6400.4190.4720.6760.321
Table 5. Ablation study on the inference strategy. For each dataset, the first row represents the results using the coarse instance map obtained by directly multiplying the segmentation map S with the Voronoi partition P derived from the coarse center-point map C. The second row represents the results using the refined instance maps obtained from Algorithm 1.
Table 5. Ablation study on the inference strategy. For each dataset, the first row represents the results using the coarse instance map obtained by directly multiplying the segmentation map S with the Voronoi partition P derived from the coarse center-point map C. The second row represents the results using the refined instance maps obtained from Algorithm 1.
DatasetMethod Dice o b j AJIDQSQPQ
Monuseg0.7650.5920.7900.7160.566
🗸0.7720.6120.7970.7140.569
CPM170.7860.6360.8200.7200.593
🗸0.7870.6390.8250.7210.597
CoNSeP0.6420.4020.4880.6690.329
🗸0.6510.4330.4930.6670.331
Table 6. The training and inference times on the Monuseg dataset. The parameter count in the table refers to the total number of parameters during the model training phase. The last two rows represent the times of our method in generating a coarse instance map and a refined instance map, respectively.
Table 6. The training and inference times on the Monuseg dataset. The parameter count in the table refers to the total number of parameters during the model training phase. The last two rows represent the times of our method in generating a coarse instance map and a refined instance map, respectively.
MethodsParamsTrainingTotalAverage
TimeInference TimeInference Time
[M][s/epoch][s][s/img]
HoverNet [25]45.6816.2527.691.978
Micro-Net [18]25.8313.1711.650.832
Mask-RCNN [53]43.9826.5838.822.773
Weak-Anno [40]24.9111.783.4020.243
WSPP [45]49.8212.065.3480.382
SC-Net [44]69.0623.589.0020.506
SPN+IEN [43]49.8643.2884.426.030
Ours(C)49.2517.827.1820.513
Ours(R)49.2517.8228.622.044
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mo, Y.; Chen, L.; Zhang, L.; Zhao, Q. Weakly Supervised Nuclei Segmentation with Point-Guided Attention and Self-Supervised Pseudo-Labeling. Bioengineering 2025, 12, 85. https://doi.org/10.3390/bioengineering12010085

AMA Style

Mo Y, Chen L, Zhang L, Zhao Q. Weakly Supervised Nuclei Segmentation with Point-Guided Attention and Self-Supervised Pseudo-Labeling. Bioengineering. 2025; 12(1):85. https://doi.org/10.3390/bioengineering12010085

Chicago/Turabian Style

Mo, Yapeng, Lijiang Chen, Lingfeng Zhang, and Qi Zhao. 2025. "Weakly Supervised Nuclei Segmentation with Point-Guided Attention and Self-Supervised Pseudo-Labeling" Bioengineering 12, no. 1: 85. https://doi.org/10.3390/bioengineering12010085

APA Style

Mo, Y., Chen, L., Zhang, L., & Zhao, Q. (2025). Weakly Supervised Nuclei Segmentation with Point-Guided Attention and Self-Supervised Pseudo-Labeling. Bioengineering, 12(1), 85. https://doi.org/10.3390/bioengineering12010085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop