Next Article in Journal
Deep Modular Bilinear Attention Network for Visual Question Answering
Previous Article in Journal
Low-Cost Ka-Band Satellite Receiver Data Preprocessing for Tropospheric Propagation Studies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stain Style Transfer for Histological Images Using S3CGAN

Department of Computer Science and Information Engineering, National University of Tainan, Tainan 700, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1044; https://doi.org/10.3390/s22031044
Submission received: 30 November 2021 / Revised: 18 January 2022 / Accepted: 26 January 2022 / Published: 28 January 2022
(This article belongs to the Section Biomedical Sensors)

Abstract

:
This study proposes a new CycleGAN-based stain transfer model, called S3CGAN, equipped with a specialized color classifier structure. The specialized color classifier can assist the generative network to conquer the existing challenge in GANs, namely the instability of the network caused by the insufficient representativeness of the training data in the initial stage of network training. The color classifier is pretrained, hence it can provide correct color information feedback to the generator during the initial network training phase. The augmented information from color classification enables the generator to generate superior results. Owing to the CycleGAN architecture, the proposed model does not require representative paired inputs. The proposed model uses U-Net and a Markovian discriminator to enhance the structural retention ability to generate images with high fidelity.

1. Introduction

Medical imaging techniques, such as ultrasound imaging, computed tomographic scanning, positron emission tomography, single-photon emission computed tomography, magnetic resonance imaging, and thermal imaging, are used for the early detection of cancer. However, histopathological biopsies are currently the only recognized method for cancer diagnosis. Biopsy technologies can now scan the entire tissues on microscope slides and produce whole slide images (WSIs). These images can then be used for automated image analysis. The production of WSIs often results in varying stain colors due to reasons such as using different batches of staining reagents, different suppliers of staining reagents, samples with different thicknesses, different staining conditions, and different image scanners. Variations in color diminish the performance of automated image analysis.
Conventional image processing techniques [1,2,3,4] have been used to resolve the problem of stain variations; however, they produce unfavorable results due to the limitations of the image processing methods. These limitations come from improper color mapping, failure to take spatial features into account, or the need to choose a representative reference image. Recently, generative adversarial networks (GANs) [5] exhibit amazing results in generating high-fidelity images. GANs consist of two networks, a generator and a discriminator. They both play an adversarial game during the training stage, where the generator tries to cheat the discriminator by generating data similar to those in the training set. The discriminator tries not to be cheated by distinguishing between real and fake data. In the testing stage, the trained generator is used to generate data that is highly realistic to real data. Conditional GANs (CGANs) [6] are a natural extension of GANs in which the limitation factors can be set for the input of the network, forming paired inputs, to further control the generated data. This controllable mechanism is suitable for generating specific images. However, one inherent weakness of CGAN is that the input data must be paired. In [7], a CGAN-based approach was proposed to solve the stain variation problem, but their stain transfer quality was not ideal. This is because they transfer stain style by colorizing gray images; however, even the same tissue can produce different grayscale values under different staining conditions. In recent years, CycleGAN [8] was developed to transfer the characteristics of one image to another, called style transfer, via treating this transfer task as an image reconstruction problem. This approach naturally solves the problem that CGAN is limited by the need for paired inputs. In [9], a CycleGAN-based approach was proposed to normalize the stain variation. However, the fidelity of the transferred images was insufficient and sometimes produced abnormal tones. These problems arose from two insufficiencies of the generator, one was the structural retention ability, the other was the color–texture coherence.
An excellent pathological image staining conversion does not change the texture structure of the original image, but the new staining must be able to color according to the texture of the original image. The former is reflected in the structural retention ability, while the latter is reflected in the color–texture coherence for a generator. Based on this consideration, we propose a new neural network model called S3CGAN (specialized-color-classifier CycleGAN). The CycleGAN architecture is adopted to eliminate the need of selecting representative paired inputs for CGAN-based approaches. The structural retention ability is enhanced through the U-Net [10] and the Markovian discriminator [11]. The U-Net is used to endow the generator with pixel-level description capability and the Markovian discriminator provides distributed locality constraints on pathological image structures. A specialized color classifier is utilized to offer coherent color–texture information to the generator. The proposed model produces superior stain transfer outcomes and effectively resolves the color variation problem for pathological images.

2. Related Work

2.1. Conventional Color Normalization

To overcome the problem of stain variation, Reinhard et al. [1] proposed a color correction method in which the colors of one image were applied to another. In this method, the stain ratio between the source image and the target image is assumed to be similar. The source image is converted from the RGB color space to lαβ color space. The color histograms of the source and target images are then aligned to complete the stain transfer. Spatial information is not considered in this method; therefore, incorrect stain transfers often occur if the source and target images have different stain ratios. Macenko et al. [2] improved the method of Reinhard et al. [1] and conducted stain separation for stain normalization. First, the source image is converted from the RGB color space to the optical density space. Next, color deconvolution is performed to calculate the stain matrix, which indicates how each stain contributes to the RGB colors. Finally, the stain matrix is used for stain normalization. Vahadane et al. [3] used a structure-preserving color normalization algorithm, which is a spectral matching method, for color correction. This method is only suitable for hematoxylin and eosin (H&E) stains because the colors are rendered differently in other stains. Khan et al. [4] used color classifiers to calculate the contributions of different stains to the RGB colors and then used stain matrices for stain normalization. Although this method achieved a superior performance to that of Reinhard et al. [1], it has certain limitations because it requires a reference image for analysis.

2.2. Color Normalization Methods Based on GANs

Cho et al. proposed a stain transfer network based on a CGAN (Figure 1) [7]. The network τ is composed of two transformations: gray-normalization G and style-generator ζ. G standardizes each stain style of color images from different institutes and ζ colorizes gray images following the stain style of certain institutes. This method does not require additional corresponding labels for support; however, the stain transfer outcomes of this method are not ideal because even the same tissue can produce different grayscale values under different staining conditions.
Shaban et al. [9] proposed a new stain transfer network called StainGAN (Figure 2), which uses the original CycleGAN architecture for modeling. This network does not require paired data and additional task labels. However, the image fidelity after the stain style conversion is not ideal, and sometimes local color drifts appear. Preserving high fidelity after stain style conversion for pathological images is crucial for disease diagnosis. Distortion of the pathological image during stain transfer increases the risk of misdiagnosis. In addition, the network does not optimize the stability of its training process; thus, its training process often fails.

3. Proposed Method

3.1. System Architecture

To overcome the aforementioned problems, we propose a new stain transfer model called the S3CGAN. The input of this model comprises RGB channels so that it can avoid the problems associated with the method presented in [7], in which only brightness information is used. The proposed model uses U-Net, which has excellent structural expression ability, as the generative network architecture. The Markovian discriminator [11], which uses local receptive fields, is used to preserve the tissue structure in the images. In addition, we propose a specialized color classifier to assist the generative network to generate images with color characteristics similar to the target domain and to stabilize the network in the initial training phase. WGAN–GP (Wasserstein GAN with gradient penalty) [12] is used to further enhance the stability of network training. The architecture of the proposed S3CGAN is shown in Figure 3, where the scenario is based on the conversion of type A stain to type B stain and vice versa. The operation process of this architecture is briefly described as follows, and the more detailed operation details will be explained in the following sections. The source image with type A stain, denoted as SrcA, is converted to a new style with type B stain, denoted as DstAB, via the generator GA2B. The converted result DstAB is forward to three different parts, namely GB2A, DB, and Ccolor. GB2A is another generator used to converts DstAB back to the original type A stain, denoted as RconBA. DB is a discriminator used to check if DstAB looks like type B stain. Ccolor is a pretrained color classifier used to assess the likelihood of DstAB being type B stain from the color perspective. During the training stage, GA2B, GB2A, and DB are updated at the same time. For DB, its update driving force comes from the loss of distinguishing between real image (SrcB) and fake image (DstAB). For GA2B, its update driving force comes from three losses, one from the reconstruction quality (SrcA vs. RconBA), one from the success rate of GA2B cheating DB, and the other from the color possible correctness of DstAB. GB2A and DA undergo the same process in the reverse direction. Detailed explanations of the different parts of the proposed network are provided in the following sections.

3.2. Generator

In this study, the S3CGAN comprises two independent generators, one of which is responsible for transferring the source domain image to the target domain and the other of which is responsible for transferring the target domain image to the source domain. In contrast to the method used in [9], in which ResNet was used as the main generative architecture, we used U-Net [10] because it has excellent image detail expression ability. This attribute is crucial because if tissue sample images are distorted during the stain transfer, the accuracy of disease diagnosis may decrease. The architecture of the U-Net is displayed in Figure 4.

3.3. Discriminator

The discriminators used by conventional GANs output a single value, which can only reflect the overall quality of the generated image but cannot reflect the quality of different parts. That is, the receptive field of conventional GANs is the entire image, which often results in low image quality in certain parts of the generated image. This property is not appropriate for histological imaging, which requires the presentation of details and high discrimination. The Markovian discriminator outputs a group of values, each of which represents the image quality of a part of the generated image (Figure 5). This discriminator maps the input onto an N × N-sized matrix. Each value in the matrix represents the similarity between a patch and the corresponding target image feature, which can be used to optimize each patch. This study used the Markovian discriminator to enhance the structural retention quality after stain transfer.

3.4. Specialized Color Classifier

GAN models mostly exhibit a poor performance if training samples that lack representativeness are used during their initial training phase. To produce high-quality stain images via using generators, the texture and color distributions must be known. In conventional GAN architectures, the distribution information is indirectly provided by the discriminator in its feedback. Training generators to learn both the texture and color distributions is difficult if the discriminator only provides two feedback categories: real and fake. A method to resolve this problem is to increase the amount of information in the feedback. The information can be enhanced via using a color classifier to provide color classification results for a generated image. Such a classifier can be constructed using two approaches. The first one involves embedding color classification functions into the discriminator. This method, also known as embedded classification, is used in the auxiliary classifier GAN (AC-GAN) [13]. The second one is used in this study, which involves developing a specialized color classifier. In the first approach, the discriminator must be optimized for discrimination and classification tasks during training, which is more challenging. In the second approach, a pretrained color classifier Ccolor is used to feedback color information to the generator. Because the color classifier has been trained, it can provide correct color information feedback to the generator during the initial network training phase, which is the most crucial phase during network optimization and the phase in which network instability is most likely to occur. The color classifier can assist the optimization process of the generator. This classifier plays a similar role as a “color consultant.” It can also reduce the burden of the discriminator, facilitate the optimization of discrimination tasks, and stabilize the network training process.
Ccolor is a simple binary classification network with a lightweight architecture, as shown in Figure 6. Ccolor uses images from the source and target domains as training data. After training, Ccolor can determine if an input image belongs to the source or target domain. The augmented information of color classification enables the generator to generate superior results during the initial training phase. Even if the training samples with less stain representativeness are used for the discriminator during the initial training phase, owing to the augmented color classification information the bias coming from the training samples can still be corrected.

3.5. Training Process

Step 1: A total of 3000 patches each were selected from type A and type B stain images. These patches were used as training samples to train the color classifier network Ccolor.
Step 2: This step corresponds with the training phase of the CycleGAN. We used the 6000 patches selected from type A and type B stain images as training data. The training data set is different from that of the first step. Figure 3 corresponds to the use of the SrcA type A stain images as the input. We first input the SrcA images by batches into the generator GA2B to produce a group of fake type B stain images (DstAB images). The DstAB images were then input into the generator GB2A to reconstruct the original input image RconA. Subsequently, the generated DstAB images and original real type B stain images (SrcB images) were used as training data to train the discriminator DB such that DB could learn to recognize whether the input images were real type B stain images. The WGAN-GP loss function LadvD used for training is defined in (1), where P A ^ is the distribution calculated from the linear combination of the real distribution P ( r e a l ( A ) ) and generative distribution P ( g e n ( A ˜ ) ) . The gradient penalty in this loss can boost the stability during network training.
Step 3: This step corresponds to the second part of the training phase of the CycleGAN. The generator was trained after training the discriminator in step 2. We used three loss functions, namely Ladv, Lcyc, and Lcolor, to optimize the generator during generator training. The generated DstAB images were input into the discriminator DB to obtain the adversarial loss Ladv, reflecting the difference between the generated and the real type B stain images. The definition of the loss function Ladv is shown in Equation (2). The restored image RconA and the original SrcA images were used to calculate the mean absolute error, which was used to quantify the cyclic loss Lcyc. In addition, the generated DstAB images were input into Ccolor to calculate the cross entropy, denoted as Lcolor, between the classification results and the correct outcomes to reflect their similarity. Equations (1)–(5) present expressions for the loss functions and target loss functions used in the network training.
The generator GB2A and discriminator DA were trained using the same way but in the reverse direction. The training stopped after multiple training iterations when the generated image became stable.
a d v D = E A ˜ ~ P g e n ( A ˜ ) [ D A ( A ˜ ) ] E A ~ P r e a l ( A ) [ D A ( A ) ] + 10 E A ^ ~ P A ^ [ ( | | A ^ D A ( A ^ ) | | 2 1 ) 2 ]
a d v = E A ~ P d a t a ( A ) [ l o g ( 1 D B ( G A 2 B ( A ) ) ) ]
c y c ( G A 2 B , G B 2 A ) = E A ~ P d a t a ( A ) [ | | G B 2 A   ( G A 2 B   ( A ) ) A | | 1 ]
c o l o r = E A ~ P d a t a ( A ) [ l o g ( C c o l o r ( G A 2 B ( A ) ) ) ]
o b j = A d v + α c y c + β c o l o r

4. Experiments

Two datasets were used in this study to evaluate the performance of the proposed model. The Camelyon 16 dataset [14] was used to evaluate the tumor classification performance, and the Mitos-Atypia-14 dataset [15] was used to evaluate the generated image quality and structure fidelity.

4.1. Datasets

The Camelyon16 challenge was held by the International Symposium on Biomedical Imaging, and the goal of this challenge was to detect metastases of breast cancer automatically in H&E-stained lymph node sections. The Camelyon16 dataset includes many WSIs of sentinel lymph nodes that were provided by Radboud University Medical Center (marked as lab 1) and University Medical Center Utrecht (marked as lab 2). This study used 3000 patch images, each from lab 1 and lab 2, to train the stain transfer model. The size of the patch images is 256 × 256 pixels. A basic tumor classifier was used as a test platform to evaluate how stain transfer affected the tumor classification performance. This tumor classifier was trained using the dataset collected by Shaban et al. [9], which contains 14,704 tumor patch images and 14,704 nontumor patch images. The aforementioned training data were labeled type B stain images (Figure 7). The test data comprised 10,816 patch images and were labeled type A stain images (Figure 8).
The Mitos-Atypia-14 dataset comprises 280 H&E-stained biopsy WSIs of the same samples captured using two scanners, namely the Aperio Scanscope XT and Hamamatsu Nanozoomer 2.0-HT, at a magnification of 20× (Figure 9). The training data comprised 3000 patch images each for both scanners. The size of these patch images is 256 × 256 pixels. The test data comprised 4741 images, each scanned by the Aperio Scanscope XT and Hamamatsu Nanozoomer 2.0-HT as the ground truth.

4.2. Evaluating the Tumor Classification Performance

This experiment was conducted with the Camelyon16 dataset, and the tumor classification performance obtained for with/without stain style transfer was employed to indirectly verify the effectiveness of the proposed stain transfer method. A simple classification network architecture, the same as Figure 6, was used in this experiment because of its weak generalization capability, such that an excellent classification performance can be obtained only when the input image features are similar to the trained image features. Thus, the classification performance can indicate the similarity between the style-transferred images and the target images. We used the type B stain images [9] to train the tumor classifier. Next, the type A stain images [9] were converted to type B stain images via different stain transfer methods. These transferred type B stain images were input into the trained tumor classifier to evaluate the classification accuracy and indirectly verify the effectiveness of the different stain transfer methods.
Figure 10 shows the stain-transferred images using different approaches. We can find that the transferred image using the proposed method has the closest overall tone to the reference image. The results presented in Table 1, annotated as “(simple classifier)”, indicate that the proposed method produced superior classification performance to the other methods. This result indirectly indicates that the produced images by the proposed method are the closest to the training dataset’s type B stain image features. In Table 1, S3CGAN* refers to the S3CGAN without a specialized color classifier. The AUC value of the S3CGAN was 2% higher than that of the S3CGAN*, which proves that the specialized color classifier enhanced not only the stability during network training but also the stain transfer performance. The AUC value of the S3CGAN was also 2% higher than that of the AC-GAN structure, which uses an embedded classifier. This result indicates that the specialized color classifier can improve the overall stain transfer performance. The performance of the S3CGAN was 4% higher than that of the StainGAN. In addition to the effects of the specialized color classifier, the usage of the U-Net and the Markovian discriminator also improved the structural retention ability. To be closer to the clinical use situation, a sophisticated tumor classifier proposed by Jiang et al. [16] was also used. The results are presented in Table 1, annotated as “(complicated classifier)”. As expected, due to the better generalization ability of this classifier, the effect of the differences in the efficacy of the stain conversion methods is relatively limited. Even so, the method proposed in this study still outperforms other methods.

4.3. Mitos-Atypia-14 Experiment

The Mitos-Atypia-14 dataset contains biopsy images obtained using two scanners; thus, this dataset can be used for obtaining the ground truth and objectively evaluating the image quality of stain-transferred images. We used the structural similarity index measure (SSIM) [17] to evaluate the similarity between a stain-transferred image and the ground truth. Moreover, we used the PSNR to evaluate the image quality of the stain-transferred images. We converted the format of the images scanned by the Aperio Scanscope XT into that of the images scanned by the Hamamatsu Nanozoomer 2.0-HT by using the stain transfer methods listed in Table 2. The images obtained using the aforementioned scanners and the ground truth were then used to calculate the SSIM and PSNR, and the corresponding results are presented in Table 2. The results presented in Table 2 indicate that the GAN-based methods achieved higher mean SSIM and mean PSNR values than did the conventional image processing methods. This result indicates that GAN models are crucial for biopsy stain transfer tasks. In addition, S3CGAN achieved higher mean SSIM and mean PSNR values than did the other methods. To understand whether the method proposed in this study achieves statistical significance over other methods on the SSIM and PSNR, p-values were also calculated. The results show that all reach statistical significance. Thus, the model developed in this study has excellent fidelity capability for histological image stain transfer, which is crucial for histological image interpretation. Information on histological image structures, such as the shapes, distributions, and the relative positions of the cell nucleus and cytoplasm, can affect the clinical interpretation outcome.

4.4. Specialized Color Classifier vs. Embedded Classifier

This section presents detailed descriptions of the specialized color classifier and embedded classifier. The specialized color classifier used by the S3CGAN is a simple and pretrained convolutional neural network architecture. Therefore, it can support GAN training in the initial training phase and assist in optimizing the generator performance. The embedded classifier is a branch network of the discriminator and is still learning during the initial training phase; thus, it cannot effectively assist in optimizing the generator performance during the training phase. The abnormal color spots in Figure 11 appear when the embedded classifier is used in the training process.

4.5. Training Stability and the Hyperparameters of the Specialized Color Classifier

To understand the effect of the specialized color classifier on the training stability of the stain style conversion, we removed the specialized color classifier from S3CGAN and conducted seven independent runs. In each run, the network was initialized first and then took a snapshot for a randomly selected outcome after 1000 epochs. The seven snapshots are shown in Figure 12a. Using the same process, seven snapshots of S3CGAN’s outcomes were also taken, and the results are shown in Figure 12b. From Figure 12, we can find that adding the specialized color classifier can indeed make the color of the image generated by the network more accurate in the iterative process.
In this study, we used two hyperparameters in Equation (5), namely α and β. The parameter α is the weight for the cycle consistency and the parameter β is the weight for the color classification loss. We followed the parameter setup of the CycleGAN and set α as 10. Table 3 presents the different values adopted for β. We found that when β was between 0.1 and 1, the generated stain color transfer results did not differ considerably; however, the overall color contrast of the generated image was stronger when β was set as 1. A stronger color contrast is more suitable for tumor classification. Hence, we set the β value as 1 when performing stain transfer.

5. Conclusions

This study proposes a new stain transfer model based on the CycleGAN called S3CGAN. This model uses the end-to-end characteristic of neural network architectures to overcome the need of selecting reference images manually. The cyclic design of the CycleGAN enables it to overcome the limitation of the CGAN, which requires paired inputs. To improve the fidelity of the generated image, the S3CGAN model utilizes a U-Net-based generator with excellent structural detail expression ability and a Markovian-based discriminator with local receptive fields. For maintaining training stability, the proposed model uses the WGAN–GP during training to prevent unbalanced training due to the performance difference between the generator and the discriminator. In addition, we propose the use of a specialized color classifier as a “color consultant”, which can increase the stability of network training and improve the quality of the generated image. The experimental results indicate that, compared with conventional image processing methods and GAN-based methods, the proposed S3CGAN generates images with superior color and textural structures. The S3CGAN has excellent fidelity capability for histological mage stain transfer, which is a crucial capability for histological image interpretation.

Author Contributions

Conceptualization, J.-S.L. and Y.-X.M.; methodology, J.-S.L. and Y.-X.M.; software, Y.-X.M.; validation, Y.-X.M.; formal analysis, Y.-X.M.; data curation, Y.-X.M.; writing—original draft preparation, J.-S.L. and Y.-X.M.; writing—review and editing, J.-S.L. and Y.-X.M.; supervision, J.-S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Science and Technology under grant number MOST 108-2221-E-024-011-MY3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color transfer between images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
  2. Macenko, M.; Niethammer, M.; Marron, J.S.; Borland, D.; Woosley, J.T.; Guan, X.; Schmitt, C.; Thomas, N.E. A method for normalizing histology slides for quantitative analysis. In Proceedings of the 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, ISBI’09, Boston, MA, USA, 28 June–1 July 2009; pp. 1107–1110. [Google Scholar]
  3. Vahadane, A.; Peng, T.; Sethi, A.; Albarqouni, S.; Wang, L.; Baust, M.; Navab, N.; Steiger, K.; Schlitter, A.M.; Esposito, I. Structure-preserving color normalization and sparse stain separation for histological images. IEEE Trans. Med. Imag. 2016, 35, 1962–1971. [Google Scholar] [CrossRef] [PubMed]
  4. Khan, A.M.; Rajpoot, N.; Treanor, D.; Magee, D. A nonlinear mapping approach to stain normalization in digital histopathology images using image-specific color deconvolution. IEEE Trans. Biomed. Eng. 2014, 61, 1729–1738. [Google Scholar] [CrossRef] [PubMed]
  5. Goodfellow, J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  6. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  7. Cho, H.; Lim, S.; Choi, G.; Min, H. Neural Stain-Style Transfer Learning Using Gan for Histopathological Images. arXiv 2017, arXiv:1710.08543. [Google Scholar]
  8. Zhu, J.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar] [CrossRef] [Green Version]
  9. Shaban, M.T.; Baur, C.; Navab, N.; Albarqouni, S. Staingan: Stain Style Transfer for Digital Histological Images. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 953–956. [Google Scholar] [CrossRef] [Green Version]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015. [Google Scholar]
  11. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  12. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of wasserstein gans. arXiv 2017, arXiv:1704.00028. [Google Scholar]
  13. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. arXiv 2016, arXiv:1610.09585. [Google Scholar]
  14. Camelyon16. Available online: https://camelyon16.grand-challenge.org/ (accessed on 8 June 2020).
  15. Mitos-Atypia-14. Available online: https://mitos-atypia-14.grand-challenge.org/ (accessed on 8 June 2020).
  16. Jiang, Y.; Chen, L.; Zhang, H.; Xiao, X. Breast cancer histopathological image classification using convolutional neural networks with small SE-ResNet module. PLoS ONE 2019, 14, e0214587. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Overview of the stain style transfer network proposed by Cho.
Figure 1. Overview of the stain style transfer network proposed by Cho.
Sensors 22 01044 g001
Figure 2. System architecture of StainGAN.
Figure 2. System architecture of StainGAN.
Sensors 22 01044 g002
Figure 3. System architecture of the S3CGAN.
Figure 3. System architecture of the S3CGAN.
Sensors 22 01044 g003
Figure 4. Architecture of U-Net.
Figure 4. Architecture of U-Net.
Sensors 22 01044 g004
Figure 5. Markovian discriminator.
Figure 5. Markovian discriminator.
Sensors 22 01044 g005
Figure 6. Architecture of the specialized color classifier.
Figure 6. Architecture of the specialized color classifier.
Sensors 22 01044 g006
Figure 7. Samples of type B stain images.
Figure 7. Samples of type B stain images.
Sensors 22 01044 g007
Figure 8. Samples of type A stain images.
Figure 8. Samples of type A stain images.
Sensors 22 01044 g008
Figure 9. Example images from the Mitos-Atypia-14 dataset.
Figure 9. Example images from the Mitos-Atypia-14 dataset.
Sensors 22 01044 g009
Figure 10. Transfer results obtained with different stain transfer methods.
Figure 10. Transfer results obtained with different stain transfer methods.
Sensors 22 01044 g010
Figure 11. Comparison of the results obtained using the specialized color classifier and embedded classifier.
Figure 11. Comparison of the results obtained using the specialized color classifier and embedded classifier.
Sensors 22 01044 g011
Figure 12. Seven randomly selected snapshots from network outcomes of seven independent runs, (a) without and (b) with the specialized color classifier using S3CGAN architecture.
Figure 12. Seven randomly selected snapshots from network outcomes of seven independent runs, (a) without and (b) with the specialized color classifier using S3CGAN architecture.
Sensors 22 01044 g012
Table 1. AUC values obtained with each stain transfer method after the transfer outcome was input into the tumor classifier.
Table 1. AUC values obtained with each stain transfer method after the transfer outcome was input into the tumor classifier.
MethodsAUC
(Simple Classifier)
AUC
(Complicated Classifier)
Reinhard [1]0.580.83
Macenko [2]0.650.85
Vahadane [3]0.770.87
Cho [7]0.730.90
StainGAN [9]0.790.91
AC-GAN structure0.810.91
S3CGAN*0.810.91
S3CGAN0.830.92
Table 2. Mean SSIM and mean PSNR values for different stain transfer methods.
Table 2. Mean SSIM and mean PSNR values for different stain transfer methods.
SSIM/p-ValuePSNR/p-Value
Reinhard [1]0.58/0.0013.4/0.00
Macenko [2]0.67/0.0014.0/0.00
Vahadane [3]0.65/0.0014.2/0.00
Cho [7]0.68/0.0020.4/0.00
StainGAN [9]0.73/0.0023.0/0.00
AC-GAN structure0.69/0.0021.1/0.00
S3CGAN*0.7524.7
S3CGAN0.7624.9
Table 3. Stain Transfer Results Obtained with the S3CGAN When Using Different β Values.
Table 3. Stain Transfer Results Obtained with the S3CGAN When Using Different β Values.
Sourceβ = 0.1β = 0.3β = 0.4β = 1
Stain A to B Sensors 22 01044 i001 Sensors 22 01044 i002 Sensors 22 01044 i003 Sensors 22 01044 i004 Sensors 22 01044 i005
Stain A to B Sensors 22 01044 i006 Sensors 22 01044 i007 Sensors 22 01044 i008 Sensors 22 01044 i009 Sensors 22 01044 i010
Stain B to A Sensors 22 01044 i011 Sensors 22 01044 i012 Sensors 22 01044 i013 Sensors 22 01044 i014 Sensors 22 01044 i015
Stain B to A Sensors 22 01044 i016 Sensors 22 01044 i017 Sensors 22 01044 i018 Sensors 22 01044 i019 Sensors 22 01044 i020
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, J.-S.; Ma, Y.-X. Stain Style Transfer for Histological Images Using S3CGAN. Sensors 2022, 22, 1044. https://doi.org/10.3390/s22031044

AMA Style

Lee J-S, Ma Y-X. Stain Style Transfer for Histological Images Using S3CGAN. Sensors. 2022; 22(3):1044. https://doi.org/10.3390/s22031044

Chicago/Turabian Style

Lee, Jiann-Shu, and Yao-Xian Ma. 2022. "Stain Style Transfer for Histological Images Using S3CGAN" Sensors 22, no. 3: 1044. https://doi.org/10.3390/s22031044

APA Style

Lee, J. -S., & Ma, Y. -X. (2022). Stain Style Transfer for Histological Images Using S3CGAN. Sensors, 22(3), 1044. https://doi.org/10.3390/s22031044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop