Next Article in Journal
Development of an In Vivo Extended-Spectrum Cephalosporin-Resistant Escherichia coli Model in Post-Weaned Pigs and Its Use in Assessment of Dietary Interventions
Next Article in Special Issue
A Fine-Grained Image Classification Approach for Dog Feces Using MC-SCMNet under Complex Backgrounds
Previous Article in Journal
Anatomical Description of Rhinoceros Iguana (Cyclura cornuta cornuta) Head by Computed Tomography, Magnetic Resonance Imaging and Gross-Sections
Previous Article in Special Issue
PolarBearVidID: A Video-Based Re-Identification Benchmark Dataset for Polar Bears
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Wound Image Segmentation: Transfer Learning from Human to Pet via Active Semi-Supervised Learning

1
Department of Physics and Astronomy, University of Bologna, 40127 Bologna, Italy
2
Department of Veterinary Medical Sciences, University of Bologna, 40064 Ozzano dell’Emilia, Italy
3
Department of Medical and Surgical Sciences, University of Bologna, 40138 Bologna, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
These authors contributed equally to this work and share seniorship.
Animals 2023, 13(6), 956; https://doi.org/10.3390/ani13060956
Submission received: 31 January 2023 / Revised: 24 February 2023 / Accepted: 6 March 2023 / Published: 7 March 2023
(This article belongs to the Special Issue Artificial Intelligence (AI) Applied to Animal Health and Welfare)

Abstract

:

Simple Summary

Appropriate wound management shortens healing times and reduces management costs, benefiting the patient in physical terms and potentially reducing the healthcare system economic burden. Artificial intelligence techniques could be used to automate the process of wound healing assessment, easing the effort required by clinicians and removing the inherent subjectivity of the evaluation. However, the training of artificial intelligence models relies on the availability of large datasets of carefully annotated data. The annotation of medical data is a time consuming and expensive process which requires the supervision of high-expertise professionals. In this work, we introduced a novel pipeline for the segmentation of pet wound images, using an advanced training strategy able to minimize human intervention for both the image annotation and wound segmentation. We implemented our solution in a novel mobile app, providing a valuable tool for pet wound treatment and a methodological approach for the generation of large image-segmentation datasets.

Abstract

Wound management is a fundamental task in standard clinical practice. Automated solutions already exist for humans, but there is a lack of applications regarding wound management for pets. Precise and efficient wound assessment is helpful to improve diagnosis and to increase the effectiveness of treatment plans for chronic wounds. In this work, we introduced a novel pipeline for the segmentation of pet wound images. Starting from a model pre-trained on human-based wound images, we applied a combination of transfer learning (TL) and active semi-supervised learning (ASSL) to automatically label a large dataset. Additionally, we provided a guideline for future applications of TL+ASSL training strategy on image datasets. We compared the effectiveness of the proposed training strategy, monitoring the performance of an EfficientNet-b3 U-Net model against the lighter solution provided by a MobileNet-v2 U-Net model. We obtained 80% of correctly segmented images after five rounds of ASSL training. The EfficientNet-b3 U-Net model significantly outperformed the MobileNet-v2 one. We proved that the number of available samples is a key factor for the correct usage of ASSL training. The proposed approach is a viable solution to reduce the time required for the generation of a segmentation dataset.

1. Introduction

Wound healing in both human and animal treatments is a complex and multidisciplinary problem [1]. The importance of precise and efficient wound assessment is helpful to improve diagnosis in acute wounds and to increase the effectiveness of treatment plans for chronic wounds [2]. The increasing usage of digital procedures in medicine is encouraging clinicians to explore new solutions to improve clinical practice and patient treatments [3,4]. Smartphone cameras provide valid and ready-to-use solutions for the acquisition of high-quality imaging data [5,6], also offering enough computational power for their real time analysis.
In veterinary medicine, as well as in humans, automated technology for wound management could help to quickly assess the severity of an open wound that may require prompt intervention in acute injuries [7,8,9]; likewise, it may objectify the state of the wound at the time of clinical presentation. Automated solutions for wound management may help in the monitoring of the subsequent wound healing process.
There are only a few AI-based solutions focusing on animal wound management, yet the global animal wound care market size was valued at USD 960.3 mln in 2020 (ref. https://www.grandviewresearch.com/industry-analysis/animal-wound-care-market, accessed on 6 March 2023). Current wound modelling studies mainly focus on laboratory mice, measuring the wound size as a key parameter [10]. Despite the growing literature on performing automated wound management applied to human patients, there is still a general lack of interest in the adaptation of the same techniques on the animal counterpart. Those automated solutions usually leverage the power of artificial neural networks. However, the deployment of these models relies on the availability of a large amount of data and manually annotated samples [11].
Features such as the size and shape of the wound areas represent fundamental quantities for the monitoring of the wound healing status and the consequent prognosis of the patient [12]. The introduction of automated solutions to address the feature extraction task could speed up the standard clinical practice and provide a standardized method of wound severity stratification. The extraction of quantitative features able to characterize the lesion in exam starts from a precise identification of the wound region of interest, i.e., the wound segmentation [13,14]. Automated wound segmentation on humans has been addressed via the development of ad hoc Convolutional Neural Network (CNN) [15,16,17] models. Equivalent solutions for pet images are still missing in the literature, probably due to poor data for the training of deep learning models.
The training of an artificial neural network usually needs supervision, but it is generally difficult to obtain great quantities of manually annotated data in clinical settings. In particular, the manual segmentation of images could, indeed, be extremely time consuming for large datasets. However, when available, manual segmentation by human experts could further suffer from imperfections, mainly caused by inter-observer variability, due to a subjective wound boundaries estimation [18].
Several approaches are able to overcome the problem of large data annotations and consequent image segmentation [19,20]. In the work of Zhou et al. [21], the authors proposed the division of each image into a series of disjointed patches, providing to the clinicians only small portions of the whole image in which to focus their manual annotation. Another interesting attempt by Mahapatra et al. [22] introduced the usage of Semi-Supervised Learning (SSL) in combination with Active Learning (AL) on Crohn’s disease; starting from only a few labelled samples, the predictions of the automated model were iteratively refined by experts in different rounds of training. Following the same training strategy, we discussed in our previous work [23] an application to human wound image segmentation using a deep learning model, obtaining state-of-the-art comparable results.
In this work, we extended our previously proposed pipeline to pet images via a transfer learning procedure. We aim to minimize the labeling effort of the clinicians, requiring no starting manual annotation at all. To perform a robust image segmentation without a labelled ground truth, we started from the model that Curti et al. [23] trained on a human wound image dataset.

2. Materials and Method

2.1. Patient Selection

A set of wound photos of dogs and cats was acquired with 4 different digital cameras at the University Veterinary Hospital of the Department of Veterinary Medical Sciences, University of Bologna, during daily clinical practice. The acquisition procedure was performed from April 2014 to June 2022, with a total of 290 selected images (PetWound dataset). The photos included spontaneous wounds of domestic dogs and cats that were brought to the institution for the treatment of the wound, and eventually other concomitant lesions. To be enrolled in the study, images must contain a cutaneous region with an open wound caused by cuts, abrasions, laceration, sores, burns, degloving injuries, avulsion wounds, or dehiscence from previous surgery. The statistics of species involved in the study are reported in Table 1.
Exclusion criteria were set from both a technical and a clinical point of view: some images were rejected because of poor quality, i.e., no sharp definition, inadequate light exposition, etc.; sometimes the kind of injuries represented could not be categorized as proper open wounds (open wound during surgery sutured wounds, and wounds with other surgical implants); some images with bloodstains or red-colored objects in the background were also excluded.

2.2. Data Acquisition

The images were acquired during clinical practice by several operators without a standardized procedure. The only guideline was to keep the wound at the center of the image. The aim was to obtain an heterogenous dataset, with different light conditions (illumination and exposition) and variable backgrounds.
All the images were stored in RGB 8-bit JPEG format, with different dimensions according to the device used. The details of the camera devices used for the acquisition are shown in Table 2.

2.3. Training Strategy

In this work, we developed an ad hoc Active Semi-Supervised Learning (ASSL) strategy, reproducing the procedure already discussed in our previous work [23]. The ASSL procedure was reiterated for several rounds of training, requiring the clinicians to simply accept or discard the output of the model after each round. We used the model trained on the human wound Deepskin dataset without any refinement to produce the initial labels for the pet wound images. This procedure substitutes the manual annotation for the initial round of the ASSL pipeline.
We used a deep CNN U-Net [24] with an EfficientNet-b3 [25] backbone as a model for automated segmentation. At each round of ASSL, the model weights were restored to the original configuration obtained by training on the Deepskin dataset, i.e., a human-based wound dataset. The use of a pre-trained configuration of the model weights during ASSL rounds facilitates the learning process, providing a solid starting point for a Transfer Learning (TL) procedure. The TL strategy allows a faster convergence of the model training, as already proved by several authors in image analyses [26,27,28].
The initial set of predicted segmentations was validated by two expert clinicians. For each validation image, the clinicians determined whether the generated segmentation was accurate according to the following binary criteria: (i) the mask must cover the entire wound area; (ii) the mask must cover only the wound area, i.e., the mask must not have holes or spurious parts; (iii) the mask shape must follow the correct wound boundaries. The validated images (and corresponding segmentation masks) which satisfied all the criteria were inserted into the training set as the starting point of the current ASSL training strategy. Ad hoc software was developed to minimize the time required for the validation procedure [29]. The validation was performed, reviewing the segmentations overlayed on the original image displayed on a high-resolution screen and without limits of time for the evaluation.
An analogous pipeline of ASSL in combination with TL was used for the training of a MobileNet-v2 U-Net [30] architecture. The MobileNet-v2 U-Net is a standard candidate for mobile implementations, and it guarantees a faster evaluation of the images with a small number of parameters to tune. The MobileNet-v2 U-Net model was pre-trained on the Deepskin dataset (obtaining segmentation performances compatible with the results achieved in our previous work [23]) and used in the ASSL training strategy as described for the EfficientNet-b3 one. The use of a smaller model as a benchmark allows a quantitative measurement of the required model complexity for the wound segmentation task in relation to the number of the available samples: small datasets are commonly analyzed with simpler deep learning models, but there are no clear guidelines about this criterion in combination with ASSL training strategy.
The implementation of the ASSL training strategy aims to progressively increase the number of correctly annotated samples. Starting from a small fraction of the original set of data labelled with the application of TL, we aimed to reach at least 80% of correct segmentations. At each round of ASSL, we randomly split the labelled data into two disjointed sets of images, using 90% of the images for the training and 10% for the performance evaluation.

2.4. Training Metrics

The efficiency of the ASSL training was evaluated according to the number of correctly segmented images after clinical evaluation. Additionally, in order to prevent possible overfitting of the model during the training stage, we monitored the efficiency of the model using standard metrics. The number of correct segmentations was evaluated on the totality of the PetWound dataset, while the efficiency metrics were quantified on 10% of images used for the performance evaluation.
In particular, the F1 and IoU metrics used for the performance evaluation are defined as:
p r e c i s i o n = T P T P + F P
r e c a l l = T P T P + F N
F 1 = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
I o U = T P T P + F N + F P
where TP, FP, and FN are the True Positive, False Positives, and False Negative scores, respectively. We trained both the models for 100 epochs at each round of ASSL, with Adam optimizer (learning rate of 10 5 ). The models were trained minimizing the Binary Focal Loss (BF) as loss function:
B F l o s s ( y t r u e ,   y p r e d ) = y p r e d α ( 1 y p r e d ) γ log ( y p r e d ) ( 1 y t r u e ) α y p r e d γ log ( 1 y p r e d )
where y t r u e   and y p r e d   are the ground truth binary mask and the predicted one, respectively. In our simulations, we used a value of α = 0.25, β = 1, and γ = 2. The data augmentation was restricted only to horizontal and vertical flips. All the simulations were performed using a 64-bit workstation machine (8 GB RAM memory and 1 CPU Intel® i5-8250U CPU, with 4 cores, and a UHD Graphics 620 Intel®). A schematic representation of the proposed pipeline is shown in Figure 1.

3. Results

3.1. Transfer Learning on EfficientNet-b3

The application of the pre-trained EfficientNet-b3 U-Net model on the PetWound dataset produced a total of 143 (49% of the whole dataset) correctly annotated images. This core set of images (identified as Round 0 of ASSL procedure) was used as a kick-start for the ASSL training strategy.
The ASSL procedure reached 80% of correctly annotated wound images after five rounds of training. The results obtained by the model along each ASSL round are reported in Table 3, expressed in terms of segmentation metric scores and number of images used for the training/testing of the model. For each round, we also reported the number of correctly annotated images according to the defined criterion.
At the final round of ASSL training, the model achieved an F1 score of 0.98 and a corresponding IoU score of 0.96. Examples of the results obtained with PetWound images are shown in Figure 2.

3.2. Transfer Learning on MobileNet-v2

The application of the pre-trained MobileNet-v2 U-Net model on the raw PetWound dataset led to the correct annotation of 113 images (39% of the whole dataset). This core set of images was used to kick-start the ASSL training strategy for a total of four rounds of training. The results obtained by the application of the MobileNet-v2 U-Net model in the ASSL pipeline are reported in Table 4.
We stopped the ASSL training strategy after four rounds of training because the model performances reached a plateau. The MobileNet-v2 U-Net model did not achieve results compatible with the EfficientNet-b3 one, showing a significant gap in terms of corrected segmentation images. Despite the introduction of the TL, the lighter model was not able to generalize on the heterogeneous PetWound dataset. These results confirmed the need for a deeper or more complex model for the use of the ASSL training strategy.

4. Discussion

In this work, we proposed the application of an ASSL training strategy in combination with TL for the automated segmentation of pet wound images. We confirmed the effectiveness of ASSL in labelling large amounts of data with minimal human effort. The introduction of TL in the training pipeline allowed us to start the ASSL without an initial manually annotated dataset, further reducing the workload required by the clinicians. In the proposed procedure, the evaluation of the experts consists only of an accept/discard method, minimizing the time and the effort required for the generation of high-quality segmentation masks.
The final model obtained with the proposed training pipeline achieved an F1 score of 0.98 and an IoU score of 0.96, confirming the goodness of the generated segmentations. As showed in Table 3, the model performances are consistent through the rounds of ASSL training with a slight improvement of the two considered metrics with respect to the values of the initial round. We must consider that the model performance could be positively biased, because at each round we were only including images for which we knew the model performed well in the previous round. However, the positive trend of the correctly segmented images along the ASSL rounds shows that the model was improving its prediction and generalization capabilities. The robustness of our training pipeline is enforced by the fact that the EfficientNet-b3 U-Net model trained on the Deepskin dataset also generalized well on the PetWound dataset, correctly annotating almost 50% of the available images without any fine-tuning of the model.
The results obtained during the ASSL using the MobileNet-v2 U-Net model were not as good as the ones obtained with the EfficientNet-b3 U-Net one. We argue that the MobileNet-v2 backbone could be not powerful enough to perform this complex segmentation task in the initial stages of ASSL, when the number of available samples is extremely limited. The MobileNet-v2 U-Net was then trained on the final set of annotations produced by the EfficientNet-b3 U-Net model, obtaining good results in terms of segmentation metrics (IoU = 0.90, F1 = 0.95 with 72% of corrected segmentations on the whole dataset). This result suggests that a simpler model, such as the MobileNet-v2 one, could need a greater number of training samples to learn the meaningful characteristics of the images in complex contests than the one of this study. This result goes against the classical belief that the optimal number of parameters involved in a deep learning model scales with the quantity of available data: the introduction of TL in the pipeline seems to remedy this problem, providing a sufficiently good starting point for the correct generalization of new data.
To our knowledge, there are no available segmentation datasets of pet wound images; thus, the PetWound images we collected, and the corresponding segmentations produced in this work, could be used as a valid starting point for the application of deep learning pipelines in similar studies. An ASSL strategy having the PetWound dataset as starting point could easily produce hundreds of other pet wound annotated images with minimal effort.
The main limitations of our dataset are the restricted number of available images and the inclusion of only two different pet species, namely dogs and cats. The expansion of the dataset with wound images of other pet species would be beneficial for the generalizability of the study. On the other hand, a strength of our dataset is the high heterogeneity of the included images, presenting different resolutions, light conditions, and backgrounds, making the dataset suitable for real clinical practice scenarios.

5. Conclusions

The aim of this work was to develop a model for the automatic segmentation of pet-wound images, starting with no manually labelled samples and using only TL and ASSL training strategies. The combination of the two training strategies proved their effectiveness in generating large amounts of annotated samples with minimal human intervention. This procedure speeds up the validation procedure by clinicians and it is proven to be a viable solution in medical analyses. This work may represent a starting point for the development of automated wound management in veterinary medicine for clinical and research activities.
We found the EfficientNet-b3 U-Net model, comparing its performances with the lighter MobileNet-v2 U-Net one, to be an optimal deep learning model for the ASSL training strategy. We also demonstrated numerically that the complexity of wound segmentation does not require complex deep learning models, showing comparable performances between the EfficientNet-b3 U-Net and the MobileNet-v2 U-Net architectures when trained on a larger set of annotated images. The inclusion of TL components in the ASSL pipeline, indeed, strengthens the generalization capabilities of the trained models.
The results obtained in this paper stand as a reliable solution to perform correct wound image segmentation. The MobileNet-v2 U-Net performances suggest that the future direction of this field could focus on implementations of smartphone-based technologies.

Author Contributions

Conceptualization, N.C., G.C. (Gianluca Carlini), C.S., D.D. and E.G.; Data curation, V.C., E.P., S.D.M., A.F., M.G. and L.P.; Formal analysis, D.B. and N.C.; Methodology, D.B., G.C. (Gianluca Carlini), C.S., D.D., G.C. (Gastone Castellani) and E.G.; Project administration, G.C. (Gastone Castellani), S.D.M., M.G., L.P. and E.G.; Resources, V.C.; Software, D.B., N.C. and G.C. (Gianluca Carlini); Supervision, G.C. (Gastone Castellani), S.D.M., M.G., L.P. and E.G.; Validation, V.C., E.P. and S.D.M.; Writing—original draft, D.B.; Writing—review and editing, N.C., G.C. (Gianluca Carlini), C.S., D.D., G.C. (Gastone Castellani), E.P., S.D.M., A.F., M.G., L.P. and E.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

This study was an observational, retrospective single-center study.

Data Availability Statement

The data used during the current study are available from the corresponding author on reasonable request. The pre-trained model and parameters used for the image segmentation are available in the Github-repository (https://github.com/Torbidos7/PetWound, accessed on 6 March 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gottrup, F. A specialized wound-healing center concept: Importance of a multidisciplinary department structure and surgical treatment facilities in the treatment of chronic wounds. Am. J. Surg. 2004, 187, S38–S43. [Google Scholar] [CrossRef] [PubMed]
  2. Anisuzzaman, D.M.; Wang, C.; Rostami, B.; Gopalakrishnan, S.; Niezgoda, J.; Yu, Z. Image-Based Artificial Intelligence in Wound Assessment: A Systematic Review. Adv. Wound Care 2022, 11, 687–709. [Google Scholar] [CrossRef]
  3. Yanase, J.; Triantaphyllou, E. A systematic survey of computer-aided diagnosis in medicine: Past and present developments. Expert Syst. Appl. 2019, 138, 112821. [Google Scholar] [CrossRef]
  4. Lucas, Y.; Niri, R.; Treuillet, S.; Douzi, H.; Castaneda, B. Wound Size Imaging: Ready for Smart Assessment and Monitoring. Adv. Wound Care 2021, 10, 641–661. [Google Scholar] [CrossRef] [PubMed]
  5. Barrett, P.M.; Topol, E.J. Smartphone Medicine. IT Prof. 2016, 18, 52–54. [Google Scholar] [CrossRef]
  6. Hernández-Neuta, I.; Neumann, F.; Brightmeyer, J.; Tis, T.B.; Madaboosi, N.; Wei, Q.; Ozcan, A.; Nilsson, M. Smartphone-based clinical diagnostics: Towards democratization of evidence-based health care. J. Intern. Med. 2019, 285, 19–39. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Basran, P.S.; Appleby, R.B. The unmet potential of artificial intelligence in veterinary medicine. Ajvr 2022, 83, 385–392. [Google Scholar] [CrossRef]
  8. Leary, D.; Basran, P.S. The role of artificial intelligence in veterinary radiation oncology. Vet Radiol. Ultrasound 2022, 63, 903–912. [Google Scholar] [CrossRef]
  9. Hennessey, E.; DiFazio, M.; Hennessey, R.; Cassel, N. Artificial intelligence in veterinary diagnostic imaging: A literature review. Vet Radiol. Ultrasound 2022, 63, 851–870. [Google Scholar] [CrossRef]
  10. Carrión, H.; Jafari, M.; Bagood, M.D.; Yang, H.Y.; Isseroff, R.R.; Gomez, M. Automatic wound detection and size estimation using deep learning algorithms. PLoS Comput. Biol. 2022, 18, e1009852. [Google Scholar] [CrossRef]
  11. Tajbakhsh, N.; Jeyaseelan, L.; Li, Q.; Chiang, J.N.; Wu, Z.; Ding, X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med. Image Anal. 2020, 63, 101693. [Google Scholar] [CrossRef] [Green Version]
  12. Gethin, G. The importance of continuous wound measuring. Wounds UK 2006, 2, 60–68. [Google Scholar]
  13. Chowdhary, C.L.; Acharjya, D.P. Segmentation and Feature Extraction in Medical Imaging: A Systematic Review. Procedia Comput. Sci. 2020, 167, 26–36. [Google Scholar] [CrossRef]
  14. Sklansky, J. Image Segmentation and Feature Extraction. IEEE Trans Syst. Man. Cybern. 1978, 8, 237–247. [Google Scholar] [CrossRef]
  15. Wang, C.; Yan, X.; Smith, M.; Kochhar, K.; Rubin, M.; Warren, S.M.; Wrobel, J.; Lee, H. A unified framework for automatic wound segmentation and analysis with deep convolutional neural networks. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 5 November 2015; pp. 2415–2418. [Google Scholar] [CrossRef]
  16. Scebba, G.; Zhang, J.; Catanzaro, S.; Mihai, C.; Distler, O.; Berli, M.; Karlen, W. Detect-and-segment: A deep learning approach to automate wound image segmentation. Inform. Med. Unlocked 2022, 29, 100884. [Google Scholar] [CrossRef]
  17. Wang, C.; Anisuzzaman, D.M.; Williamson, V.; Dhar, M.K.; Rostami, B.; Niezigoda, Z.; Sandeep, G.; Yu, Z. Fully automatic wound segmentation with deep convolutional neural networks. Sci. Rep. 2020, 10, 21897. [Google Scholar] [CrossRef] [PubMed]
  18. Haghpanah, S.; Bogie, K.; Wang, X.; Banks, P.G.; Ho, C.H. Reliability of Electronic Versus Manual Wound Measurement Techniques. Arch. Phys. Med. Rehabil. 2006, 87, 1396–1402. [Google Scholar] [CrossRef] [PubMed]
  19. Lahiri, A.; Jain, V.; Mondal, A.; Biswas, P.K. Retinal Vessel Segmentation Under Extreme Low Annotation: A Gan Based Semi-Supervised Approach. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Virtual, 25–28 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 418–422. [Google Scholar] [CrossRef]
  20. Zheng, H.; Yang, L.; Chen, J.; Han, J.; Zhang, Y.; Liang, P.; Zhao, Z.; Wang, C.; Chen, D.Z. Biomedical Image Segmentation via Representative Annotation. AAAI 2019, 33, 5901–5908. [Google Scholar] [CrossRef] [Green Version]
  21. Zhou, J.; Cao, R.; Kang, J.; Guo, K.; Xu, Y. An Efficient High-Quality Medical Lesion Image Data Labeling Method Based on Active Learning. IEEE Access 2020, 8, 144331–144342. [Google Scholar] [CrossRef]
  22. Mahapatra, D.; Schüffler, P.J.; Tielbeek, J.A.W.; Vos, F.M.; Buhmann, J.M. Semi-Supervised and Active Learning for Automatic Segmentation of Crohn’s Disease. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2013; Mori, K., Sakuma, I., Sato, Y., Barillot, C., Navab, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 214–221. [Google Scholar]
  23. Curti, N.; Merli, Y.; Zengarini, C.; Giampieri, E.; Merlotti, A.; Dall’Olio, D.; Marcelli, E.; Bianchi, T.; Castellani, G. Effectiveness of Semi-Supervised Active Learning in Automated Wound Image Segmentation. Int. J. Mol. Sci. 2023, 24, 706. [Google Scholar] [CrossRef]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arxiv 2015, arXiv:1505.04597. [Google Scholar] [CrossRef]
  25. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arxiv 2019, arXiv:1905.11946. [Google Scholar] [CrossRef]
  26. Yang, L.; Hanneke, S.; Carbonell, J. A theory of transfer learning with applications to active learning. Mach. Learn. 2013, 90, 161–189. [Google Scholar] [CrossRef]
  27. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Transfer learning for medical image classification: A literature review. BMC Med. Imaging 2022, 22, 69. [Google Scholar] [CrossRef]
  28. Kora, P.; Ooi, C.P.; Faust, O.; Raghavendra, U.; Gudigar, A.; Chan, W.Y.; Meenakshi, K.; Swaraja, K.; Plawiak, P.; Acharya, U.R. Transfer learning techniques for medical image analysis: A review. Biocybern. Biomed. Eng. 2022, 42, 79–107. [Google Scholar] [CrossRef]
  29. Curti, N. Active_Learning_Validator. Utility HTML+JS Script for the Validation of Images in Active Semi-Supervised Learning Projects. Available online: https://github.com/Nico-Curti/active_learning_validator (accessed on 6 March 2023).
  30. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. arxiv 2019, arXiv:1801.04381. Available online: http://arxiv.org/abs/1801.04381 (accessed on 11 January 2023).
Figure 1. Representation of the pipeline implemented for the training of the EfficientNet-b3 model. (a) The images acquired during clinical practice are used as data for the ASSL procedure. (b) At the first round, the initial set of labelled images was obtained by the application of the model pre-trained on the Deepskin dataset, without any refinement of the model parameters. (c) The obtained set of labelled images was used as the kick-start for the ASSL training strategy, performing a TL from the human-based wounds to the pet ones. At each round of training, the model was reset to the initial conditions, i.e., the parameters obtained by the training on the Deepskin dataset.
Figure 1. Representation of the pipeline implemented for the training of the EfficientNet-b3 model. (a) The images acquired during clinical practice are used as data for the ASSL procedure. (b) At the first round, the initial set of labelled images was obtained by the application of the model pre-trained on the Deepskin dataset, without any refinement of the model parameters. (c) The obtained set of labelled images was used as the kick-start for the ASSL training strategy, performing a TL from the human-based wounds to the pet ones. At each round of training, the model was reset to the initial conditions, i.e., the parameters obtained by the training on the Deepskin dataset.
Animals 13 00956 g001
Figure 2. Example of segmentations obtained by the EfficientNet-b3 model at the end of the 5th round of ASSL training. In green, we marked the contours of the segmentation masks produced by the model, highlighting with the alpha level the region of interest identified. The aspect ratio of the images was manually adjusted to be coherent among them.
Figure 2. Example of segmentations obtained by the EfficientNet-b3 model at the end of the 5th round of ASSL training. In green, we marked the contours of the segmentation masks produced by the model, highlighting with the alpha level the region of interest identified. The aspect ratio of the images was manually adjusted to be coherent among them.
Animals 13 00956 g002
Table 1. Description of the images involved in the study. The number of images is split according to the pet species of the patients. The three rows of the table show the number of initial images collected; the number of images after exclusion of inadequate ones; the different wounds considered starting from the included images, since the same wound could have been acquired at different time points in dogs and cats.
Table 1. Description of the images involved in the study. The number of images is split according to the pet species of the patients. The three rows of the table show the number of initial images collected; the number of images after exclusion of inadequate ones; the different wounds considered starting from the included images, since the same wound could have been acquired at different time points in dogs and cats.
DogCatTot.
N° initial images301105406
N° included images20882290
N° included wounds13047177
Table 2. Main specifics of the camera devices used for the acquisition of the photos. The PetWound database is composed of images acquired with 4 different camera devices. In detail, there are 3 different smartphone digital cameras and 1 Olympus digital camera.
Table 2. Main specifics of the camera devices used for the acquisition of the photos. The PetWound database is composed of images acquired with 4 different camera devices. In detail, there are 3 different smartphone digital cameras and 1 Olympus digital camera.
Redmi Note 9 PRORedmi Note 5ASUS Z017DOlympus
Imaging CORP u770sW
F-stopf/1.9f/1.9f/2f/3.5
Exposition1/33 s1/25 s1/100 s1/15 s
Iso sensibilityISO-330ISO-200ISO 71ISO-71
Focal distance5 mm4 mm4 mm4 mm
Focal length25 mm24 mmNANA
Table 3. Results obtained by the EfficientNet-b3 model in the ASSL training strategy on the PetWound dataset. For each round, we reported the number of images used for the training and validation of the model, the number of correctly annotated images, and the segmentation metric scores. We considered as Round 0 the one performed with the model trained only on the Deepskin dataset.
Table 3. Results obtained by the EfficientNet-b3 model in the ASSL training strategy on the PetWound dataset. For each round, we reported the number of images used for the training and validation of the model, the number of correctly annotated images, and the segmentation metric scores. We considered as Round 0 the one performed with the model trained only on the Deepskin dataset.
Round 1Round 2Round 3Round 4Round 5
N° training images127 (88%)155 (86.5%)170 (87.5%)197 (88.2%)203 (88.5%)
N° validation images16 (12%)24 (13.5%)24 (12.5%)24 (10.8%)24 (10.5%)
N° correct segmentation179 (62%)194 (67%)221 (76%)227 (78%)232 (80%)
F1 score0.980.980.980.980.98
IoU score0.950.960.960.960.96
Table 4. Results obtained by the MobileNet-v2 U-Net model in the ASSL training strategy on the PetWound dataset. For each round, we reported the number of images used for the training and validation of the model, the number of correctly annotated images, and the segmentation metric scores. We considered as Round 0 the one performed with the model trained only on the Deepskin dataset.
Table 4. Results obtained by the MobileNet-v2 U-Net model in the ASSL training strategy on the PetWound dataset. For each round, we reported the number of images used for the training and validation of the model, the number of correctly annotated images, and the segmentation metric scores. We considered as Round 0 the one performed with the model trained only on the Deepskin dataset.
Round 1Round 2Round 3Round 4
N° training
images
97 (86%)119 (88%)122 (88.5%)127 (88.5%)
N° validation
images
16 (14%)16 (12%)16 (11.5%)16 (11.5%)
N° correct
segmentation
135 (47%)138 (48%)143 (49%)145 (50%)
F1 score0.970.920.930.94
IoU score0.940.850.870.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Buschi, D.; Curti, N.; Cola, V.; Carlini, G.; Sala, C.; Dall’Olio, D.; Castellani, G.; Pizzi, E.; Del Magno, S.; Foglia, A.; et al. Automated Wound Image Segmentation: Transfer Learning from Human to Pet via Active Semi-Supervised Learning. Animals 2023, 13, 956. https://doi.org/10.3390/ani13060956

AMA Style

Buschi D, Curti N, Cola V, Carlini G, Sala C, Dall’Olio D, Castellani G, Pizzi E, Del Magno S, Foglia A, et al. Automated Wound Image Segmentation: Transfer Learning from Human to Pet via Active Semi-Supervised Learning. Animals. 2023; 13(6):956. https://doi.org/10.3390/ani13060956

Chicago/Turabian Style

Buschi, Daniele, Nico Curti, Veronica Cola, Gianluca Carlini, Claudia Sala, Daniele Dall’Olio, Gastone Castellani, Elisa Pizzi, Sara Del Magno, Armando Foglia, and et al. 2023. "Automated Wound Image Segmentation: Transfer Learning from Human to Pet via Active Semi-Supervised Learning" Animals 13, no. 6: 956. https://doi.org/10.3390/ani13060956

APA Style

Buschi, D., Curti, N., Cola, V., Carlini, G., Sala, C., Dall’Olio, D., Castellani, G., Pizzi, E., Del Magno, S., Foglia, A., Giunti, M., Pisoni, L., & Giampieri, E. (2023). Automated Wound Image Segmentation: Transfer Learning from Human to Pet via Active Semi-Supervised Learning. Animals, 13(6), 956. https://doi.org/10.3390/ani13060956

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop