Deciphering Medicine: The Role of Explainable Artificial Intelligence in Healthcare Innovations

A special issue of Bioengineering (ISSN 2306-5354). This special issue belongs to the section "Biosignal Processing".

Deadline for manuscript submissions: closed (31 January 2025) | Viewed by 15574

Special Issue Editors


E-Mail Website
Guest Editor
Department of Bioengineering, Speed School of Engineering, University of Louisville, Louisville, KY 40292, USA
Interests: medical image analysis; artificial intelligence in medicine; deep learning; computer-aided diagnostics; precsion medicine; diagnostics and prognostic markers; bigdata in medicine
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Computers and Control Systems Engineering, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt
Interests: artificial intelligence (AI); machine learning; deep learning; robotics; metaheuristics; computer-assisted diagnosis systems; computer vision; bioinspired optimization algorithms; smart systems engineering
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

In an era where artificial intelligence (AI) is rapidly transforming the landscape of healthcare, the need for transparency and understandability in AI algorithms is critical. This Special Issue, "Deciphering Medicine: The Role of Explainable Artificial Intelligence in Healthcare Innovations", seeks to bridge the gap between advanced AI technologies and their practical, ethical, and efficient application in medical settings.

Focus and Scope:

We invite authors to submit original research, reviews, and insightful studies that focus on the development, implementation, and evaluation of explainable AI systems in medical diagnostics and treatment. This issue aims to highlight innovative methodologies, case studies, and frameworks that enhance the interpretability and transparency of AI models, thereby fostering trust and reliability among healthcare professionals and patients.

Key Themes:

  • The development of explainable AI models for diagnosis, prognosis, and treatment planning.
  • Ethical implications and considerations in deploying AI in medical settings.
  • Case studies showcasing the successful implementation of explainable AI in clinical practice.
  • Advances in machine learning and deep learning that enhance transparency and interpretability.
  • The integration of AI with traditional medical knowledge to improve patient outcomes.
  • User-centric approaches to designing explainable AI systems in healthcare.
  • Regulatory and policy perspectives on the use of AI in medical diagnostics and treatment.

Submissions:

We welcome submissions from researchers, practitioners, and thought leaders in the fields of computer science, medical informatics, bioengineering, and related disciplines. Articles should emphasize not only the technological aspects of AI, but also its practical implications, user experience, and ethical considerations in a medical context.

By focusing on explainable AI in healthcare, this Special Issue aims to illuminate the path towards the more transparent, ethical, and effective integration of AI in medicine, ultimately contributing to improved patient care and healthcare outcomes.

Dr. Mohamed Shehata
Prof. Dr. Mostafa Elhosseini
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Bioengineering is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • deciphering medicine
  • explainable AI
  • machine learning
  • deep learning
  • medical diagnostics and treatment

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (9 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Review

13 pages, 2076 KiB  
Article
Use of Multimodal Artificial Intelligence in Surgical Instrument Recognition
by Syed Ali Haider, Olivia A. Ho, Sahar Borna, Cesar A. Gomez-Cabello, Sophia M. Pressman, Dave Cole, Ajai Sehgal, Bradley C. Leibovich and Antonio Jorge Forte
Bioengineering 2025, 12(1), 72; https://doi.org/10.3390/bioengineering12010072 - 15 Jan 2025
Viewed by 596
Abstract
Accurate identification of surgical instruments is crucial for efficient workflows and patient safety within the operating room, particularly in preventing complications such as retained surgical instruments. Artificial Intelligence (AI) models have shown the potential to automate this process. This study evaluates the accuracy [...] Read more.
Accurate identification of surgical instruments is crucial for efficient workflows and patient safety within the operating room, particularly in preventing complications such as retained surgical instruments. Artificial Intelligence (AI) models have shown the potential to automate this process. This study evaluates the accuracy of publicly available Large Language Models (LLMs)—ChatGPT-4, ChatGPT-4o, and Gemini—and a specialized commercial mobile application, Surgical-Instrument Directory (SID 2.0), in identifying surgical instruments from images. The study utilized a dataset of 92 high-resolution images of 25 surgical instruments (retractors, forceps, scissors, and trocars) photographed from multiple angles. Model performance was evaluated using accuracy, weighted precision, recall, and F1 score. ChatGPT-4o exhibited the highest accuracy (89.1%) in categorizing instruments (e.g., scissors, forceps). SID 2.0 (77.2%) and ChatGPT-4 (76.1%) achieved comparable accuracy, while Gemini (44.6%) demonstrated lower accuracy in this task. For precise subtype identification of instrument names (like “Mayo scissors” or “Kelly forceps”), all models had low accuracy, with SID 2.0 having an accuracy of 39.1%, followed by ChatGPT-4o (33.69%). Subgroup analysis revealed ChatGPT-4 and 4o recognized trocars in all instances. Similarly, Gemini identified surgical scissors in all instances. In conclusion, publicly available LLMs can reliably identify surgical instruments at the category level, with ChatGPT-4o demonstrating an overall edge. However, precise subtype identification remains a challenge for all models. These findings highlight the potential of AI-driven solutions to enhance surgical-instrument management and underscore the need for further refinements to improve accuracy and support patient safety. Full article
Show Figures

Figure 1

24 pages, 7865 KiB  
Article
Population and Co-Occurrence Characteristics of Diagnoses and Comorbidities in Coronary Artery Disease Patients: A Case Study from a Hospital in Guangxi, China
by Jiaojiao Wang, Zhixuan Qi, Xiliang Liu, Xin Li, Zhidong Cao, Daniel Dajun Zeng and Hong Wang
Bioengineering 2024, 11(12), 1284; https://doi.org/10.3390/bioengineering11121284 - 18 Dec 2024
Viewed by 748
Abstract
Coronary artery disease (CAD) remains a major global health concern, significantly contributing to morbidity and mortality. This study aimed to investigate the co-occurrence patterns of diagnoses and comorbidities in CAD patients using a network-based approach. A retrospective analysis was conducted on 195 hospitalized [...] Read more.
Coronary artery disease (CAD) remains a major global health concern, significantly contributing to morbidity and mortality. This study aimed to investigate the co-occurrence patterns of diagnoses and comorbidities in CAD patients using a network-based approach. A retrospective analysis was conducted on 195 hospitalized CAD patients from a single hospital in Guangxi, China, with data collected on age, sex, and comorbidities. Network analysis, supported by sensitivity analysis, revealed key diagnostic clusters and comorbidity hubs, with hypertension emerging as the central node in the co-occurrence network. Unstable angina and myocardial infarction were identified as central diagnoses, frequently co-occurring with metabolic conditions such as diabetes. The results also highlighted significant age- and sex-specific differences in CAD diagnoses and comorbidities. Sensitivity analysis confirmed the robustness of the network structure and identified clusters, despite the limitations of sample size and data source. Modularity analysis uncovered distinct clusters, illustrating the complex interplay between cardiovascular and metabolic disorders. These findings provide valuable insights into the relationships between CAD and its comorbidities, emphasizing the importance of integrated, personalized management strategies. Future studies with larger, multi-center datasets and longitudinal designs are needed to validate these results and explore the temporal dynamics of CAD progression. Full article
Show Figures

Figure 1

12 pages, 3323 KiB  
Article
Identifying Ventricular Dysfunction Indicators in Electrocardiograms via Artificial Intelligence-Driven Analysis
by Hisaki Makimoto, Takayuki Okatani, Masanori Suganuma, Tomoyuki Kabutoya, Takahide Kohro, Yukiko Agata, Yukiyo Ogata, Kenji Harada, Redi Llubani, Alexandru Bejinariu, Obaida R. Rana, Asuka Makimoto, Elisabetha Gharib, Anita Meissner, Malte Kelm and Kazuomi Kario
Bioengineering 2024, 11(11), 1069; https://doi.org/10.3390/bioengineering11111069 - 26 Oct 2024
Viewed by 941
Abstract
Recent studies highlight artificial intelligence’s ability to identify ventricular dysfunction via electrocardiograms (ECGs); however, specific indicative waveforms remain unclear. This study analysed ECG and echocardiography data from 17,422 cases in Japan and Germany. We developed 10-layer convolutional neural networks to detect left ventricular [...] Read more.
Recent studies highlight artificial intelligence’s ability to identify ventricular dysfunction via electrocardiograms (ECGs); however, specific indicative waveforms remain unclear. This study analysed ECG and echocardiography data from 17,422 cases in Japan and Germany. We developed 10-layer convolutional neural networks to detect left ventricular ejection fractions below 50%, using four-fold cross-validation. Model performance, evaluated among different ECG configurations (3 s strips, single-beat, and two-beat overlay) and segments (PQRST, QRST, P, QRS, and PQRS), showed two-beat ECGs performed best, followed by single-beat models, surpassing 3 s models in both internal and external validations. Single-beat models revealed limb leads, particularly I and aVR, as most indicative of dysfunction. An analysis indicated segments from QRS to T-wave were most revealing, with P segments enhancing model performance. This study confirmed that dual-beat ECGs enabled the most precise ventricular function classification, and segments from the P- to T-wave in ECGs were more effective for assessing ventricular dysfunction, with leads I and aVR offering higher diagnostic utility. Full article
Show Figures

Graphical abstract

16 pages, 2605 KiB  
Article
Applying a Deep Learning Model for Total Kidney Volume Measurement in Autosomal Dominant Polycystic Kidney Disease
by Jia-Lien Hsu, Anandakumar Singaravelan, Chih-Yun Lai, Zhi-Lin Li, Chia-Nan Lin, Wen-Shuo Wu, Tze-Wah Kao and Pei-Lun Chu
Bioengineering 2024, 11(10), 963; https://doi.org/10.3390/bioengineering11100963 - 26 Sep 2024
Viewed by 1126
Abstract
Background: Autosomal dominant polycystic kidney disease (ADPKD) is the most common hereditary renal disease leading to end-stage renal disease. Total kidney volume (TKV) measurement has been considered as a surrogate in the evaluation of disease severity and prognostic predictor of ADPKD. However, the [...] Read more.
Background: Autosomal dominant polycystic kidney disease (ADPKD) is the most common hereditary renal disease leading to end-stage renal disease. Total kidney volume (TKV) measurement has been considered as a surrogate in the evaluation of disease severity and prognostic predictor of ADPKD. However, the traditional manual measurement of TKV by medical professionals is labor-intensive, time-consuming, and human error prone. Materials and methods: In this investigation, we conducted TKV measurements utilizing magnetic resonance imaging (MRI) data. The dataset consisted of 30 patients with ADPKD and 10 healthy individuals. To calculate TKV, we trained models using both coronal- and axial-section MRI images. The process involved extracting images in Digital Imaging and Communications in Medicine (DICOM) format, followed by augmentation and labeling. We employed a U-net model for image segmentation, generating mask images of the target areas. Subsequent post-processing steps and TKV estimation were performed based on the outputs obtained from these mask images. Results: The average TKV, as assessed by medical professionals from the testing dataset, was 1501.84 ± 965.85 mL with axial-section images and 1740.31 ± 1172.21 mL with coronal-section images, respectively (p = 0.73). Utilizing the deep learning model, the mean TKV derived from axial- and coronal-section images was 1536.33 ± 958.68 mL and 1636.25 ± 964.67 mL, respectively (p = 0.85). The discrepancy in mean TKV between medical professionals and the deep learning model was 44.23 ± 58.69 mL with axial-section images (p = 0.8) and 329.12 ± 352.56 mL with coronal-section images (p = 0.9), respectively. The average variability in TKV measurement was 21.6% with the coronal-section model and 3.95% with the axial-section model. The axial-section model demonstrated a mean Dice Similarity Coefficient (DSC) of 0.89 ± 0.27 and an average patient-wise Jaccard coefficient of 0.86 ± 0.27, while the mean DSC and Jaccard coefficient of the coronal-section model were 0.82 ± 0.29 and 0.77 ± 0.31, respectively. Conclusion: The integration of deep learning into image processing and interpretation is becoming increasingly prevalent in clinical practice. In our pilot study, we conducted a comparative analysis of the performance of a deep learning model alongside corresponding axial- and coronal-section models, a comparison that has been less explored in prior research. Our findings suggest that our deep learning model for TKV measurement performs comparably to medical professionals. However, we observed that varying image orientations could introduce measurement bias. Specifically, our AI model exhibited superior performance with axial-section images compared to coronal-section images. Full article
Show Figures

Graphical abstract

24 pages, 13781 KiB  
Article
High-Precision Skin Disease Diagnosis through Deep Learning on Dermoscopic Images
by Sadia Ghani Malik, Syed Shahryar Jamil, Abdul Aziz, Sana Ullah, Inam Ullah and Mohammed Abohashrh
Bioengineering 2024, 11(9), 867; https://doi.org/10.3390/bioengineering11090867 - 27 Aug 2024
Cited by 1 | Viewed by 3252
Abstract
Dermatological conditions are primarily prevalent in humans and are primarily caused by environmental and climatic fluctuations, as well as various other reasons. Timely identification is the most effective remedy to avert minor ailments from escalating into severe conditions. Diagnosing skin illnesses is consistently [...] Read more.
Dermatological conditions are primarily prevalent in humans and are primarily caused by environmental and climatic fluctuations, as well as various other reasons. Timely identification is the most effective remedy to avert minor ailments from escalating into severe conditions. Diagnosing skin illnesses is consistently challenging for health practitioners. Presently, they rely on conventional methods, such as examining the condition of the skin. State-of-the-art technologies can enhance the accuracy of skin disease diagnosis by utilizing data-driven approaches. This paper presents a Computer Assisted Diagnosis (CAD) framework that has been developed to detect skin illnesses at an early stage. We suggest a computationally efficient and lightweight deep learning model that utilizes a CNN architecture. We then do thorough experiments to compare the performance of shallow and deep learning models. The CNN model under consideration consists of seven convolutional layers and has obtained an accuracy of 87.64% when applied to three distinct disease categories. The studies were conducted using the International Skin Imaging Collaboration (ISIC) dataset, which exclusively consists of dermoscopic images. This study enhances the field of skin disease diagnostics by utilizing state-of-the-art technology, attaining exceptional levels of accuracy, and striving for efficiency improvements. The unique features and future considerations of this technology create opportunities for additional advancements in the automated diagnosis of skin diseases and tailored treatment. Full article
Show Figures

Figure 1

24 pages, 9356 KiB  
Article
Deciphering Factors Contributing to Cost-Effective Medicine Using Machine Learning
by Bowen Long, Jinfeng Zhou, Fangya Tan and Srikar Bellur
Bioengineering 2024, 11(8), 818; https://doi.org/10.3390/bioengineering11080818 - 12 Aug 2024
Viewed by 1192
Abstract
This study uses machine learning to identify critical factors influencing the cost-effectiveness of over-the-counter (OTC) medications. By developing a novel cost-effectiveness rating (CER) based on user ratings and prices, we analyzed data from Amazon. The findings indicate that Flexible Spending Account (FSA)/Health Savings [...] Read more.
This study uses machine learning to identify critical factors influencing the cost-effectiveness of over-the-counter (OTC) medications. By developing a novel cost-effectiveness rating (CER) based on user ratings and prices, we analyzed data from Amazon. The findings indicate that Flexible Spending Account (FSA)/Health Savings Account (HSA) eligibility, symptom treatment range, safety warnings, special effects, active ingredients, and packaging size significantly impact cost-effectiveness across cold, allergy, digestion, and pain relief medications. Medications eligible for FSA or HSA funds, treating a broader range of symptoms, and having smaller packaging are perceived as more cost-effective. Cold medicines with safety warnings were cost-effective due to their lower average price and effective ingredients like phenylephrine and acetaminophen. Allergy medications with kid-friendly features showed higher cost-effectiveness, and ingredients like calcium, famotidine, and magnesium boosted the cost-effectiveness of digestion medicines. These insights help consumers make informed purchasing decisions and assist manufacturers and retailers in enhancing product competitiveness. Overall, this research supports better decision-making in the pharmaceutical industry by highlighting factors that drive cost-effective medication purchases. Full article
Show Figures

Figure 1

21 pages, 3747 KiB  
Article
ViT-PSO-SVM: Cervical Cancer Predication Based on Integrating Vision Transformer with Particle Swarm Optimization and Support Vector Machine
by Abdulaziz AlMohimeed, Mohamed Shehata, Nora El-Rashidy, Sherif Mostafa, Amira Samy Talaat and Hager Saleh
Bioengineering 2024, 11(7), 729; https://doi.org/10.3390/bioengineering11070729 - 18 Jul 2024
Cited by 3 | Viewed by 1767
Abstract
Cervical cancer (CCa) is the fourth most prevalent and common cancer affecting women worldwide, with increasing incidence and mortality rates. Hence, early detection of CCa plays a crucial role in improving outcomes. Non-invasive imaging procedures with good diagnostic performance are desirable and have [...] Read more.
Cervical cancer (CCa) is the fourth most prevalent and common cancer affecting women worldwide, with increasing incidence and mortality rates. Hence, early detection of CCa plays a crucial role in improving outcomes. Non-invasive imaging procedures with good diagnostic performance are desirable and have the potential to lessen the degree of intervention associated with the gold standard, biopsy. Recently, artificial intelligence-based diagnostic models such as Vision Transformers (ViT) have shown promising performance in image classification tasks, rivaling or surpassing traditional convolutional neural networks (CNNs). This paper studies the effect of applying a ViT to predict CCa using different image benchmark datasets. A newly developed approach (ViT-PSO-SVM) was presented for boosting the results of the ViT based on integrating the ViT with particle swarm optimization (PSO), and support vector machine (SVM). First, the proposed framework extracts features from the Vision Transformer. Then, PSO is used to reduce the complexity of extracted features and optimize feature representation. Finally, a softmax classification layer is replaced with an SVM classification model to precisely predict CCa. The models are evaluated using two benchmark cervical cell image datasets, namely SipakMed and Herlev, with different classification scenarios: two, three, and five classes. The proposed approach achieved 99.112% accuracy and 99.113% F1-score for SipakMed with two classes and achieved 97.778% accuracy and 97.805% F1-score for Herlev with two classes outperforming other Vision Transformers, CNN models, and pre-trained models. Finally, GradCAM is used as an explainable artificial intelligence (XAI) tool to visualize and understand the regions of a given image that are important for a model’s prediction. The obtained experimental results demonstrate the feasibility and efficacy of the developed ViT-PSO-SVM approach and hold the promise of providing a robust, reliable, accurate, and non-invasive diagnostic tool that will lead to improved healthcare outcomes worldwide. Full article
Show Figures

Graphical abstract

16 pages, 11727 KiB  
Article
Toward Interpretable Cell Image Representation and Abnormality Scoring for Cervical Cancer Screening Using Pap Smears
by Yu Ando, Junghwan Cho, Nora Jee-Young Park, Seokhwan Ko and Hyungsoo Han
Bioengineering 2024, 11(6), 567; https://doi.org/10.3390/bioengineering11060567 - 4 Jun 2024
Cited by 1 | Viewed by 1104
Abstract
Screening is critical for prevention and early detection of cervical cancer but it is time-consuming and laborious. Supervised deep convolutional neural networks have been developed to automate pap smear screening and the results are promising. However, the interest in using only normal samples [...] Read more.
Screening is critical for prevention and early detection of cervical cancer but it is time-consuming and laborious. Supervised deep convolutional neural networks have been developed to automate pap smear screening and the results are promising. However, the interest in using only normal samples to train deep neural networks has increased owing to the class imbalance problems and high-labeling costs that are both prevalent in healthcare. In this study, we introduce a method to learn explainable deep cervical cell representations for pap smear cytology images based on one-class classification using variational autoencoders. Findings demonstrate that a score can be calculated for cell abnormality without training models with abnormal samples, and we localize abnormality to interpret our results with a novel metric based on absolute difference in cross-entropy in agglomerative clustering. The best model that discriminates squamous cell carcinoma (SCC) from normals gives 0.908±0.003 area under operating characteristic curve (AUC) and one that discriminates high-grade epithelial lesion (HSIL) 0.920±0.002 AUC. Compared to other clustering methods, our method enhances the V-measure and yields higher homogeneity scores, which more effectively isolate different abnormality regions, aiding in the interpretation of our results. Evaluation using an external dataset shows that our model can discriminate abnormality without the need for additional training of deep models. Full article
Show Figures

Figure 1

Review

Jump to: Research

18 pages, 743 KiB  
Review
Artificial Intelligence Support for Informal Patient Caregivers: A Systematic Review
by Sahar Borna, Michael J. Maniaci, Clifton R. Haider, Cesar A. Gomez-Cabello, Sophia M. Pressman, Syed Ali Haider, Bart M. Demaerschalk, Jennifer B. Cowart and Antonio Jorge Forte
Bioengineering 2024, 11(5), 483; https://doi.org/10.3390/bioengineering11050483 - 12 May 2024
Cited by 1 | Viewed by 2985
Abstract
This study aims to explore how artificial intelligence can help ease the burden on caregivers, filling a gap in current research and healthcare practices due to the growing challenge of an aging population and increased reliance on informal caregivers. We conducted a search [...] Read more.
This study aims to explore how artificial intelligence can help ease the burden on caregivers, filling a gap in current research and healthcare practices due to the growing challenge of an aging population and increased reliance on informal caregivers. We conducted a search with Google Scholar, PubMed, Scopus, IEEE Xplore, and Web of Science, focusing on AI and caregiving. Our inclusion criteria were studies where AI supports informal caregivers, excluding those solely for data collection. Adhering to PRISMA 2020 guidelines, we eliminated duplicates and screened for relevance. From 947 initially identified articles, 10 met our criteria, focusing on AI’s role in aiding informal caregivers. These studies, conducted between 2012 and 2023, were globally distributed, with 80% employing machine learning. Validation methods varied, with Hold-Out being the most frequent. Metrics across studies revealed accuracies ranging from 71.60% to 99.33%. Specific methods, like SCUT in conjunction with NNs and LibSVM, showcased accuracy between 93.42% and 95.36% as well as F-measures spanning 93.30% to 95.41%. AUC values indicated model performance variability, ranging from 0.50 to 0.85 in select models. Our review highlights AI’s role in aiding informal caregivers, showing promising results despite different approaches. AI tools provide smart, adaptive support, improving caregivers’ effectiveness and well-being. Full article
Show Figures

Graphical abstract

Back to TopTop