Next Issue
Volume 7, March
Previous Issue
Volume 6, September
 
 

Mach. Learn. Knowl. Extr., Volume 6, Issue 4 (December 2024) – 34 articles

Cover Story (view full-size image): Brain tumors are among the deadliest elements of cancers, and early detection is crucial for improving patient outcomes. Although an MRI is the gold standard for diagnosing brain tumors, manual analysis is often affected by radiologist fatigue and subjectivity. This study introduces a novel computer-aided diagnosis (CAD) framework for multi-class brain tumor classification from MRI scans. The framework leverages pre-trained deep learning models and explainable AI techniques to enhance both diagnostic accuracy and interpretability. A user-friendly detection system ensures seamless clinical integration. Evaluated on a public benchmark dataset, the system achieves nearly 99% accuracy, offering significant promise in improving diagnostic precision and facilitating timely interventions. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
17 pages, 986 KiB  
Article
Reliable and Faithful Generative Explainers for Graph Neural Networks
by Yiqiao Li, Jianlong Zhou, Boyuan Zheng, Niusha Shafiabady and Fang Chen
Mach. Learn. Knowl. Extr. 2024, 6(4), 2913-2929; https://doi.org/10.3390/make6040139 - 18 Dec 2024
Viewed by 635
Abstract
Graph neural networks (GNNs) have been effectively implemented in a variety of real-world applications, although their underlying work mechanisms remain a mystery. To unveil this mystery and advocate for trustworthy decision-making, many GNN explainers have been proposed. However, existing explainers often face significant [...] Read more.
Graph neural networks (GNNs) have been effectively implemented in a variety of real-world applications, although their underlying work mechanisms remain a mystery. To unveil this mystery and advocate for trustworthy decision-making, many GNN explainers have been proposed. However, existing explainers often face significant challenges, such as the following: (1) explanations being tied to specific instances; (2) limited generalisability to unseen graphs; (3) potential generation of invalid graph structures; and (4) restrictions to particular tasks (e.g., node classification, graph classification). To address these challenges, we propose a novel explainer, GAN-GNNExplainer, which employs a generator to produce explanations and a discriminator to oversee the generation process, enhancing the reliability of the outputs. Despite its advantages, GAN-GNNExplainer still struggles with generating faithful explanations and underperforms on real-world datasets. To overcome these shortcomings, we introduce ACGAN-GNNExplainer, an approach that improves upon GAN-GNNExplainer by using a more robust discriminator that consistently monitors the generation process, thereby producing explanations that are both reliable and faithful. Extensive experiments on both synthetic and real-world graph datasets demonstrate the superiority of our proposed methods over existing GNN explainers. Full article
Show Figures

Graphical abstract

21 pages, 16988 KiB  
Article
An End-to-End Adaptive Method for Remaining Useful Life Prediction of Rolling Bearings Using Time–Frequency Image Features
by Liang Chen, Hao Wang, Linshu Meng, Zhenzhen Xu, Lin Xue and Mingfa Ren
Mach. Learn. Knowl. Extr. 2024, 6(4), 2892-2912; https://doi.org/10.3390/make6040138 - 16 Dec 2024
Viewed by 715
Abstract
The deep learning model has attracted widespread attention in the field of rolling bearing remaining useful life (RUL) prediction due to its advantages of less reliance on prior knowledge, high accuracy, and strong generalization. However, a large number of prediction models use very [...] Read more.
The deep learning model has attracted widespread attention in the field of rolling bearing remaining useful life (RUL) prediction due to its advantages of less reliance on prior knowledge, high accuracy, and strong generalization. However, a large number of prediction models use very complicated artificial feature extraction and selection methods to build the original input features of the deep learning model and health indicator. These approaches do not fully exploit the capabilities of deep learning models as they continue to heavily rely on prior knowledge, The accuracy of their predictions largely hinges on the quality of the input features, and the generalization of manually crafted features remains uncertain. To address these challenges, in this paper, an end-to-end prediction model for the remaining useful life of rolling bearings is proposed, which is divided into three modules. First, a short-term Fourier transform module is incorporated into the model to automatically obtain the time–frequency information of the signal. Then, the convolutional next (ConvNext) module, which is a simple and efficient pure convolutional neural network, is utilized to extract features from the spectrogram. Finally, we capture the short-term dependence and long-term dependence by two parallel channels Transformer and self-attention convolutional long short-term memory (SA-ConvLSTM), and the self-attention mechanism is employed for the adaptive prediction of the bearing’s remaining useful life. Through integration with artificial intelligence, this method proposes a high-performance solution for predicting the remaining useful life of bearings. It has minimal reliance on manual labor, stronger fitting capabilities, and can be widely used for predicting the remaining useful life of bearings. Full article
(This article belongs to the Topic Applications of NLP, AI, and ML in Software Engineering)
Show Figures

Figure 1

16 pages, 538 KiB  
Article
What ChatGPT Has to Say About Its Topological Structure: The Anyon Hypothesis
by Michel Planat and Marcelo Amaral
Mach. Learn. Knowl. Extr. 2024, 6(4), 2876-2891; https://doi.org/10.3390/make6040137 - 15 Dec 2024
Viewed by 962
Abstract
Large language models (LLMs) achieve remarkable predictive capabilities but remain opaque in their internal reasoning, creating a pressing need for more interpretable artificial intelligence. Here, we propose bridging this explanatory gap by drawing on concepts from topological quantum computing (TQC), specifically the anyonic [...] Read more.
Large language models (LLMs) achieve remarkable predictive capabilities but remain opaque in their internal reasoning, creating a pressing need for more interpretable artificial intelligence. Here, we propose bridging this explanatory gap by drawing on concepts from topological quantum computing (TQC), specifically the anyonic frameworks arising from SU(2)k theories. Anyons interpolate between fermions and bosons, offering a mathematical language that may illuminate the latent structure and decision-making processes within LLMs. By examining how these topological constructs relate to token interactions and contextual dependencies in neural architectures, we aim to provide a fresh perspective on how meaning and coherence emerge. After eliciting insights from ChatGPT and exploring low-level cases of SU(2)k models, we argue that the machinery of modular tensor categories and topological phases could inform more transparent, stable, and robust AI systems. This interdisciplinary approach suggests that quantum-theoretic principles may underpin a novel understanding of explainable AI. Full article
21 pages, 2964 KiB  
Article
Prediction of Drivers’ Red-Light Running Behaviour in Connected Vehicle Environments Using Deep Recurrent Neural Networks
by Md Mostafizur Rahman Komol, Mohammed Elhenawy, Jack Pinnow, Mahmoud Masoud, Andry Rakotonirainy, Sebastien Glaser, Merle Wood and David Alderson
Mach. Learn. Knowl. Extr. 2024, 6(4), 2855-2875; https://doi.org/10.3390/make6040136 - 11 Dec 2024
Viewed by 845
Abstract
Red-light running at signalised intersections poses a significant safety risk, necessitating advanced predictive technologies to predict red-light violation behaviour, especially for advanced red-light warning (ARLW) systems. This research leverages Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) models to forecast the red-light [...] Read more.
Red-light running at signalised intersections poses a significant safety risk, necessitating advanced predictive technologies to predict red-light violation behaviour, especially for advanced red-light warning (ARLW) systems. This research leverages Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) models to forecast the red-light running and stopping behaviours of drivers in connected vehicles. We utilised data from the Ipswich Connected Vehicle Pilot (ICVP) in Queensland, Australia, which gathered naturalistic driving data from 355 connected vehicles at 29 signalised intersections. These vehicles broadcast Cooperative Awareness Messages (CAM) within the Cooperative Intelligent Transport Systems (C-ITS), providing kinematic inputs such as vehicle speed, speed limits, longitudinal and lateral accelerations, and yaw rate. These variables were monitored at 100-millisecond intervals for durations from 1 to 4 s before reaching various distances from the stop line. Our results indicate that the LSTM model outperforms the GRU in predicting both red-light running and stopping behaviours with high accuracy. However, the pre-trained GRU model performs better in predicting red-light running specifically, making it valuable in applications requiring early violation prediction. Implementing these models can enhance red-light violation countermeasures, such as dynamic all-red extension (DARE), decreasing the likelihood of severe collisions and enhancing road users’ safety. Full article
Show Figures

Figure 1

26 pages, 2059 KiB  
Article
Continual Semi-Supervised Malware Detection
by Matthew Chin and Roberto Corizzo
Mach. Learn. Knowl. Extr. 2024, 6(4), 2829-2854; https://doi.org/10.3390/make6040135 - 10 Dec 2024
Viewed by 920
Abstract
Detecting malware has become extremely important with the increasing exposure of computational systems and mobile devices to online services. However, the rapidly evolving nature of malicious software makes this task particularly challenging. Despite the significant number of machine learning works for malware detection [...] Read more.
Detecting malware has become extremely important with the increasing exposure of computational systems and mobile devices to online services. However, the rapidly evolving nature of malicious software makes this task particularly challenging. Despite the significant number of machine learning works for malware detection proposed in the last few years, limited interest has been devoted to continual learning approaches, which could allow models to showcase effective performance in challenging and dynamic scenarios while being computationally efficient. Moreover, most of the research works proposed thus far adopt a fully supervised setting, which relies on fully labelled data and appears to be impractical in a rapidly evolving malware landscape. In this paper, we address malware detection from a continual semi-supervised one-class learning perspective, which only requires normal/benign data and empowers models with a greater degree of flexibility, allowing them to detect multiple malware types with different morphology. Specifically, we assess the effectiveness of two replay strategies on anomaly detection models and analyze their performance in continual learning scenarios with three popular malware detection datasets (CIC-AndMal2017, CIC-MalMem-2022, and CIC-Evasive-PDFMal2022). Our evaluation shows that replay-based strategies can achieve competitive performance in terms of continual ROC-AUC with respect to the considered baselines and bring new perspectives and insights on this topic. Full article
Show Figures

Figure 1

21 pages, 6514 KiB  
Systematic Review
Digital Assistance Systems to Implement Machine Learning in Manufacturing: A Systematic Review
by Jannik Rosemeyer, Marta Pinzone and Joachim Metternich
Mach. Learn. Knowl. Extr. 2024, 6(4), 2808-2828; https://doi.org/10.3390/make6040134 - 4 Dec 2024
Viewed by 1082
Abstract
Implementing machine learning technologies in manufacturing environment relies heavily on human expertise in terms of domain and machine learning knowledge. Yet, the required machine learning knowledge is often not available in manufacturing companies. A possible solution to overcome this competence gap and let [...] Read more.
Implementing machine learning technologies in manufacturing environment relies heavily on human expertise in terms of domain and machine learning knowledge. Yet, the required machine learning knowledge is often not available in manufacturing companies. A possible solution to overcome this competence gap and let domain experts with limited machine learning programming skills build viable applications are digital assistance systems that support the implementation. At the present, there is no comprehensive overview over corresponding assistance systems. Thus, within this study a systematic literature review based on the PRISMA-P process was conducted. Twenty-nine papers were identified and analyzed in depth regarding machine learning use case, required resources and research outlook. Six key findings as well as requirements for future developments are derived from the investigation. As such, the existing assistance systems basically focus on technical aspects whereas the integration of the users as well as validation in industrial environments lack behind. Future assistance systems should put more emphasis on the users and integrate them both in development and validation. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

25 pages, 7107 KiB  
Article
Parameter-Efficient Fine-Tuning of Large Pretrained Models for Instance Segmentation Tasks
by Nermeen Abou Baker, David Rohrschneider and Uwe Handmann
Mach. Learn. Knowl. Extr. 2024, 6(4), 2783-2807; https://doi.org/10.3390/make6040133 - 2 Dec 2024
Viewed by 1382
Abstract
Research and applications in artificial intelligence have recently shifted with the rise of large pretrained models, which deliver state-of-the-art results across numerous tasks. However, the substantial increase in parameters introduces a need for parameter-efficient training strategies. Despite significant advancements, limited research has explored [...] Read more.
Research and applications in artificial intelligence have recently shifted with the rise of large pretrained models, which deliver state-of-the-art results across numerous tasks. However, the substantial increase in parameters introduces a need for parameter-efficient training strategies. Despite significant advancements, limited research has explored parameter-efficient fine-tuning (PEFT) methods in the context of transformer-based models for instance segmentation. Addressing this gap, this study investigates the effectiveness of PEFT methods, specifically adapters and Low-Rank Adaptation (LoRA), applied to two models across four benchmark datasets. Integrating sequentially arranged adapter modules and applying LoRA to deformable attention—explored here for the first time—achieves competitive performance while fine-tuning only about 1–6% of model parameters, a marked improvement over the 40–55% required in traditional fine-tuning. Key findings indicate that using 2–3 adapters per transformer block offers an optimal balance of performance and efficiency. Furthermore, LoRA, exhibits strong parameter efficiency when applied to deformable attention, and in certain cases surpasses adapter configurations. These results show that the impact of PEFT techniques varies based on dataset complexity and model architecture, underscoring the importance of context-specific tuning. Overall, this work demonstrates the potential of PEFT to enable scalable, customizable, and computationally efficient transfer learning for instance segmentation tasks. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

30 pages, 1413 KiB  
Tutorial
Deep Learning with Convolutional Neural Networks: A Compact Holistic Tutorial with Focus on Supervised Regression
by Yansel Gonzalez Tejeda and Helmut A. Mayer
Mach. Learn. Knowl. Extr. 2024, 6(4), 2753-2782; https://doi.org/10.3390/make6040132 - 30 Nov 2024
Viewed by 1052
Abstract
In this tutorial, we present a compact and holistic discussion of Deep Learning with a focus on Convolutional Neural Networks (CNNs) and supervised regression. While there are numerous books and articles on the individual topics we cover, comprehensive and detailed tutorials that address [...] Read more.
In this tutorial, we present a compact and holistic discussion of Deep Learning with a focus on Convolutional Neural Networks (CNNs) and supervised regression. While there are numerous books and articles on the individual topics we cover, comprehensive and detailed tutorials that address deep learning from a foundational yet rigorous and accessible perspective are rare. Most resources on CNNs are either too advanced, focusing on cutting-edge architectures, or too narrow, addressing only specific applications like image classification. This tutorial not only summarizes the most relevant concepts but also provides an in-depth exploration of each, offering a complete yet agile set of ideas. Moreover, we highlight the powerful synergy between learning theory, statistics, and machine learning, which together underpin the deep learning and CNN frameworks. We aim for this tutorial to serve as an optimal resource for students, professors, and anyone interested in understanding the foundations of deep learning. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

15 pages, 3785 KiB  
Article
Optimizing Ingredient Substitution Using Large Language Models to Enhance Phytochemical Content in Recipes
by Luís Rita, Joshua Southern, Ivan Laponogov, Kyle Higgins and Kirill Veselkov
Mach. Learn. Knowl. Extr. 2024, 6(4), 2738-2752; https://doi.org/10.3390/make6040131 - 26 Nov 2024
Viewed by 1429
Abstract
In the emerging field of computational gastronomy, aligning culinary practices with scientifically supported nutritional goals is increasingly important. This study explores how large language models (LLMs) can be applied to optimize ingredient substitutions in recipes, specifically to enhance the phytochemical content of meals. [...] Read more.
In the emerging field of computational gastronomy, aligning culinary practices with scientifically supported nutritional goals is increasingly important. This study explores how large language models (LLMs) can be applied to optimize ingredient substitutions in recipes, specifically to enhance the phytochemical content of meals. Phytochemicals are bioactive compounds found in plants, which, based on preclinical studies, may offer potential health benefits. We fine-tuned models, including OpenAI’s GPT-3.5-Turbo, DaVinci-002, and Meta’s TinyLlama-1.1B, using an ingredient substitution dataset. These models were used to predict substitutions that enhance the phytochemical content and to create a corresponding enriched recipe dataset. Our approach improved the top ingredient prediction accuracy on substitution tasks, from the baseline 34.53 ± 0.10% to 38.03 ± 0.28% on the original substitution dataset and from 40.24 ± 0.36% to 54.46 ± 0.29% on a refined version of the same dataset. These substitutions led to the creation of 1951 phytochemically enriched ingredient pairings and 1639 unique recipes. While this approach demonstrates potential in optimizing ingredient substitutions, caution must be taken when drawing conclusions about health benefits, as the claims are based on preclinical evidence. This research represents a step forward in using AI to promote healthier eating practices, providing potential pathways for integrating computational methods with nutritional science. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

16 pages, 789 KiB  
Article
Node-Centric Pruning: A Novel Graph Reduction Approach
by Hossein Shokouhinejad, Roozbeh Razavi-Far, Griffin Higgins and Ali A. Ghorbani
Mach. Learn. Knowl. Extr. 2024, 6(4), 2722-2737; https://doi.org/10.3390/make6040130 - 22 Nov 2024
Viewed by 1221
Abstract
In the era of rapidly expanding graph-based applications, efficiently managing large-scale graphs has become a critical challenge. This paper introduces an innovative graph reduction technique, Node-Centric Pruning (NCP), designed to simplify complex graphs while preserving their essential structural properties, thereby enhancing the scalability [...] Read more.
In the era of rapidly expanding graph-based applications, efficiently managing large-scale graphs has become a critical challenge. This paper introduces an innovative graph reduction technique, Node-Centric Pruning (NCP), designed to simplify complex graphs while preserving their essential structural properties, thereby enhancing the scalability and maintaining performance of downstream Graph Neural Networks (GNNs). Our proposed approach strategically prunes less significant nodes and refines the graph structure, ensuring that critical topological properties are maintained. By carefully evaluating node significance based on advanced connectivity metrics, our method preserves the topology and ensures high performance in downstream machine learning tasks. Extensive experimentation demonstrates that our proposed method not only maintains the integrity and functionality of the original graph but also significantly improves the computational efficiency and preserves the classification performance of GNNs. These enhancements in computational efficiency and resource management make our technique particularly valuable for deploying GNNs in real-world applications, where handling large, complex datasets effectively is crucial. This advancement represents a significant step toward making GNNs more practical and effective for a wide range of applications in both industry and academia. Full article
Show Figures

Figure 1

34 pages, 1657 KiB  
Article
A Study on Text Classification in the Age of Large Language Models
by Paul Trust and Rosane Minghim
Mach. Learn. Knowl. Extr. 2024, 6(4), 2688-2721; https://doi.org/10.3390/make6040129 - 21 Nov 2024
Viewed by 1330
Abstract
Large language models (LLMs) have recently made significant advances, excelling in tasks like question answering, summarization, and machine translation. However, their enormous size and hardware requirements make them less accessible to many in the machine learning community. To address this, techniques such as [...] Read more.
Large language models (LLMs) have recently made significant advances, excelling in tasks like question answering, summarization, and machine translation. However, their enormous size and hardware requirements make them less accessible to many in the machine learning community. To address this, techniques such as quantization, prefix tuning, weak supervision, low-rank adaptation, and prompting have been developed to customize these models for specific applications. While these methods have mainly improved text generation, their implications for the text classification task are not thoroughly studied. Our research intends to bridge this gap by investigating how variations like model size, pre-training objectives, quantization, low-rank adaptation, prompting, and various hyperparameters influence text classification tasks. Our overall conclusions show the following: 1—even with synthetic labels, fine-tuning works better than prompting techniques, and increasing model size does not always improve classification performance; 2—discriminatively trained models generally perform better than generatively pre-trained models; and 3—fine-tuning models at 16-bit precision works much better than using 8-bit or 4-bit models, but the performance drop from 8-bit to 4-bit is smaller than from 16-bit to 8-bit. In another scale of our study, we conducted experiments with different settings for low-rank adaptation (LoRA) and quantization, finding that increasing LoRA dropout negatively affects classification performance. We did not find a clear link between the LoRA attention dimension (rank) and performance, observing only small differences between standard LoRA and its variants like rank-stabilized LoRA and weight-decomposed LoRA. Additional observations to support model setup for classification tasks are presented in our analyses. Full article
Show Figures

Figure 1

29 pages, 2409 KiB  
Article
Enhancing Hierarchical Sales Forecasting with Promotional Data: A Comparative Study Using ARIMA and Deep Neural Networks
by Mariana Teixeira, José Manuel Oliveira and Patrícia Ramos
Mach. Learn. Knowl. Extr. 2024, 6(4), 2659-2687; https://doi.org/10.3390/make6040128 - 19 Nov 2024
Viewed by 922
Abstract
Retailers depend on accurate sales forecasts to effectively plan operations and manage supply chains. These forecasts are needed across various levels of aggregation, making hierarchical forecasting methods essential for the retail industry. As competition intensifies, the use of promotions has become a widespread [...] Read more.
Retailers depend on accurate sales forecasts to effectively plan operations and manage supply chains. These forecasts are needed across various levels of aggregation, making hierarchical forecasting methods essential for the retail industry. As competition intensifies, the use of promotions has become a widespread strategy, significantly impacting consumer purchasing behavior. This study seeks to improve forecast accuracy by incorporating promotional data into hierarchical forecasting models. Using a sales dataset from a major Portuguese retailer, base forecasts are generated for different hierarchical levels using ARIMA models and Multi-Layer Perceptron (MLP) neural networks. Reconciliation methods including bottom-up, top-down, and optimal reconciliation with OLS and WLS (struct) estimators are employed. The results show that MLPs outperform ARIMA models for forecast horizons longer than one day. While the addition of regressors enhances ARIMA’s accuracy, it does not yield similar improvements for MLP. MLPs present a compelling balance of simplicity and efficiency, outperforming ARIMA in flexibility while offering faster training times and lower computational demands compared to more complex deep learning models, making them highly suitable for practical retail forecasting applications. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

20 pages, 1278 KiB  
Article
Application of Bayesian Neural Networks in Healthcare: Three Case Studies
by Lebede Ngartera, Mahamat Ali Issaka and Saralees Nadarajah
Mach. Learn. Knowl. Extr. 2024, 6(4), 2639-2658; https://doi.org/10.3390/make6040127 - 16 Nov 2024
Viewed by 1660
Abstract
This study aims to explore the efficacy of Bayesian Neural Networks (BNNs) in enhancing predictive modeling for healthcare applications. Advancements in artificial intelligence have significantly improved predictive modeling capabilities, with BNNs offering a probabilistic framework that addresses the inherent uncertainty and variability in [...] Read more.
This study aims to explore the efficacy of Bayesian Neural Networks (BNNs) in enhancing predictive modeling for healthcare applications. Advancements in artificial intelligence have significantly improved predictive modeling capabilities, with BNNs offering a probabilistic framework that addresses the inherent uncertainty and variability in healthcare data. This study demonstrates the real-world applicability of BNNs through three key case studies: personalized diabetes treatment, early Alzheimer’s disease detection, and predictive modeling for HbA1c levels. By leveraging the Bayesian approach, these models provide not only enhanced predictive accuracy but also uncertainty quantification, a critical factor in clinical decision making. While the findings are promising, future research should focus on optimizing scalability and integration for real-world applications. This work lays a foundation for future studies, including the development of rating scales based on BNN predictions to improve clinical outcomes. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

21 pages, 603 KiB  
Article
Diversifying Multi-Head Attention in the Transformer Model
by Nicholas Ampazis and Flora Sakketou
Mach. Learn. Knowl. Extr. 2024, 6(4), 2618-2638; https://doi.org/10.3390/make6040126 - 12 Nov 2024
Viewed by 1485
Abstract
Recent studies have shown that, due to redundancy, some heads of the Transformer model can be pruned without diminishing the efficiency of the model. In this paper, we propose a constrained optimization algorithm based on Hebbian learning, which trains specific layers in the [...] Read more.
Recent studies have shown that, due to redundancy, some heads of the Transformer model can be pruned without diminishing the efficiency of the model. In this paper, we propose a constrained optimization algorithm based on Hebbian learning, which trains specific layers in the Transformer architecture in order to enforce diversification between the different heads in the multi-head attention module. The diversification of the heads is achieved through a single-layer feed-forward neural network that is added to the Transformer architecture and is trained with the proposed algorithm. We utilize the algorithm in three different architectural variations of the baseline Transformer model. In addition to the diversification of the heads, the proposed methodology can be used to prune the heads that capture redundant information. Experiments on diverse NLP tasks, including machine translation, text summarization, question answering and large language modeling, show that our proposed approach consistently improves the performance of baseline Transformer models. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

17 pages, 765 KiB  
Article
Data Reconciliation-Based Hierarchical Fusion of Machine Learning Models
by Pál Péter Hanzelik, Alex Kummer and János Abonyi
Mach. Learn. Knowl. Extr. 2024, 6(4), 2601-2617; https://doi.org/10.3390/make6040125 - 11 Nov 2024
Viewed by 798
Abstract
In the context of hierarchical system modeling, ensuring constraints between different hierarchy levels are met, so, for instance, ensuring the aggregation constraints are satisfied, is essential. However, modelling and forecasting each element of the hierarchy independently introduce errors. To mitigate this balance error, [...] Read more.
In the context of hierarchical system modeling, ensuring constraints between different hierarchy levels are met, so, for instance, ensuring the aggregation constraints are satisfied, is essential. However, modelling and forecasting each element of the hierarchy independently introduce errors. To mitigate this balance error, it is recommended to employ an optimal data reconciliation technique with an emphasis on measurement and modeling errors. In this study, three different machine learning methods for development were investigated. The first method involves no data reconciliation, relying solely on machine learning models built independently at each hierarchical level. The second approach incorporates measurement errors by adjusting the measured data to satisfy each constraint, and the machine learning model is developed based on this dataset. The third method is based on directly fine-tuning the machine learning predictions based on the prediction errors of each model. The three methods were compared using three case studies with different complexities, namely mineral composition estimation with 9 elements, forecasting of retail sales with 14 elements, and waste deposition forecasting with more than 3000 elements. From the results of this study, the conclusion can be drawn that the third method performs the best, and reliable machine learning models can be developed. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

31 pages, 2280 KiB  
Perspective
Adaptive AI Alignment: Established Resources for Aligning Machine Learning with Human Intentions and Values in Changing Environments
by Stephen Fox
Mach. Learn. Knowl. Extr. 2024, 6(4), 2570-2600; https://doi.org/10.3390/make6040124 - 6 Nov 2024
Viewed by 1803
Abstract
AI Alignment is a term used to summarize the aim of making artificial intelligence (AI) systems behave in line with human intentions and values. There has been little consideration in previous AI Alignment studies of the need for AI Alignment to be adaptive [...] Read more.
AI Alignment is a term used to summarize the aim of making artificial intelligence (AI) systems behave in line with human intentions and values. There has been little consideration in previous AI Alignment studies of the need for AI Alignment to be adaptive in order to contribute to the survival of human organizations in changing environments. This research gap is addressed here by defining human intentions and values in terms of survival biophysics: entropy, complexity, and adaptive behavior. Furthermore, although technology alignment has been a focus of studies for more than thirty years, there has been little consideration in AI Alignment studies of established resources for aligning technologies. Unlike the current focus of AI Alignment on addressing potential AI risks, technology alignment is generally focused on aligning with opportunities. Established resources include the critical realist philosophy of science, scientific theories, total quality management practices, technology alignment methods, engineering techniques, and technology standards. Here, these established resources are related to the alignment of different types of machine learning with different levels of human organizations. In addition, established resources are related to a well-known hypothetical extreme example of AI Misalignment, and to major constructs in the AI Alignment literature. Overall, it is argued that AI Alignment needs to be adaptive in order for human organizations to be able to survive in changing environments, and that established resources can facilitate Adaptive AI Alignment which addresses risks while focusing on opportunities. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

55 pages, 2140 KiB  
Review
A Review on Machine Learning Deployment Patterns and Key Features in the Prediction of Preeclampsia
by Louise Pedersen, Magdalena Mazur-Milecka, Jacek Ruminski and Stefan Wagner
Mach. Learn. Knowl. Extr. 2024, 6(4), 2515-2569; https://doi.org/10.3390/make6040123 - 5 Nov 2024
Viewed by 968
Abstract
Previous reviews have investigated machine learning (ML) models used to predict the risk of developing preeclampsia. However, they have not addressed the intended deployment of these models throughout pregnancy, nor have they detailed feature performance. This study aims to provide an overview of [...] Read more.
Previous reviews have investigated machine learning (ML) models used to predict the risk of developing preeclampsia. However, they have not addressed the intended deployment of these models throughout pregnancy, nor have they detailed feature performance. This study aims to provide an overview of existing ML models and their intended deployment patterns and performance, along with identified features of high importance. This review followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 guidelines. The search was performed in January and February 2024. It included all papers published before March 2024 obtained from the scientific databases: PubMed, Engineering Village, the Association for Computing Machinery, Scopus, and Web of Science. Of a total of 198 identified studies, 18 met the inclusion criteria. Among these, 11 showed the intent to use the ML model as a single-use tool, two intended a dual-use, and two intended multiple-use. Ten studies listed the features of the highest importance, with systolic and diastolic blood pressure, mean arterial pressure, and hypertension frequently mentioned as critical predictors. Notably, three of the four studies proposing dual or multiple-use models were conducted in 2023 and 2024, while the remaining study is from 2009. No single ML model emerged as superior across the subgroups of PE. Incorporating body mass index alongside hypertension and either mean arterial pressure, diastolic blood pressure, or systolic blood pressure as features may enhance performance. The deployment patterns mainly focused on single use during gestational weeks 11+0 to 14+1. Full article
(This article belongs to the Special Issue Machine Learning in Data Science)
Show Figures

Figure 1

21 pages, 1223 KiB  
Article
Advancing Life Cycle Assessment of Sustainable Green Hydrogen Production Using Domain-Specific Fine-Tuning by Large Language Models Augmentation
by Yajing Chen, Urs Liebau, Shreyas Mysore Guruprasad, Iaroslav Trofimenko and Christine Minke
Mach. Learn. Knowl. Extr. 2024, 6(4), 2494-2514; https://doi.org/10.3390/make6040122 - 4 Nov 2024
Viewed by 1452
Abstract
Assessing the sustainable development of green hydrogen and assessing its potential environmental impacts using the Life Cycle Assessment is crucial. Challenges in LCA, like missing environmental data, are often addressed using machine learning, such as artificial neural networks. However, to find an ML [...] Read more.
Assessing the sustainable development of green hydrogen and assessing its potential environmental impacts using the Life Cycle Assessment is crucial. Challenges in LCA, like missing environmental data, are often addressed using machine learning, such as artificial neural networks. However, to find an ML solution, researchers need to read extensive literature or consult experts. This research demonstrates how customised LLMs, trained with domain-specific papers, can help researchers overcome these challenges. By starting small by consolidating papers focused on the LCA of proton exchange membrane water electrolysis, which produces green hydrogen, and ML applications in LCA. These papers are uploaded to OpenAI to create the LlamaIndex, enabling future queries. Using the LangChain framework, researchers query the customised model (GPT-3.5-turbo), receiving tailored responses. The results demonstrate that customised LLMs can assist researchers in providing suitable ML solutions to address data inaccuracies and gaps. The ability to quickly query an LLM and receive an integrated response across relevant sources presents an improvement over manually retrieving and reading individual papers. This shows that leveraging fine-tuned LLMs can empower researchers to conduct LCAs more efficiently and effectively. Full article
(This article belongs to the Special Issue Sustainable Applications for Machine Learning)
Show Figures

Figure 1

47 pages, 1749 KiB  
Article
Automatic Extraction and Visualization of Interaction Networks for German Fairy Tales
by David Schmidt and Frank Puppe
Mach. Learn. Knowl. Extr. 2024, 6(4), 2447-2493; https://doi.org/10.3390/make6040121 - 1 Nov 2024
Viewed by 1036
Abstract
Interaction networks are a method of displaying the significant characters in a narrative text and their interactions. We automatically construct interaction networks from dialogues in German fairy tales by the Brothers Grimm and subsequently visualize these networks. This requires the combination of algorithms [...] Read more.
Interaction networks are a method of displaying the significant characters in a narrative text and their interactions. We automatically construct interaction networks from dialogues in German fairy tales by the Brothers Grimm and subsequently visualize these networks. This requires the combination of algorithms for several tasks: coreference resolution for the identification of characters and their appearances, as well as speaker/addressee detection and the detection of dialogue boundaries for the identification of interactions. After an evaluation of the individual algorithms, the predicted networks are evaluated against benchmarks established by networks based on manually annotated coreference and speaker/addressee information. The evaluation focuses on specific components of the predicted networks, such as the nodes, as well as the overall network, employing a newly devised score. This is followed by an analysis of various types of errors that the algorithms can make, like a coreference resolution algorithm not realizing that the frog has transformed into a prince, and their impact on the created networks. We find that the quality of many predicted networks is satisfactory for use cases in which the reliability of edges and character types are not of critical importance. However, there is considerable room for improvement. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

12 pages, 1295 KiB  
Article
Lexical Error Guard: Leveraging Large Language Models for Enhanced ASR Error Correction
by Mei Si, Omar Cobas and Michael Fababeir
Mach. Learn. Knowl. Extr. 2024, 6(4), 2435-2446; https://doi.org/10.3390/make6040120 - 29 Oct 2024
Viewed by 2636
Abstract
Error correction is a vital element in modern automatic speech recognition (ASR) systems. A significant portion of ASR error correction work is closely integrated within specific ASR systems, which creates challenges for adapting these solutions to different ASR frameworks. This research introduces Lexical [...] Read more.
Error correction is a vital element in modern automatic speech recognition (ASR) systems. A significant portion of ASR error correction work is closely integrated within specific ASR systems, which creates challenges for adapting these solutions to different ASR frameworks. This research introduces Lexical Error Guard (LEG), which leverages the extensive pre-trained knowledge of large language models (LLMs) and employs instructional learning to create an adaptable error correction system compatible with various ASR platforms. Additionally, a parameter-efficient fine-tuning method is utilized using quantized low-rank adaptation (QLoRA) to facilitate fast training of the system. Tested on the LibriSpeech data corpus, the results indicate that LEG improves ASR results when used with various Whisper model sizes. Improvements in WER are made, with a decrease from 2.27% to 2.21% on the “Test Clean” dataset for Whisper Large with beam search. Improvements on the “Test Other” dataset for Whisper Large with beam search are also made, from 4.93% to 4.72%. Full article
Show Figures

Figure 1

13 pages, 1761 KiB  
Article
Leveraging Multi-Modality and Enhanced Temporal Networks for Robust Violence Detection
by Gwangho Na, Jaepil Ko and Kyungjoo Cheoi
Mach. Learn. Knowl. Extr. 2024, 6(4), 2422-2434; https://doi.org/10.3390/make6040119 - 28 Oct 2024
Viewed by 945
Abstract
In this paper, we present a novel model that enhances performance by extending the dual-modality TEVAD model—originally leveraging visual and textual information—into a multi-modal framework that integrates visual, audio, and textual data. Additionally, we refine the multi-scale temporal network (MTN) to improve feature [...] Read more.
In this paper, we present a novel model that enhances performance by extending the dual-modality TEVAD model—originally leveraging visual and textual information—into a multi-modal framework that integrates visual, audio, and textual data. Additionally, we refine the multi-scale temporal network (MTN) to improve feature extraction across multiple temporal scales between video snippets. Using the XD-Violence dataset, which includes audio data for violence detection, we conduct experiments to evaluate various feature fusion methods. The proposed model achieves an average precision (AP) of 83.9%, surpassing the performance of single-modality approaches (visual: 73.9%, audio: 67.1%, textual: 29.9%) and dual-modality approaches (visual + audio: 78.8%, visual + textual: 78.5%). These findings demonstrate that the proposed model outperforms models based on the original MTN and reaffirm the efficacy of multi-modal approaches in enhancing violence detection compared to single- or dual-modality methods. Full article
Show Figures

Figure 1

22 pages, 15428 KiB  
Article
Towards Self-Conscious AI Using Deep ImageNet Models: Application for Blood Cell Classification
by Mohamad Abou Ali, Fadi Dornaika and Ignacio Arganda-Carreras
Mach. Learn. Knowl. Extr. 2024, 6(4), 2400-2421; https://doi.org/10.3390/make6040118 - 21 Oct 2024
Viewed by 2254
Abstract
The exceptional performance of ImageNet competition winners in image classification has led AI researchers to repurpose these models for a whole range of tasks using transfer learning (TL). TL has been hailed for boosting performance, shortening learning time and reducing computational effort. Despite [...] Read more.
The exceptional performance of ImageNet competition winners in image classification has led AI researchers to repurpose these models for a whole range of tasks using transfer learning (TL). TL has been hailed for boosting performance, shortening learning time and reducing computational effort. Despite these benefits, issues such as data sparsity and the misrepresentation of classes can diminish these gains, occasionally leading to misleading TL accuracy scores. This research explores the innovative concept of endowing ImageNet models with a self-awareness that enables them to recognize their own accumulated knowledge and experience. Such self-awareness is expected to improve their adaptability in various domains. We conduct a case study using two different datasets, PBC and BCCD, which focus on blood cell classification. The PBC dataset provides high-resolution images with abundant data, while the BCCD dataset is hindered by limited data and inferior image quality. To compensate for these discrepancies, we use data augmentation for BCCD and undersampling for both datasets to achieve balance. Subsequent pre-processing generates datasets of different size and quality, all geared towards blood cell classification. We extend conventional evaluation tools with novel metrics—“accuracy difference” and “loss difference”—to detect overfitting or underfitting and evaluate their utility as potential indicators for learning behavior and promoting the self-confidence of ImageNet models. Our results show that these metrics effectively track learning progress and improve the reliability and overall performance of ImageNet models in new applications. This study highlights the transformative potential of turning ImageNet models into self-aware entities that significantly improve their robustness and efficiency in various AI tasks. This groundbreaking approach opens new perspectives for increasing the effectiveness of transfer learning in real-world AI implementations. Full article
Show Figures

Figure 1

25 pages, 4300 KiB  
Article
Machine Learning Monte Carlo Approaches and Statistical Physics Notions to Characterize Bacterial Species in Human Microbiota
by Michele Bellingeri, Leonardo Mancabelli, Christian Milani, Gabriele Andrea Lugli, Roberto Alfieri, Massimiliano Turchetto, Marco Ventura and Davide Cassi
Mach. Learn. Knowl. Extr. 2024, 6(4), 2375-2399; https://doi.org/10.3390/make6040117 - 18 Oct 2024
Viewed by 835
Abstract
Recent studies have shown correlations between the microbiota’s composition and various health conditions. Machine learning (ML) techniques are essential for analyzing complex biological data, particularly in microbiome research. ML methods help analyze large datasets to uncover microbiota patterns and understand how these patterns [...] Read more.
Recent studies have shown correlations between the microbiota’s composition and various health conditions. Machine learning (ML) techniques are essential for analyzing complex biological data, particularly in microbiome research. ML methods help analyze large datasets to uncover microbiota patterns and understand how these patterns affect human health. This study introduces a novel approach combining statistical physics with the Monte Carlo (MC) methods to characterize bacterial species in the human microbiota. We assess the significance of bacterial species in different age groups by using notions of statistical distances to evaluate species prevalence and abundance across age groups and employing MC simulations based on statistical mechanics principles. Our findings show that the microbiota composition experiences a significant transition from early childhood to adulthood. Species such as Bifidobacterium breve and Veillonella parvula decrease with age, while others like Agathobaculum butyriciproducens and Eubacterium rectale increase. Additionally, low-prevalence species may hold significant importance in characterizing age groups. Finally, we propose an overall species ranking by integrating the methods proposed here in a multicriteria classification strategy. Our research provides a comprehensive tool for microbiota analysis using statistical notions, ML techniques, and MC simulations. Full article
(This article belongs to the Special Issue Advances in Machine and Deep Learning)
Show Figures

Figure 1

20 pages, 2409 KiB  
Article
Systematic Analysis of Retrieval-Augmented Generation-Based LLMs for Medical Chatbot Applications
by Arunabh Bora and Heriberto Cuayáhuitl
Mach. Learn. Knowl. Extr. 2024, 6(4), 2355-2374; https://doi.org/10.3390/make6040116 - 18 Oct 2024
Viewed by 4852
Abstract
Artificial Intelligence (AI) has the potential to revolutionise the medical and healthcare sectors. AI and related technologies could significantly address some supply-and-demand challenges in the healthcare system, such as medical AI assistants, chatbots and robots. This paper focuses on tailoring LLMs to medical [...] Read more.
Artificial Intelligence (AI) has the potential to revolutionise the medical and healthcare sectors. AI and related technologies could significantly address some supply-and-demand challenges in the healthcare system, such as medical AI assistants, chatbots and robots. This paper focuses on tailoring LLMs to medical data utilising a Retrieval-Augmented Generation (RAG) database to evaluate their performance in a computationally resource-constrained environment. Existing studies primarily focus on fine-tuning LLMs on medical data, but this paper combines RAG and fine-tuned models and compares them against base models using RAG or only fine-tuning. Open-source LLMs (Flan-T5-Large, LLaMA-2-7B, and Mistral-7B) are fine-tuned using the medical datasets Meadow-MedQA and MedMCQA. Experiments are reported for response generation and multiple-choice question answering. The latter uses two distinct methodologies: Type A, as standard question answering via direct choice selection; and Type B, as language generation and probability confidence score generation of choices available. Results in the medical domain revealed that Fine-tuning and RAG are crucial for improved performance, and that methodology Type A outperforms Type B. Full article
Show Figures

Figure 1

19 pages, 4843 KiB  
Article
Long-Range Bird Species Identification Using Directional Microphones and CNNs
by Tiago Garcia, Luís Pina, Magnus Robb, Jorge Maria, Roel May and Ricardo Oliveira
Mach. Learn. Knowl. Extr. 2024, 6(4), 2336-2354; https://doi.org/10.3390/make6040115 - 16 Oct 2024
Viewed by 1164
Abstract
This study explores the integration of directional microphones with convolutional neural networks (CNNs) for long-range bird species identification. By employing directional microphones, we aimed to capture high-resolution audio from specific directions, potentially improving the clarity of bird calls over extended distances. Our approach [...] Read more.
This study explores the integration of directional microphones with convolutional neural networks (CNNs) for long-range bird species identification. By employing directional microphones, we aimed to capture high-resolution audio from specific directions, potentially improving the clarity of bird calls over extended distances. Our approach involved processing these recordings with CNNs trained on a diverse dataset of bird calls. The results demonstrated that the system is capable of systematically identifying bird species up to 150 m, reaching 280 m for species vocalizing at frequencies greater than 1000 Hz and clearly distinct from background noise. The furthest successful detection was obtained at 510 m. While the method showed promise in enhancing the identification process compared to traditional techniques, there were notable limitations in the clarity of the audio recordings. These findings suggest that while the integration of directional microphones and CNNs for long-range bird species identification is promising, further refinement is needed to fully realize the benefits of this approach. Future efforts should focus on improving the audio-capture technology to reduce ambient noise and enhance the system’s overall performance in long-range bird species identification. Full article
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)
Show Figures

Figure 1

15 pages, 4443 KiB  
Article
Potato Leaf Disease Detection Based on a Lightweight Deep Learning Model
by Chao-Yun Chang and Chih-Chin Lai
Mach. Learn. Knowl. Extr. 2024, 6(4), 2321-2335; https://doi.org/10.3390/make6040114 - 14 Oct 2024
Viewed by 2250
Abstract
Traditional methods of agricultural disease detection rely primarily on manual observation, which is not only time-consuming and labor-intensive, but also prone to human error. The advent of deep learning has revolutionized plant disease detection by providing more accurate and efficient solutions. The management [...] Read more.
Traditional methods of agricultural disease detection rely primarily on manual observation, which is not only time-consuming and labor-intensive, but also prone to human error. The advent of deep learning has revolutionized plant disease detection by providing more accurate and efficient solutions. The management of potato diseases is critical to the agricultural industry, as these diseases can lead to substantial losses in crop production. The prompt identification and classification of potato leaf diseases are essential to mitigating such losses. In this paper, we present a novel approach that integrates a lightweight convolutional neural network architecture, RegNetY-400MF, with transfer learning techniques to accurately identify seven different types of potato leaf diseases. The proposed method not only enhances the precision of potato leaf disease detection but also reduces the computational and storage demands, with a mere 0.40 GFLOPs and a model size of 16.8 MB. This makes it well-suited for use on edge devices with limited resources, enabling real-time disease detection in agricultural environments. The experimental results demonstrated that the accuracy of the proposed method in identifying seven potato leaf diseases was 90.68%, providing a comprehensive solution for potato crop management. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

18 pages, 893 KiB  
Article
Temporal Relational Graph Convolutional Network Approach to Financial Performance Prediction
by Brindha Priyadarshini Jeyaraman, Bing Tian Dai and Yuan Fang
Mach. Learn. Knowl. Extr. 2024, 6(4), 2303-2320; https://doi.org/10.3390/make6040113 - 10 Oct 2024
Viewed by 1646
Abstract
Accurately predicting financial entity performance remains a challenge due to the dynamic nature of financial markets and vast unstructured textual data. Financial knowledge graphs (FKGs) offer a structured representation for tackling this problem by representing complex financial relationships and concepts. However, constructing a [...] Read more.
Accurately predicting financial entity performance remains a challenge due to the dynamic nature of financial markets and vast unstructured textual data. Financial knowledge graphs (FKGs) offer a structured representation for tackling this problem by representing complex financial relationships and concepts. However, constructing a comprehensive and accurate financial knowledge graph that captures the temporal dynamics of financial entities is non-trivial. We introduce FintechKG, a comprehensive financial knowledge graph developed through a three-dimensional information extraction process that incorporates commercial entities and temporal dimensions and uses a financial concept taxonomy that ensures financial domain entity and relationship extraction. We propose a temporal and relational graph convolutional network (RGCN)-based representation for FintechKG data across multiple timesteps, which captures temporal dependencies. This representation is then combined with FinBERT embeddings through a projection layer, enabling a richer feature space. To demonstrate the efficacy of FintechKG, we evaluate its performance using the example task of financial performance prediction. A logistic regression model uses these combined features and social media embeddings for performance prediction. We classify whether the revenue will increase or decrease. This approach demonstrates the effectiveness of FintechKG combined with textual information for accurate financial forecasting. Our work contributes a systematic FKG construction method and a framework that utilizes both relational and textual embeddings for improved financial performance prediction. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

21 pages, 3922 KiB  
Article
Maximising Synergy: The Benefits of a Joint Implementation of Knowledge Management and Artificial Intelligence System Standards
by Natalia Khazieva, Alena Pauliková and Henrieta Hrablik Chovanová
Mach. Learn. Knowl. Extr. 2024, 6(4), 2282-2302; https://doi.org/10.3390/make6040112 - 8 Oct 2024
Cited by 2 | Viewed by 2004
Abstract
Implementing management systems in organisations of all types and sizes often raises the following question: “What benefits will this bring?” Initial resistance and criticism are common as potential challenges are identified during the implementation process. To address this, it is essential to highlight [...] Read more.
Implementing management systems in organisations of all types and sizes often raises the following question: “What benefits will this bring?” Initial resistance and criticism are common as potential challenges are identified during the implementation process. To address this, it is essential to highlight the advantages of these systems and engage stakeholders in supporting management efforts. While the planning, implementation, use, maintenance, auditing, and improvement of management systems are generally voluntary, certification is frequently driven by external factors, particularly customer demands. Employees also stand to gain significantly, with knowledge and information serving as valuable resources, especially for leveraging artificial intelligence. This article explores the management’s readiness to adopt and fully utilise two management systems based on international standards: the ISO 30401 Knowledge management system (KMS) and the ISO/IEC 42001 Artificial intelligence management system (AIMS). Through interviews, we assess the challenges and solutions associated with implementing these systems, whether planned or partially adopted. The findings illustrate the synergistic benefits of integrating the KMS and AIMS, demonstrating how their combined use can enhance Integrated Management Systems (IMSs). Such integration supports comprehensive planning, operation, and performance evaluation of processes and services while also promoting continuous improvement. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

34 pages, 15730 KiB  
Article
Empowering Brain Tumor Diagnosis through Explainable Deep Learning
by Zhengkun Li and Omar Dib
Mach. Learn. Knowl. Extr. 2024, 6(4), 2248-2281; https://doi.org/10.3390/make6040111 - 7 Oct 2024
Cited by 1 | Viewed by 3540
Abstract
Brain tumors are among the most lethal diseases, and early detection is crucial for improving patient outcomes. Currently, magnetic resonance imaging (MRI) is the most effective method for early brain tumor detection due to its superior imaging quality for soft tissues. However, manual [...] Read more.
Brain tumors are among the most lethal diseases, and early detection is crucial for improving patient outcomes. Currently, magnetic resonance imaging (MRI) is the most effective method for early brain tumor detection due to its superior imaging quality for soft tissues. However, manual analysis of brain MRI scans is prone to errors, largely influenced by the radiologists’ experience and fatigue. To address these challenges, computer-aided diagnosis (CAD) systems are more significant. These advanced computer vision techniques such as deep learning provide accurate predictions based on medical images, enhancing diagnostic precision and reliability. This paper presents a novel CAD framework for multi-class brain tumor classification. The framework employs six pre-trained deep learning models as the base and incorporates comprehensive data preprocessing and augmentation strategies to enhance computational efficiency. To address issues related to transparency and interpretability in deep learning models, Gradient-weighted Class Activation Mapping (Grad-CAM) is utilized to visualize the decision-making processes involved in tumor classification from MRI scans. Additionally, a user-friendly Brain Tumor Detection System has been developed using Streamlit, demonstrating its practical applicability in real-world settings and providing a valuable tool for clinicians. All simulation results are derived from a public benchmark dataset, showing that the proposed framework achieves state-of-the-art performance, with accuracy approaching 99% in ResNet-50, Xception, and InceptionV3 models. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

16 pages, 6338 KiB  
Article
Bayesian Optimization Using Simulation-Based Multiple Information Sources over Combinatorial Structures
by Antonio Sabbatella, Andrea Ponti, Antonio Candelieri and Francesco Archetti
Mach. Learn. Knowl. Extr. 2024, 6(4), 2232-2247; https://doi.org/10.3390/make6040110 - 5 Oct 2024
Viewed by 967
Abstract
Bayesian optimization due to its flexibility and sample efficiency has become a standard approach for simulation optimization. To reduce this problem, one can resort to cheaper surrogates of the objective function. Examples are ubiquitous, from protein engineering or material science to tuning machine [...] Read more.
Bayesian optimization due to its flexibility and sample efficiency has become a standard approach for simulation optimization. To reduce this problem, one can resort to cheaper surrogates of the objective function. Examples are ubiquitous, from protein engineering or material science to tuning machine learning algorithms, where one could use a subset of the full training set or even a smaller related dataset. Cheap information sources in the optimization scheme have been studied in the literature as the multi-fidelity optimization problem. Of course, cheaper sources may hold some promise toward tractability, but cheaper models offer an incomplete model inducing unknown bias and epistemic uncertainty. In this manuscript, we are concerned with the discrete case, where fx,wi is the value of the performance measure associated with the environmental condition wi and p(wi) represents the relevance of the condition wi (i.e., the probability of occurrence or the fraction of time this condition occurs). The main contribution of this paper is the proposal of a Gaussian-based framework, called augmented Gaussian process (AGP), based on sparsification, originally proposed for continuous functions and its generalization in this paper to stochastic optimization using different risk profiles for combinatorial optimization. The AGP leverages sample and cost-efficient Bayesian optimization (BO) of multiple information sources and supports a new acquisition function to select the new source–location pair considering the cost of the source and the (location-dependent) model discrepancy. An extensive set of computational results supports risk-aware optimization based on CVaR (conditional value-at-risk). Computational experiments confirm the actual performance of the MISO-AGP method and the hyperparameter optimization on benchmark functions and real-world problems. Full article
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop