Next Issue
Volume 6, December
Previous Issue
Volume 6, June
 
 

Mach. Learn. Knowl. Extr., Volume 6, Issue 3 (September 2024) – 40 articles

Cover Story (view full-size image): This study explores the impact of climate change on soil health by focusing on the temperature sensitivity of soil microbial respiration (Q10). Leveraging Explainable Artificial Intelligence (XAI), the research uncovers the key chemical, physical, and microbiological soil factors that influence Q10 values. Our findings reveal the pivotal role of the soil microbiome in driving soil respiration responses to warming. By identifying these critical variables, the study provides essential insights into soil carbon dynamics, informing the development of innovative strategies for climate change mitigation and sustainable soil management. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
19 pages, 3565 KiB  
Article
A Multi-Objective Framework for Balancing Fairness and Accuracy in Debiasing Machine Learning Models
by Rashmi Nagpal, Ariba Khan, Mihir Borkar and Amar Gupta
Mach. Learn. Knowl. Extr. 2024, 6(3), 2130-2148; https://doi.org/10.3390/make6030105 - 20 Sep 2024
Viewed by 1429
Abstract
Machine learning algorithms significantly impact decision-making in high-stakes domains, necessitating a balance between fairness and accuracy. This study introduces an in-processing, multi-objective framework that leverages the Reject Option Classification (ROC) algorithm to simultaneously optimize fairness and accuracy while safeguarding protected attributes such as [...] Read more.
Machine learning algorithms significantly impact decision-making in high-stakes domains, necessitating a balance between fairness and accuracy. This study introduces an in-processing, multi-objective framework that leverages the Reject Option Classification (ROC) algorithm to simultaneously optimize fairness and accuracy while safeguarding protected attributes such as age and gender. Our approach seeks a multi-objective optimization solution that balances accuracy, group fairness loss, and individual fairness loss. The framework integrates fairness objectives without relying on a weighted summation method, instead focusing on directly optimizing the trade-offs. Empirical evaluations on publicly available datasets, including German Credit, Adult Income, and COMPAS, reveal several significant findings: the ROC-based approach demonstrates superior performance, achieving an accuracy of 94.29%, an individual fairness loss of 0.04, and a group fairness loss of 0.06 on the German Credit dataset. These results underscore the effectiveness of our framework, particularly the ROC component, in enhancing both the fairness and performance of machine learning models. Full article
Show Figures

Figure 1

19 pages, 3484 KiB  
Article
Efficient Visual-Aware Fashion Recommendation Using Compressed Node Features and Graph-Based Learning
by Umar Subhan Malhi, Junfeng Zhou, Abdur Rasool and Shahbaz Siddeeq
Mach. Learn. Knowl. Extr. 2024, 6(3), 2111-2129; https://doi.org/10.3390/make6030104 - 15 Sep 2024
Viewed by 821
Abstract
In fashion e-commerce, predicting item compatibility using visual features remains a significant challenge. Current recommendation systems often struggle to incorporate high-dimensional visual data into graph-based learning models effectively. This limitation presents a substantial opportunity to enhance the precision and effectiveness of fashion recommendations. [...] Read more.
In fashion e-commerce, predicting item compatibility using visual features remains a significant challenge. Current recommendation systems often struggle to incorporate high-dimensional visual data into graph-based learning models effectively. This limitation presents a substantial opportunity to enhance the precision and effectiveness of fashion recommendations. In this paper, we present the Visual-aware Graph Convolutional Network (VAGCN). This novel framework helps improve how visual features can be incorporated into graph-based learning systems for fashion item compatibility predictions. The VAGCN framework employs a deep-stacked autoencoder to convert the input image’s high-dimensional raw CNN visual features into more manageable low-dimensional representations. In addition to improving feature representation, the GCN can also reason more intelligently about predictions, which would not be possible without this compression. The GCN encoder processes nodes in the graph to capture structural and feature correlation. Following the GCN encoder, the refined embeddings are input to a multi-layer perceptron (MLP) to calculate compatibility scores. The approach extends to using neighborhood information only during the testing phase to help with training efficiency and generalizability in practical scenarios, a key characteristic of our model. By leveraging its ability to capture latent visual features and neighborhood-based learning, VAGCN thoroughly investigates item compatibility across various categories. This method significantly improves predictive accuracy, consistently outperforming existing benchmarks. These contributions tackle significant scalability and computational efficiency challenges, showcasing the potential transformation of recommendation systems through enhanced feature representation, paving the way for further innovations in the fashion domain. Full article
(This article belongs to the Special Issue Machine Learning in Data Science)
Show Figures

Figure 1

15 pages, 5455 KiB  
Article
Show Me Once: A Transformer-Based Approach for an Assisted-Driving System
by Federico Pacini, Pierpaolo Dini and Luca Fanucci
Mach. Learn. Knowl. Extr. 2024, 6(3), 2096-2110; https://doi.org/10.3390/make6030103 - 13 Sep 2024
Viewed by 726
Abstract
Operating a powered wheelchair involves significant risks and requires considerable cognitive effort to maintain effective awareness of the surrounding environment. Therefore, people with significant disabilities are at a higher risk, leading to a decrease in their social interactions, which can impact their overall [...] Read more.
Operating a powered wheelchair involves significant risks and requires considerable cognitive effort to maintain effective awareness of the surrounding environment. Therefore, people with significant disabilities are at a higher risk, leading to a decrease in their social interactions, which can impact their overall health and well-being. Thus, we propose an intelligent driving-assistance system that innovatively uses Transformers, typically employed in Natural Language Processing, for navigation and a retrieval mechanism, allowing users to specify their destinations using natural language. The system records the areas visited and enables users to pinpoint these locations through descriptions, which will be considered later in the retrieval phase. Taking a foundational model, the system is fine-tuned with simulated data. The preliminary results demonstrate the system’s effectiveness compared to non-assisted solutions and its readiness for deployment on edge devices. Full article
(This article belongs to the Special Issue Advances in Machine and Deep Learning)
Show Figures

Figure 1

22 pages, 1814 KiB  
Article
A Data Science and Sports Analytics Approach to Decode Clutch Dynamics in the Last Minutes of NBA Games
by Vangelis Sarlis, Dimitrios Gerakas and Christos Tjortjis
Mach. Learn. Knowl. Extr. 2024, 6(3), 2074-2095; https://doi.org/10.3390/make6030102 - 13 Sep 2024
Viewed by 1761
Abstract
This research investigates clutch performance in the National Basketball Association (NBA) with a focus on the final minutes of contested games. By employing advanced data science techniques, we aim to identify key factors that enhance winning probabilities during these critical moments. The study [...] Read more.
This research investigates clutch performance in the National Basketball Association (NBA) with a focus on the final minutes of contested games. By employing advanced data science techniques, we aim to identify key factors that enhance winning probabilities during these critical moments. The study introduces the Estimation of Clutch Competency (EoCC) metric, which is a novel formula designed to evaluate players’ impact under pressure. Examining player performance statistics over twenty seasons, this research addresses a significant gap in the literature regarding the quantification of clutch moments and challenges conventional wisdom in basketball analytics. Our findings deal valuable insights into player efficiency during the final minutes and its impact on the probabilities of a positive outcome. The EoCC metric’s validation through comparison with the NBA Clutch Player of the Year voting results demonstrates its effectiveness in identifying top performers in high-pressure situations. Leveraging state-of-the-art data science techniques and algorithms, this study analyzes play data to uncover key factors contributing to a team’s success in pivotal moments. This research not only enhances the theoretical understanding of clutch dynamics but also provides practical insights for coaches, analysts, and the broader sports community. It contributes to more informed decision making in high-stakes basketball environments, advancing the field of sports analytics. Full article
Show Figures

Figure 1

25 pages, 8181 KiB  
Article
A Novel Integration of Data-Driven Rule Generation and Computational Argumentation for Enhanced Explainable AI
by Lucas Rizzo, Damiano Verda, Serena Berretta and Luca Longo
Mach. Learn. Knowl. Extr. 2024, 6(3), 2049-2073; https://doi.org/10.3390/make6030101 - 12 Sep 2024
Viewed by 694
Abstract
Explainable Artificial Intelligence (XAI) is a research area that clarifies AI decision-making processes to build user trust and promote responsible AI. Hence, a key scientific challenge in XAI is the development of methods that generate transparent and interpretable explanations while maintaining scalability and [...] Read more.
Explainable Artificial Intelligence (XAI) is a research area that clarifies AI decision-making processes to build user trust and promote responsible AI. Hence, a key scientific challenge in XAI is the development of methods that generate transparent and interpretable explanations while maintaining scalability and effectiveness in complex scenarios. Rule-based methods in XAI generate rules that can potentially explain AI inferences, yet they can also become convoluted in large scenarios, hindering their readability and scalability. Moreover, they often lack contrastive explanations, leaving users uncertain why specific predictions are preferred. To address this scientific problem, we explore the integration of computational argumentation—a sub-field of AI that models reasoning processes through defeasibility—into rule-based XAI systems. Computational argumentation enables arguments modelled from rules to be retracted based on new evidence. This makes it a promising approach to enhancing rule-based methods for creating more explainable AI systems. Nonetheless, research on their integration remains limited despite the appealing properties of rule-based systems and computational argumentation. Therefore, this study also addresses the applied challenge of implementing such an integration within practical AI tools. The study employs the Logic Learning Machine (LLM), a specific rule-extraction technique, and presents a modular design that integrates input rules into a structured argumentation framework using state-of-the-art computational argumentation methods. Experiments conducted on binary classification problems using various datasets from the UCI Machine Learning Repository demonstrate the effectiveness of this integration. The LLM technique excelled in producing a manageable number of if-then rules with a small number of premises while maintaining high inferential capacity for all datasets. In turn, argument-based models achieved comparable results to those derived directly from if-then rules, leveraging a concise set of rules and excelling in explainability. In summary, this paper introduces a novel approach for efficiently and automatically generating arguments and their interactions from data, addressing both scientific and applied challenges in advancing the application and deployment of argumentation systems in XAI. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

16 pages, 2094 KiB  
Article
Graph Convolutional Networks for Predicting Cancer Outcomes and Stage: A Focus on cGAS-STING Pathway Activation
by Mateo Sokač, Borna Skračić, Danijel Kučak and Leo Mršić
Mach. Learn. Knowl. Extr. 2024, 6(3), 2033-2048; https://doi.org/10.3390/make6030100 - 11 Sep 2024
Viewed by 1492
Abstract
The study presented in this paper evaluated gene expression profiles from The Cancer Genome Atlas (TCGA). To reduce complexity, we focused on genes in the cGAS–STING pathway, crucial for cytosolic DNA detection and immune response. The study analyzes three clinical variables: disease-specific survival [...] Read more.
The study presented in this paper evaluated gene expression profiles from The Cancer Genome Atlas (TCGA). To reduce complexity, we focused on genes in the cGAS–STING pathway, crucial for cytosolic DNA detection and immune response. The study analyzes three clinical variables: disease-specific survival (DSS), overall survival (OS), and tumor stage. To effectively utilize the high-dimensional gene expression data, we needed to find a way to project these data meaningfully. Since gene pathways can be represented as graphs, a novel method of presenting genomics data using graph data structure was employed, rather than the conventional tabular format. To leverage the gene expression data represented as graphs, we utilized a graph convolutional network (GCN) machine learning model in conjunction with the genetic algorithm optimization technique. This allowed for obtaining an optimal graph representation topology and capturing important activations within the pathway for each use case, enabling a more insightful analysis of the cGAS–STING pathway and its activations across different cancer types and clinical variables. To tackle the problem of unexplainable AI, graph visualization alongside the integrated gradients method was employed to explain the GCN model’s decision-making process, identifying key nodes (genes) in the cGAS–STING pathway. This approach revealed distinct molecular mechanisms, enhancing interpretability. This study demonstrates the potential of GCNs combined with explainable AI to analyze gene expression, providing insights into cancer progression. Further research with more data is needed to validate these findings. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

15 pages, 5499 KiB  
Article
Correlating Histopathological Microscopic Images of Creutzfeldt–Jakob Disease with Clinical Typology Using Graph Theory and Artificial Intelligence
by Carlos Martínez, Susana Teijeira, Patricia Domínguez, Silvia Campanioni, Laura Busto, José A. González-Nóvoa, Jacobo Alonso, Eva Poveda, Beatriz San Millán and César Veiga
Mach. Learn. Knowl. Extr. 2024, 6(3), 2018-2032; https://doi.org/10.3390/make6030099 - 7 Sep 2024
Viewed by 829
Abstract
Creutzfeldt–Jakob disease (CJD) is a rare, degenerative, and fatal brain disorder caused by abnormal proteins called prions. This research introduces a novel approach combining AI and graph theory to analyze histopathological microscopic images of brain tissues affected by CJD. The detection and quantification [...] Read more.
Creutzfeldt–Jakob disease (CJD) is a rare, degenerative, and fatal brain disorder caused by abnormal proteins called prions. This research introduces a novel approach combining AI and graph theory to analyze histopathological microscopic images of brain tissues affected by CJD. The detection and quantification of spongiosis, characterized by the presence of vacuoles in the brain tissue, plays a crucial role in aiding the accurate diagnosis of CJD. The proposed methodology employs image processing techniques to identify these pathological features in high-resolution medical images. By developing an automatic pipeline for the detection of spongiosis, we aim to overcome some limitations of manual feature extraction. The results demonstrate that our method correctly identifies and characterize spongiosis and allows the extraction of features that will help to better understand the spongiosis patterns in different CJD patients. Full article
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)
Show Figures

Figure 1

21 pages, 748 KiB  
Systematic Review
Tertiary Review on Explainable Artificial Intelligence: Where Do We Stand?
by Frank van Mourik, Annemarie Jutte, Stijn E. Berendse, Faiza A. Bukhsh and Faizan Ahmed
Mach. Learn. Knowl. Extr. 2024, 6(3), 1997-2017; https://doi.org/10.3390/make6030098 - 30 Aug 2024
Viewed by 1302
Abstract
Research into explainable artificial intelligence (XAI) methods has exploded over the past five years. It is essential to synthesize and categorize this research and, for this purpose, multiple systematic reviews on XAI mapped out the landscape of the existing methods. To understand how [...] Read more.
Research into explainable artificial intelligence (XAI) methods has exploded over the past five years. It is essential to synthesize and categorize this research and, for this purpose, multiple systematic reviews on XAI mapped out the landscape of the existing methods. To understand how these methods have developed and been applied and what evidence has been accumulated through model training and analysis, we carried out a tertiary literature review that takes as input systematic literature reviews published between 1992 and 2023. We evaluated 40 systematic literature review papers and presented binary tabular overviews of researched XAI methods and their respective characteristics, such as the scope, scale, input data, explanation data, and machine learning models researched. We identified seven distinct characteristics and organized them into twelve specific categories, culminating in the creation of comprehensive research grids. Within these research grids, we systematically documented the presence or absence of research mentions for each pairing of characteristic and category. We identified 14 combinations that are open to research. Our findings reveal a significant gap, particularly in categories like the cross-section of feature graphs and numerical data, which appear to be notably absent or insufficiently addressed in the existing body of research and thus represent a future research road map. Full article
(This article belongs to the Special Issue Machine Learning in Data Science)
Show Figures

Figure 1

28 pages, 1736 KiB  
Article
Black Box Adversarial Reprogramming for Time Series Feature Classification in Ball Bearings’ Remaining Useful Life Classification
by Alexander Bott, Felix Schreyer, Alexander Puchta and Jürgen Fleischer
Mach. Learn. Knowl. Extr. 2024, 6(3), 1969-1996; https://doi.org/10.3390/make6030097 - 27 Aug 2024
Viewed by 930
Abstract
Standard ML relies on ample data, but limited availability poses challenges. Transfer learning offers a solution by leveraging pre-existing knowledge. Yet many methods require access to the model’s internal aspects, limiting applicability to white box models. To address this, Tsai, Chen and Ho [...] Read more.
Standard ML relies on ample data, but limited availability poses challenges. Transfer learning offers a solution by leveraging pre-existing knowledge. Yet many methods require access to the model’s internal aspects, limiting applicability to white box models. To address this, Tsai, Chen and Ho introduced Black Box Adversarial Reprogramming for transfer learning with black box models. While tested primarily in image classification, this paper explores its potential in time series classification, particularly predictive maintenance. We develop an adversarial reprogramming concept tailored to black box time series classifiers. Our study focuses on predicting the Remaining Useful Life of rolling bearings. We construct a comprehensive ML pipeline, encompassing feature engineering and model fine-tuning, and compare results with traditional transfer learning. We investigate the impact of hyperparameters and training parameters on model performance, demonstrating the successful application of Black Box Adversarial Reprogramming to time series data. The method achieved a weighted F1-score of 0.77, although it exhibited significant stochastic fluctuations, with scores ranging from 0.3 to 0.77 due to randomness in gradient estimation. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

16 pages, 956 KiB  
Article
Assessing Fine-Tuned NER Models with Limited Data in French: Automating Detection of New Technologies, Technological Domains, and Startup Names in Renewable Energy
by Connor MacLean and Denis Cavallucci
Mach. Learn. Knowl. Extr. 2024, 6(3), 1953-1968; https://doi.org/10.3390/make6030096 - 27 Aug 2024
Viewed by 2462
Abstract
Achieving carbon neutrality by 2050 requires unprecedented technological, economic, and sociological changes. With time as a scarce resource, it is crucial to base decisions on relevant facts and information to avoid misdirection. This study aims to help decision makers quickly find relevant information [...] Read more.
Achieving carbon neutrality by 2050 requires unprecedented technological, economic, and sociological changes. With time as a scarce resource, it is crucial to base decisions on relevant facts and information to avoid misdirection. This study aims to help decision makers quickly find relevant information related to companies and organizations in the renewable energy sector. In this study, we propose fine-tuning five RNN and transformer models trained for French on a new category, “TECH”. This category is used to classify technological domains and new products. In addition, as the model is fine-tuned on news related to startups, we note an improvement in the detection of startup and company names in the “ORG” category. We further explore the capacities of the most effective model to accurately predict entities using a small amount of training data. We show the progression of the model from being trained on several hundred to several thousand annotations. This analysis allows us to demonstrate the potential of these models to extract insights without large corpora, allowing us to reduce the long process of annotating custom training data. This approach is used to automatically extract new company mentions as well as to extract technologies and technology domains that are currently being discussed in the news in order to better analyze industry trends. This approach further allows to group together mentions of specific energy domains with the companies that are actively developing new technologies in the field. Full article
Show Figures

Figure 1

17 pages, 2683 KiB  
Article
Forecasting the Right Crop Nutrients for Specific Crops Based on Collected Data Using an Artificial Neural Network (ANN)
by Sairoel Amertet and Girma Gebresenbet
Mach. Learn. Knowl. Extr. 2024, 6(3), 1936-1952; https://doi.org/10.3390/make6030095 - 26 Aug 2024
Viewed by 974
Abstract
In farming technologies, it is difficult to properly provide the accurate crop nutrients for respective crops. For this reason, farmers are experiencing enormous problems. Although various types of machine learning (deep learning and convolutional neural networks) have been used to identify crop diseases, [...] Read more.
In farming technologies, it is difficult to properly provide the accurate crop nutrients for respective crops. For this reason, farmers are experiencing enormous problems. Although various types of machine learning (deep learning and convolutional neural networks) have been used to identify crop diseases, as has crop classification-based image processing, they have failed to forecast accurate crop nutrients for various crops, as crop nutrients are numerical instead of visual. Neural networks represent an opportunity for the precision agriculture sector to more accurately forecast crop nutrition. Recent technological advancements in neural networks have begun to provide greater precision, with an array of opportunities in pattern recognition. Neural networks represent an opportunity to effectively solve numerical data problems. The aim of the current study is to estimate the right crop nutrients for the right crops based on the data collected using an artificial neural network. The crop data were collected from the MNIST dataset. To forecast the precise nutrients for the crops, ANN models were developed. The entire system was simulated in a MATLAB environment. The obtained results for forecasting accurate nutrients were 99.997%, 99.996%, and 99.997% for validation, training, and testing, respectively. Therefore, the proposed algorithm is suitable for forecasting accurate crop nutrients for the crops. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

15 pages, 1283 KiB  
Article
Optimal Knowledge Distillation through Non-Heuristic Control of Dark Knowledge
by Darian Onchis, Codruta Istin and Ioan Samuila
Mach. Learn. Knowl. Extr. 2024, 6(3), 1921-1935; https://doi.org/10.3390/make6030094 - 22 Aug 2024
Viewed by 1237
Abstract
In this paper, a method is introduced to control the dark knowledge values also known as soft targets, with the purpose of improving the training by knowledge distillation for multi-class classification tasks. Knowledge distillation effectively transfers knowledge from a larger model to a [...] Read more.
In this paper, a method is introduced to control the dark knowledge values also known as soft targets, with the purpose of improving the training by knowledge distillation for multi-class classification tasks. Knowledge distillation effectively transfers knowledge from a larger model to a smaller model to achieve efficient, fast, and generalizable performance while retaining much of the original accuracy. The majority of deep neural models used for classification tasks append a SoftMax layer to generate output probabilities and it is usual to take the highest score and consider it the inference of the model, while the rest of the probability values are generally ignored. The focus is on those probabilities as carriers of dark knowledge and our aim is to quantify the relevance of dark knowledge, not heuristically as provided in the literature so far, but with an inductive proof on the SoftMax operational limits. These limits are further pushed by using an incremental decision tree with information gain split. The user can set a desired precision and an accuracy level to obtain a maximal temperature setting for a continual classification process. Moreover, by fitting both the hard targets and the soft targets, one obtains an optimal knowledge distillation effect that mitigates better catastrophic forgetting. The strengths of our method come from the possibility of controlling the amount of distillation transferred non-heuristically and the agnostic application of this model-independent study. Full article
Show Figures

Figure 1

28 pages, 7677 KiB  
Article
Visual Reasoning and Multi-Agent Approach in Multimodal Large Language Models (MLLMs): Solving TSP and mTSP Combinatorial Challenges
by Mohammed Elhenawy, Ahmad Abutahoun, Taqwa I. Alhadidi, Ahmed Jaber, Huthaifa I. Ashqar, Shadi Jaradat, Ahmed Abdelhay, Sebastien Glaser and Andry Rakotonirainy
Mach. Learn. Knowl. Extr. 2024, 6(3), 1894-1920; https://doi.org/10.3390/make6030093 - 13 Aug 2024
Cited by 2 | Viewed by 1004
Abstract
Multimodal Large Language Models (MLLMs) harness comprehensive knowledge spanning text, images, and audio to adeptly tackle complex problems. This study explores the ability of MLLMs in visually solving the Traveling Salesman Problem (TSP) and Multiple Traveling Salesman Problem (mTSP) using images that portray [...] Read more.
Multimodal Large Language Models (MLLMs) harness comprehensive knowledge spanning text, images, and audio to adeptly tackle complex problems. This study explores the ability of MLLMs in visually solving the Traveling Salesman Problem (TSP) and Multiple Traveling Salesman Problem (mTSP) using images that portray point distributions on a two-dimensional plane. We introduce a novel approach employing multiple specialized agents within the MLLM framework, each dedicated to optimizing solutions for these combinatorial challenges. We benchmarked our multi-agent model solutions against the Google OR tools, which served as the baseline for comparison. The results demonstrated that both multi-agent models—Multi-Agent 1, which includes the initializer, critic, and scorer agents, and Multi-Agent 2, which comprises only the initializer and critic agents—significantly improved the solution quality for TSP and mTSP problems. Multi-Agent 1 excelled in environments requiring detailed route refinement and evaluation, providing a robust framework for sophisticated optimizations. In contrast, Multi-Agent 2, focusing on iterative refinements by the initializer and critic, proved effective for rapid decision-making scenarios. These experiments yield promising outcomes, showcasing the robust visual reasoning capabilities of MLLMs in addressing diverse combinatorial problems. The findings underscore the potential of MLLMs as powerful tools in computational optimization, offering insights that could inspire further advancements in this promising field. Full article
Show Figures

Figure 1

23 pages, 4393 KiB  
Article
Balancing Results from AI-Based Geostatistics versus Fuzzy Inference by Game Theory Bargaining to Improve a Groundwater Monitoring Network
by Masoumeh Hashemi, Richard C. Peralta and Matt Yost
Mach. Learn. Knowl. Extr. 2024, 6(3), 1871-1893; https://doi.org/10.3390/make6030092 - 9 Aug 2024
Viewed by 1415
Abstract
An artificial intelligence-based geostatistical optimization algorithm was developed to upgrade a test Iranian aquifer’s existing groundwater monitoring network. For that aquifer, a preliminary study revealed that a Multi-Layer Perceptron Artificial Neural Network (MLP-ANN) more accurately determined temporally average water table elevations than geostatistical [...] Read more.
An artificial intelligence-based geostatistical optimization algorithm was developed to upgrade a test Iranian aquifer’s existing groundwater monitoring network. For that aquifer, a preliminary study revealed that a Multi-Layer Perceptron Artificial Neural Network (MLP-ANN) more accurately determined temporally average water table elevations than geostatistical kriging, spline, and inverse distance weighting. Because kriging is usually used in that area for water table estimation, the developed algorithm used MLP-ANN to guide kriging, and Genetic Algorithm (GA) to determine locations for new monitoring well location(s). For possible annual fiscal budgets allowing 1–12 new wells, 12 sets of optimal new well locations are reported. Each set has the locations of new wells that would minimize the squared difference between the time-averaged heads developed by kriging versus MLP-ANN. Also, to simultaneously consider local expertise, the algorithm used fuzzy inference to quantify an expert’s satisfaction with the number of new wells. Then, the algorithm used symmetric bargaining (Nash, Kalai–Smorodinsky, and area monotonic) to present an upgradation strategy that balanced professional judgment and heuristic optimization. In essence, the algorithm demonstrates the systematic application of relatively new computational practices to a common situation worldwide. Full article
(This article belongs to the Special Issue Sustainable Applications for Machine Learning)
Show Figures

Figure 1

14 pages, 7188 KiB  
Article
Accuracy Improvement of Debonding Damage Detection Technology in Composite Blade Joints for 20 kW Class Wind Turbine
by Hakgeun Kim, Hyeongjin Kim and Kiweon Kang
Mach. Learn. Knowl. Extr. 2024, 6(3), 1857-1870; https://doi.org/10.3390/make6030091 - 7 Aug 2024
Viewed by 775
Abstract
Securing the structural safety of blades has become crucial, owing to the increasing size and weight of blades resulting from the recent development of large wind turbines. Composites are primarily used for blade manufacturing because of their high specific strength and specific stiffness. [...] Read more.
Securing the structural safety of blades has become crucial, owing to the increasing size and weight of blades resulting from the recent development of large wind turbines. Composites are primarily used for blade manufacturing because of their high specific strength and specific stiffness. However, in composite blades, joints may experience fractures from the loads generated during wind turbine operation, leading to deformation caused by changes in structural stiffness. In this study, 7132 debonding damage data, classified by damage type, position, and size, were selected to predict debonding damage based on natural frequency. The change in the natural frequency caused by debonding damage was acquired through finite element (FE) modeling and modal analysis. Synchronization between the FE analysis model and manufactured blades was achieved through modal testing and data analysis. Finally, the relationship between debonding damage and the change in natural frequency was examined using artificial neural network techniques. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

17 pages, 786 KiB  
Article
A Parallel Approach to Enhance the Performance of Supervised Machine Learning Realized in a Multicore Environment
by Ashutosh Ghimire and Fathi Amsaad
Mach. Learn. Knowl. Extr. 2024, 6(3), 1840-1856; https://doi.org/10.3390/make6030090 - 2 Aug 2024
Viewed by 1894
Abstract
Machine learning models play a critical role in applications such as image recognition, natural language processing, and medical diagnosis, where accuracy and efficiency are paramount. As datasets grow in complexity, so too do the computational demands of classification techniques. Previous research has achieved [...] Read more.
Machine learning models play a critical role in applications such as image recognition, natural language processing, and medical diagnosis, where accuracy and efficiency are paramount. As datasets grow in complexity, so too do the computational demands of classification techniques. Previous research has achieved high accuracy but required significant computational time. This paper proposes a parallel architecture for Ensemble Machine Learning Models, harnessing multicore CPUs to expedite performance. The primary objective is to enhance machine learning efficiency without compromising accuracy through parallel computing. This study focuses on benchmark ensemble models including Random Forest, XGBoost, ADABoost, and K Nearest Neighbors. These models are applied to tasks such as wine quality classification and fraud detection in credit card transactions. The results demonstrate that, compared to single-core processing, machine learning tasks run 1.7 times and 3.8 times faster for small and large datasets on quad-core CPUs, respectively. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

22 pages, 2817 KiB  
Article
Enhanced Graph Representation Convolution: Effective Inferring Gene Regulatory Network Using Graph Convolution Network with Self-Attention Graph Pooling Layer
by Duaa Mohammad Alawad, Ataur Katebi and Md Tamjidul Hoque
Mach. Learn. Knowl. Extr. 2024, 6(3), 1818-1839; https://doi.org/10.3390/make6030089 - 1 Aug 2024
Viewed by 1374
Abstract
Studying gene regulatory networks (GRNs) is paramount for unraveling the complexities of biological processes and their associated disorders, such as diabetes, cancer, and Alzheimer’s disease. Recent advancements in computational biology have aimed to enhance the inference of GRNs from gene expression data, a [...] Read more.
Studying gene regulatory networks (GRNs) is paramount for unraveling the complexities of biological processes and their associated disorders, such as diabetes, cancer, and Alzheimer’s disease. Recent advancements in computational biology have aimed to enhance the inference of GRNs from gene expression data, a non-trivial task given the networks’ intricate nature. The challenge lies in accurately identifying the myriad interactions among transcription factors and target genes, which govern cellular functions. This research introduces a cutting-edge technique, EGRC (Effective GRN Inference applying Graph Convolution with Self-Attention Graph Pooling), which innovatively conceptualizes GRN reconstruction as a graph classification problem, where the task is to discern the links within subgraphs that encapsulate pairs of nodes. By leveraging Spearman’s correlation, we generate potential subgraphs that bring nonlinear associations between transcription factors and their targets to light. We use mutual information to enhance this, capturing a broader spectrum of gene interactions. Our methodology bifurcates these subgraphs into ‘Positive’ and ‘Negative’ categories. ‘Positive’ subgraphs are those where a transcription factor and its target gene are connected, including interactions among their neighbors. ‘Negative’ subgraphs, conversely, denote pairs without a direct connection. EGRC utilizes dual graph convolution network (GCN) models that exploit node attributes from gene expression profiles and graph embedding techniques to classify these. The performance of EGRC is substantiated by comprehensive evaluations using the DREAM5 datasets. Notably, EGRC attained an AUROC of 0.856 and an AUPR of 0.841 on the E. coli dataset. In contrast, the in silico dataset achieved an AUROC of 0.5058 and an AUPR of 0.958. Furthermore, on the S. cerevisiae dataset, EGRC recorded an AUROC of 0.823 and an AUPR of 0.822. These results underscore the robustness of EGRC in accurately inferring GRNs across various organisms. The advanced performance of EGRC represents a substantial advancement in the field, promising to deepen our comprehension of the intricate biological processes and their implications in both health and disease. Full article
Show Figures

Figure 1

20 pages, 671 KiB  
Article
Learning Optimal Dynamic Treatment Regime from Observational Clinical Data through Reinforcement Learning
by Seyum Abebe, Irene Poli, Roger D. Jones and Debora Slanzi
Mach. Learn. Knowl. Extr. 2024, 6(3), 1798-1817; https://doi.org/10.3390/make6030088 - 30 Jul 2024
Viewed by 1150
Abstract
In medicine, dynamic treatment regimes (DTRs) have emerged to guide personalized treatment decisions for patients, accounting for their unique characteristics. However, existing methods for determining optimal DTRs face limitations, often due to reliance on linear models unsuitable for complex disease analysis and a [...] Read more.
In medicine, dynamic treatment regimes (DTRs) have emerged to guide personalized treatment decisions for patients, accounting for their unique characteristics. However, existing methods for determining optimal DTRs face limitations, often due to reliance on linear models unsuitable for complex disease analysis and a focus on outcome prediction over treatment effect estimation. To overcome these challenges, decision tree-based reinforcement learning approaches have been proposed. Our study aims to evaluate the performance and feasibility of such algorithms: tree-based reinforcement learning (T-RL), DTR-Causal Tree (DTR-CT), DTR-Causal Forest (DTR-CF), stochastic tree-based reinforcement learning (SL-RL), and Q-learning with Random Forest. Using real-world clinical data, we conducted experiments to compare algorithm performances. Evaluation metrics included the proportion of correctly assigned patients to recommended treatments and the empirical mean with standard deviation of expected counterfactual outcomes based on estimated optimal treatment strategies. This research not only highlights the potential of decision tree-based reinforcement learning for dynamic treatment regimes but also contributes to advancing personalized medicine by offering nuanced and effective treatment recommendations. Full article
(This article belongs to the Section Learning)
16 pages, 1999 KiB  
Article
Insights from Augmented Data Integration and Strong Regularization in Drug Synergy Prediction with SynerGNet
by Mengmeng Liu, Gopal Srivastava, J. Ramanujam and Michal Brylinski
Mach. Learn. Knowl. Extr. 2024, 6(3), 1782-1797; https://doi.org/10.3390/make6030087 - 29 Jul 2024
Viewed by 887
Abstract
SynerGNet is a novel approach to predicting drug synergy against cancer cell lines. In this study, we discuss in detail the construction process of SynerGNet, emphasizing its comprehensive design tailored to handle complex data patterns. Additionally, we investigate a counterintuitive phenomenon when integrating [...] Read more.
SynerGNet is a novel approach to predicting drug synergy against cancer cell lines. In this study, we discuss in detail the construction process of SynerGNet, emphasizing its comprehensive design tailored to handle complex data patterns. Additionally, we investigate a counterintuitive phenomenon when integrating more augmented data into the training set results in an increase in testing loss alongside improved predictive accuracy. This sheds light on the nuanced dynamics of model learning. Further, we demonstrate the effectiveness of strong regularization techniques in mitigating overfitting, ensuring the robustness and generalization ability of SynerGNet. Finally, the continuous performance enhancements achieved through the integration of augmented data are highlighted. By gradually increasing the amount of augmented data in the training set, we observe substantial improvements in model performance. For instance, compared to models trained exclusively on the original data, the integration of the augmented data can lead to a 5.5% increase in the balanced accuracy and a 7.8% decrease in the false positive rate. Through rigorous benchmarks and analyses, our study contributes valuable insights into the development and optimization of predictive models in biomedical research. Full article
(This article belongs to the Special Issue Machine Learning in Data Science)
Show Figures

Figure 1

20 pages, 1138 KiB  
Article
Diverse Machine Learning for Forecasting Goal-Scoring Likelihood in Elite Football Leagues
by Christina Markopoulou, George Papageorgiou and Christos Tjortjis
Mach. Learn. Knowl. Extr. 2024, 6(3), 1762-1781; https://doi.org/10.3390/make6030086 - 28 Jul 2024
Cited by 1 | Viewed by 1623
Abstract
The field of sports analytics has grown rapidly, with a primary focus on performance forecasting, enhancing the understanding of player capabilities, and indirectly benefiting team strategies and player development. This work aims to forecast and comparatively evaluate players’ goal-scoring likelihood in four elite [...] Read more.
The field of sports analytics has grown rapidly, with a primary focus on performance forecasting, enhancing the understanding of player capabilities, and indirectly benefiting team strategies and player development. This work aims to forecast and comparatively evaluate players’ goal-scoring likelihood in four elite football leagues (Premier League, Bundesliga, La Liga, and Serie A) by mining advanced statistics from 2017 to 2023. Six types of machine learning (ML) models were developed and tested individually through experiments on the comprehensive datasets collected for these leagues. We also tested the upper 30th percentile of the best-performing players based on their performance in the last season, with varied features evaluated to enhance prediction accuracy in distinct scenarios. The results offer insights into the forecasting abilities of those leagues, identifying the best forecasting methodologies and the factors that most significantly contribute to the prediction of players’ goal-scoring. XGBoost consistently outperformed other models in most experiments, yielding the most accurate results and leading to a well-generalized model. Notably, when applied to Serie A, it achieved a mean absolute error (MAE) of 1.29. This study provides insights into ML-based performance prediction, advancing the field of player performance forecasting. Full article
Show Figures

Figure 1

42 pages, 16635 KiB  
Article
Towards AI Dashboards in Financial Services: Design and Implementation of an AI Development Dashboard for Credit Assessment
by Mustafa Pamuk and Matthias Schumann
Mach. Learn. Knowl. Extr. 2024, 6(3), 1720-1761; https://doi.org/10.3390/make6030085 - 27 Jul 2024
Viewed by 1699
Abstract
Financial institutions are increasingly turning to artificial intelligence (AI) to improve their decision-making processes and gain a competitive edge. Due to the iterative process of AI development, it is mandatory to have a structured process in place, from the design to the deployment [...] Read more.
Financial institutions are increasingly turning to artificial intelligence (AI) to improve their decision-making processes and gain a competitive edge. Due to the iterative process of AI development, it is mandatory to have a structured process in place, from the design to the deployment of AI-based services in the finance industry. This process must include the required validation and coordination with regulatory authorities. An appropriate dashboard can help to shape and structure the process of model development, e.g., for credit assessment in the finance industry. In addition, the analysis of datasets must be included as an important part of the dashboard to understand the reasons for changes in model performance. Furthermore, a dashboard can undertake documentation tasks to make the process of model development traceable, explainable, and transparent, as required by regulatory authorities in the finance industry. This can offer a comprehensive solution for financial companies to optimize their models, improve regulatory compliance, and ultimately foster sustainable growth in an increasingly competitive market. In this study, we investigate the requirements and provide a prototypical dashboard to create, manage, compare, and validate AI models to be used in the credit assessment of private customers. Full article
(This article belongs to the Special Issue Sustainable Applications for Machine Learning)
Show Figures

Figure 1

21 pages, 3362 KiB  
Article
Assessing the Value of Transfer Learning Metrics for Radio Frequency Domain Adaptation
by Lauren J. Wong, Braeden P. Muller, Sean McPherson and Alan J. Michaels
Mach. Learn. Knowl. Extr. 2024, 6(3), 1699-1719; https://doi.org/10.3390/make6030084 - 25 Jul 2024
Viewed by 747
Abstract
The use of transfer learning (TL) techniques has become common practice in fields such as computer vision (CV) and natural language processing (NLP). Leveraging prior knowledge gained from data with different distributions, TL offers higher performance and reduced training time, but has yet [...] Read more.
The use of transfer learning (TL) techniques has become common practice in fields such as computer vision (CV) and natural language processing (NLP). Leveraging prior knowledge gained from data with different distributions, TL offers higher performance and reduced training time, but has yet to be fully utilized in applications of machine learning (ML) and deep learning (DL) techniques and applications related to wireless communications, a field loosely termed radio frequency machine learning (RFML). This work examines whether existing transferability metrics, used in other modalities, might be useful in the context of RFML. Results show that the two existing metrics tested, Log Expected Empirical Prediction (LEEP) and Logarithm of Maximum Evidence (LogME), correlate well with post-transfer accuracy and can therefore be used to select source models for radio frequency (RF) domain adaptation and to predict post-transfer accuracy. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

26 pages, 3308 KiB  
Article
Enhancing Visitor Forecasting with Target-Concatenated Autoencoder and Ensemble Learning
by Ray-I Chang, Chih-Yung Tsai and Yu-Wei Chang
Mach. Learn. Knowl. Extr. 2024, 6(3), 1673-1698; https://doi.org/10.3390/make6030083 - 25 Jul 2024
Viewed by 840
Abstract
Accurate forecasting of inbound visitor numbers is crucial for effective planning and resource allocation in the tourism industry. Preceding forecasting algorithms primarily focused on time series analysis, often overlooking influential factors such as economic conditions. Regression models, on the other hand, face challenges [...] Read more.
Accurate forecasting of inbound visitor numbers is crucial for effective planning and resource allocation in the tourism industry. Preceding forecasting algorithms primarily focused on time series analysis, often overlooking influential factors such as economic conditions. Regression models, on the other hand, face challenges when dealing with high-dimensional data. Previous autoencoders for feature selection do not simultaneously incorporate feature and target information simultaneously, potentially limiting their effectiveness in improving predictive performance. This study presents a novel approach that combines a target-concatenated autoencoder (TCA) with ensemble learning to enhance the accuracy of tourism demand predictions. The TCA method integrates the prediction target into the training process, ensuring that the learned feature representations are optimized for specific forecasting tasks. Extensive experiments conducted on the Taiwan and Hawaii datasets demonstrate that the proposed TCA method significantly outperforms traditional feature selection techniques and other advanced algorithms in terms of the mean absolute percentage error (MAPE), mean absolute error (MAE), and coefficient of determination (R2). The results show that TCA combined with XGBoost achieves MAPE values of 3.3947% and 4.0059% for the Taiwan and Hawaii datasets, respectively, indicating substantial improvements over existing methods. Additionally, the proposed approach yields better R2 and MAE metrics than existing methods, further demonstrating its effectiveness. This study highlights the potential of TCA in providing reliable and accurate forecasts, thereby supporting strategic planning, infrastructure development, and sustainable growth in the tourism sector. Future research is advised to explore real-time data integration, expanded feature sets, and hybrid modeling approaches to further enhance the capabilities of the proposed framework. Full article
(This article belongs to the Special Issue Sustainable Applications for Machine Learning)
Show Figures

Figure 1

3 pages, 500 KiB  
Reply
Reply to Damaševičius, R. Comment on “Novozhilova et al. More Capable, Less Benevolent: Trust Perceptions of AI Systems across Societal Contexts. Mach. Learn. Knowl. Extr. 2024, 6, 342–366”
by Ekaterina Novozhilova, Kate Mays, Sejin Paik and James Katz
Mach. Learn. Knowl. Extr. 2024, 6(3), 1670-1672; https://doi.org/10.3390/make6030082 - 22 Jul 2024
Viewed by 584
Abstract
We would like to thank Dr [...] Full article
(This article belongs to the Special Issue Fairness and Explanation for Trustworthy AI)
3 pages, 498 KiB  
Comment
Comment on Novozhilova et al. More Capable, Less Benevolent: Trust Perceptions of AI Systems across Societal Contexts. Mach. Learn. Knowl. Extr. 2024, 6, 342–366
by Robertas Damaševičius
Mach. Learn. Knowl. Extr. 2024, 6(3), 1667-1669; https://doi.org/10.3390/make6030081 - 22 Jul 2024
Cited by 1 | Viewed by 588
Abstract
The referenced article [...] Full article
(This article belongs to the Section Learning)
16 pages, 2887 KiB  
Article
Global and Local Interpretable Machine Learning Allow Early Prediction of Unscheduled Hospital Readmission
by Rafael Ruiz de San Martín, Catalina Morales-Hernández, Carmen Barberá, Carlos Martínez-Cortés, Antonio Jesús Banegas-Luna, Francisco José Segura-Méndez, Horacio Pérez-Sánchez, Isabel Morales-Moreno and Juan José Hernández-Morante
Mach. Learn. Knowl. Extr. 2024, 6(3), 1653-1666; https://doi.org/10.3390/make6030080 - 17 Jul 2024
Viewed by 903
Abstract
Nowadays, most of the health expenditure is due to chronic patients who are readmitted several times for their pathologies. Personalized prevention strategies could be developed to improve the management of these patients. The aim of the present work was to develop local predictive [...] Read more.
Nowadays, most of the health expenditure is due to chronic patients who are readmitted several times for their pathologies. Personalized prevention strategies could be developed to improve the management of these patients. The aim of the present work was to develop local predictive models using interpretable machine learning techniques to early identify individual unscheduled hospital readmissions. To do this, a retrospective, case-control study, based on information regarding patient readmission in 2018–2019, was conducted. After curation of the initial dataset (n = 76,210), the final number of participants was n = 29,026. A machine learning analysis was performed following several algorithms using unscheduled hospital readmissions as dependent variable. Local model-agnostic interpretability methods were also performed. We observed a 13% rate of unscheduled hospital readmissions cases. There were statistically significant differences regarding age and days of stay (p < 0.001 in both cases). A logistic regression model revealed chronic therapy (odds ratio: 3.75), diabetes mellitus history (odds ratio: 1.14), and days of stay (odds ratio: 1.02) as relevant factors. Machine learning algorithms yielded better results regarding sensitivity and other metrics. Following, this procedure, days of stay and age were the most important factors to predict unscheduled hospital readmissions. Interestingly, other variables like allergies and adverse drug reaction antecedents were relevant. Individualized prediction models also revealed a high sensitivity. In conclusion, our study identified significant factors influencing unscheduled hospital readmissions, emphasizing the impact of age and length of stay. We introduced a personalized risk model for predicting hospital readmissions with notable accuracy. Future research should include more clinical variables to refine this model further. Full article
Show Figures

Figure 1

20 pages, 4689 KiB  
Article
Extending Multi-Output Methods for Long-Term Aboveground Biomass Time Series Forecasting Using Convolutional Neural Networks
by Efrain Noa-Yarasca, Javier M. Osorio Leyton and Jay P. Angerer
Mach. Learn. Knowl. Extr. 2024, 6(3), 1633-1652; https://doi.org/10.3390/make6030079 - 17 Jul 2024
Viewed by 1245
Abstract
Accurate aboveground vegetation biomass forecasting is essential for livestock management, climate impact assessments, and ecosystem health. While artificial intelligence (AI) techniques have advanced time series forecasting, a research gap in predicting aboveground biomass time series beyond single values persists. This study introduces RECMO [...] Read more.
Accurate aboveground vegetation biomass forecasting is essential for livestock management, climate impact assessments, and ecosystem health. While artificial intelligence (AI) techniques have advanced time series forecasting, a research gap in predicting aboveground biomass time series beyond single values persists. This study introduces RECMO and DirRecMO, two multi-output methods for forecasting aboveground vegetation biomass. Using convolutional neural networks, their efficacy is evaluated across short-, medium-, and long-term horizons on six Kenyan grassland biomass datasets, and compared with that of existing single-output methods (Recursive, Direct, and DirRec) and multi-output methods (MIMO and DIRMO). The results indicate that single-output methods are superior for short-term predictions, while both single-output and multi-output methods exhibit a comparable effectiveness in long-term forecasts. RECMO and DirRecMO outperform established multi-output methods, demonstrating a promising potential for biomass forecasting. This study underscores the significant impact of multi-output size on forecast accuracy, highlighting the need for optimal size adjustments and showcasing the proposed methods’ flexibility in long-term forecasts. Short-term predictions show less significant differences among methods, complicating the identification of the best performer. However, clear distinctions emerge in medium- and long-term forecasts, underscoring the greater importance of method choice for long-term predictions. Moreover, as the forecast horizon extends, errors escalate across all methods, reflecting the challenges of predicting distant future periods. This study suggests advancing hybrid models (e.g., RECMO and DirRecMO) to improve extended horizon forecasting. Future research should enhance adaptability, investigate multi-output impacts, and conduct comparative studies across diverse domains, datasets, and AI algorithms for robust insights. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

14 pages, 3044 KiB  
Article
Examining the Global Patent Landscape of Artificial Intelligence-Driven Solutions for COVID-19
by Fabio Mota, Luiza Amara Maciel Braga, Bernardo Pereira Cabral, Natiele Carla da Silva Ferreira, Cláudio Damasceno Pinto, José Aguiar Coelho and Luiz Anastacio Alves
Mach. Learn. Knowl. Extr. 2024, 6(3), 1619-1632; https://doi.org/10.3390/make6030078 - 16 Jul 2024
Viewed by 1308
Abstract
Artificial Intelligence (AI) technologies have been widely applied to tackle Coronavirus Disease 2019 (COVID-19) challenges, from diagnosis to prevention. Patents are a valuable source for understanding the AI technologies used in the COVID-19 context, allowing the identification of the current technological scenario, fields [...] Read more.
Artificial Intelligence (AI) technologies have been widely applied to tackle Coronavirus Disease 2019 (COVID-19) challenges, from diagnosis to prevention. Patents are a valuable source for understanding the AI technologies used in the COVID-19 context, allowing the identification of the current technological scenario, fields of application, and research, development, and innovation trends. This study aimed to analyze the global patent landscape of AI applications related to COVID-19. To do so, we analyzed AI-related COVID-19 patent metadata collected in the Derwent Innovations Index using systematic review, bibliometrics, and network analysis., Our results show diagnosis as the most frequent application field, followed by prevention. Deep Learning algorithms, such as Convolutional Neural Network (CNN), were predominantly used for diagnosis, while Machine Learning algorithms, such as Support Vector Machine (SVM), were mainly used for prevention. The most frequent International Patent Classification Codes were related to computing arrangements based on specific computational models, information, and communication technology for detecting, monitoring, or modeling epidemics or pandemics, and methods or arrangements for pattern recognition using electronic means. The most central algorithms of the two-mode network were CNN, SVM, and Random Forest (RF), while the most central application fields were diagnosis, prevention, and forecast. The most significant connection between algorithms and application fields occurred between CNN and diagnosis. Our findings contribute to a better understanding of the technological landscape involving AI and COVID-19, and we hope they can inform future research and development’s decision making and planning. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

22 pages, 1933 KiB  
Article
Learning Effective Good Variables from Physical Data
by Giulio Barletta, Giovanni Trezza and Eliodoro Chiavazzo
Mach. Learn. Knowl. Extr. 2024, 6(3), 1597-1618; https://doi.org/10.3390/make6030077 - 12 Jul 2024
Viewed by 902
Abstract
We assume that a sufficiently large database is available, where a physical property of interest and a number of associated ruling primitive variables or observables are stored. We introduce and test two machine learning approaches to discover possible groups or combinations of primitive [...] Read more.
We assume that a sufficiently large database is available, where a physical property of interest and a number of associated ruling primitive variables or observables are stored. We introduce and test two machine learning approaches to discover possible groups or combinations of primitive variables, regardless of data origin, being it numerical or experimental: the first approach is based on regression models, whereas the second on classification models. The variable group (here referred to as the new effective good variable) can be considered as successfully found when the physical property of interest is characterized by the following effective invariant behavior: in the first method, invariance of the group implies invariance of the property up to a given accuracy; in the other method, upon partition of the physical property values into two or more classes, invariance of the group implies invariance of the class. For the sake of illustration, the two methods are successfully applied to two popular empirical correlations describing the convective heat transfer phenomenon and to the Newton’s law of universal gravitation. Full article
Show Figures

Graphical abstract

18 pages, 31818 KiB  
Article
Deep Learning-Powered Optical Microscopy for Steel Research
by Šárka Mikmeková, Martin Zouhar, Jan Čermák, Ondřej Ambrož, Patrik Jozefovič, Ivo Konvalina, Eliška Materna Mikmeková and Jiří Materna
Mach. Learn. Knowl. Extr. 2024, 6(3), 1579-1596; https://doi.org/10.3390/make6030076 - 11 Jul 2024
Viewed by 1053
Abstract
The success of machine learning (ML) models in object or pattern recognition naturally leads to ML being employed in the classification of the microstructure of steel surfaces. Light optical microscopy (LOM) is the traditional imaging process in this field. However, the increasing use [...] Read more.
The success of machine learning (ML) models in object or pattern recognition naturally leads to ML being employed in the classification of the microstructure of steel surfaces. Light optical microscopy (LOM) is the traditional imaging process in this field. However, the increasing use of ML to extract or relate more aspects of the aforementioned materials and the limitations of LOM motivated us to provide an improvement to the established image acquisition process. In essence, we perform style transfer from LOM to scanning electron microscopy (SEM) combined with “intelligent” upscaling. This is achieved by employing an ML model trained on a multimodal dataset to generate an SEM-like image from the corresponding LOM image. This transformation, in our opinion, which is corroborated by a detailed analysis of the source, target and prediction, successfully pushes the limits of LOM in the case of steel surfaces. The expected consequence is the improvement of the precise characterization of advanced multiphase steels’ structure based on these transformed LOM images. Full article
(This article belongs to the Section Learning)
Show Figures

Graphical abstract

Previous Issue
Next Issue
Back to TopTop