Predicting an Optimal Medication/Prescription Regimen for Patient Discordant Chronic Comorbidities Using Multi-Output Models
Abstract
:1. Introduction
- Enhanced Decision-Making—this research strives to empower healthcare professionals to tailor medication regimens more effectively for patients with DCCs, thereby enabling them to make more informed and precise decisions.
- Empowering Patients—the study aspires to provide patients with a deeper understanding of alternative medication options, allowing them to actively participate in their healthcare decisions and receive more personalized treatment plans.
- Facilitating Informed Discussions—patients equipped with knowledge from this study will be better equipped to engage in meaningful discussions with their care teams regarding medication options. This in turn, leads to improved treatment outcomes.
2. Related Works
2.1. Recommender and Decision Support Systems
2.2. Machine Learning Tools and Algorithms for Healthcare
3. Proposed Framework
3.1. Study Design
3.1.1. Participant Selection
3.1.2. Participant and Data Screening
- Are there instances in the responses that appeared unrelated to the provided question?
- Do the responses contain nonsensical phrases or incoherent language?
- Is there any indication from the responses that suggests participants might have encountered difficulties in comprehending the questions?
3.1.3. Participant Data Quality and Cleaning
3.1.4. Feature Selection
3.2. Exploring the Potential Machine Learning Algorithms
3.2.1. Technical Backgrounds of Machine Learning Algorithms Adapted in This Study
- Random Forest: Random forest is a widely used ensemble learning algorithm known for its ability to create robust and accurate models. Operating as an ensemble of decision trees, each trained on a random subset of data and features, it combines their predictions for enhanced accuracy and robustness [31]. Random forest has been used in medical recommender systems for disease diagnosis, medication recommendations [32], treatment planning [33], patient outcome prediction [34], and drug–drug interaction prediction [32,35,36]. Its capability to handle heterogeneous medical data and generate interpretable results makes it a valuable tool for developing effective medical recommender systems.
- K-Nearest Neighbors (KNN) K-nearest neighbors (KNN) is a supervised machine learning algorithm used for classification and regression tasks [37]. It operates by assessing the similarity of data points in a multi-dimensional space and making predictions based on the majority class of its k-nearest neighbors. KNN has been widely applied for type 2 diabetes diagnosis [38], medication recommendations [32], personalized treatment planning [33], and drug interaction detection [32,35,36]. Such applications use similarities in patient data to enhance healthcare decision-making and outcomes. Thus, proper selection of distance metrics, as well as addressing issues such as feature selection and missing data, are crucial when using such applications.
- AdaBoost: AdaBoost, short for adaptive boosting, is an ensemble learning method developed to enhance the performance of weak learners [39]. It operates by assigning weights to training instances, giving more weight to misclassified examples in each iteration. Sequential iterations build an ensemble of weak classifiers, with each subsequent classifier correcting the errors of its predecessors. Classifier weighting is based on accuracy, and a final ensemble classifier is formed by combining the weighted base classifiers [39]. AdaBoost has been used in medication and treatment recommendation systems [40] to provide personalized health advice and brain MR image classification and analysis [41,42].
- XGBoost: XGBoost, or extreme gradient boosting, is a highly efficient and versatile machine learning algorithm known for its proficiency in structured and tabular data tasks [43]. It operates on the gradient boosting framework, sequentially building an ensemble of decision trees to correct errors, and includes L1 and L2 regularization terms to prevent overfitting. XGBoost has been applied in type 2 diabetes prediction and diagnosis [44], personalized treatment recommendation, drug interaction prediction [45], and health behavior recommendations [44]. These applications use XGBoost’s structured data handling capabilities and efficiency for improved clinical decision-making and patient outcomes.
3.2.2. Predicting Optimal Medication Combination
3.3. Iteration 1: Base Classification Models
3.3.1. Data Format
3.3.2. Algorithms
- Random Forest: We utilized a random forest classifier and implemented it as our classification model. We instantiated the classifier with 100 decision trees and a maximum depth of 12. We set the random state to 2 for consistent results during model training and then fitted the model. After training the model, we assessed its accuracy on the training set using the ‘score’ method. The accuracy score was calculated by comparing the recommendations made by the trained model to the true labels in the training set. In Figure 3, we can see how tree construction starts from the root node, and at each step, the algorithm selects the best split point based on certain criteria, usually maximizing information gain. The obtained results, in terms of accuracy, were not high.
- K-Nearest Neighbors Classifier: KNN classifier is a non-parametric and instance-based learning method that makes recommendations based on the similarity of the input data to its neighboring data points [47]. We configured the model with the following parameters: ‘algorithm’ set to ‘kd_tree’, ‘leaf_size’ set to 10, ‘metric’ set to Euclidean, ‘n_jobs’ set to 10, ‘n_neighbors’ set to 4, ‘p’ set to 3, and ‘weights’ set to ‘uniform’. We performed the tuning by implementing both Euclidean and Minkowski distance. We achieved the best results using the Euclidean distance paired with uniform weights, 3 neighbors.
- Extreme Gradient Boosting (XGBoost) Classifier: The XGBoost classifier is an implementation of the gradient boosting algorithm, which is known for its effectiveness in various machine learning tasks [48]. We instantiated with specific hyperparameters including 200 estimators, a maximum depth of 4, a learning rate of 0.1, subsampling and feature subsampling rates of 0.2, regularization parameters of 0.01 for both L1 and L2 regularization, a gamma value of 0.05, and a random state of 32. Finally, we used the trained classifier to predict the labels for the testing data, and the predicted labels are stored in the ‘preds’ variable for further analysis or evaluation.
- Adaptive Boost (AdaBoost) Classifier: To instantiate the AdaBoost classifier, we configured the following parameters: ‘n_estimators’ set to 128, ‘learning_rate’ set to 0.001, and ‘random_state’ set to 42. These parameters control the number of weak classifiers to combine, the learning rate of the model, and the random seed for reproducibility. The use of a single output classifier in the initial iteration of the project had its limitations. While this approach is designed to predict or categorize a single target variable, it may not be the optimum solution for problems involving multiple output variables or complex dependencies among the variables. Additionally, the single output classifier may have limited complexity, hindering its ability to capture intricate patterns in the data. The choice of a single output classifier also restricted the modeling of the concatenated target values representing multiple medications. These limitations highlight the need to explore alternative approaches that can better handle multiple output variables and capture the complexity of the problem at hand.
3.4. Iteration 2: Multi-Class Multi-Output Classification Models
3.4.1. Data Format
3.4.2. Algorithms
- Random Forest with Multi-Output Classifier: To enable multi-output functionality, the random forest classifier is wrapped inside a multi-output classifier. This wrapper allows the random forest classifier to handle multiple target variables simultaneously. We updated the classifier to have 100 decision trees and a maximum depth of 12. We set the random state to 2 for reproducibility. We updated the n_jobs parameter to 2, indicating the number of parallel jobs to use for model training. As a result of these modifications, we observed an improvement in the accuracy score. The accuracy score provides a measure of the model’s performance, and the increase in accuracy suggests that the multi-output random forest classifier can better capture the underlying patterns and make more accurate predictions on the given task.
- K-Nearest Neighbor with Multi-Output Classifier: The KNN algorithm works by finding the K-nearest neighbors to a given input data point in the feature space and then averaging their corresponding target variable values to make a prediction. We instantiated a K-nearest neighbors regressor (KNN) with the following configuration: algorithm = ‘kd_tree’, leaf_size = 20, metric = ‘minkowski’, n_jobs = 15, n_neighbors = 2, and weights = ‘uniform’. Then, we wrapped this regressor inside a multi-output classifier, allowing it to handle multiple target variables simultaneously. To assess the accuracy of the regressor on the training set, we used the ’score’ method, which calculates the coefficient of determination (R-squared score). The R-squared score represents the proportion of variance in the target variables that can be explained by the regressor.
- Extreme Gradient Boosting (XGBoost) with Multi-Output Classifier: XGBoost (extreme gradient boosting) is an advanced machine learning algorithm that belongs to the gradient boosting family of models. We instantiated an XGBoost classifier (XGBClassifier) with the provided hyperparameters. We set the number of estimators to 200, the maximum depth of each tree to 2, the learning rate to 0.1, the subsample ratio to 0.5, the column subsampling ratio to 0.5, the L1 regularization term to 0.01, the L2 regularization term to 0.01, the minimum loss reduction required for a split (gamma) to 0.05, and the random state to 32 for reproducibility. To enable multi-output functionality, we have wrapped the XGBoost classifier inside a multi-output classifier. This approach leveraged the boosting technique and ensemble learning, enabling us to make accurate predictions for multiple target variables.
- Adaptive Boosting (AdaBoost) with Multi-Output Classifier: AdaBoostClassifier is a powerful algorithm that can improve the performance of classification models by combining multiple weak classifiers into a strong ensemble classifier. We instantiated an AdaBoost classifier with specific hyperparameters, including the number of estimators set to 64, the learning rate set to 0.01, and the random state set to 128. To enable multi-output functionality, we wrap the AdaBoostClassifier inside a multi-output classifier.
3.5. Iteration 3: Classification Model Optimization
4. Results and Discussion
4.1. Analysis of Iterations 1, 2, and 3
4.1.1. Iteration 1
4.1.2. Iteration 2
4.1.3. Iteration 3
4.2. Assessing Algorithms’ Performance
4.3. Benchmarking Suite
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Zheng, D.D.; Loewenstein, D.A.; Christ, S.L.; Feaster, D.J.; Lam, B.L.; McCollister, K.E.; Curiel-Cid, R.E.; Lee, D.J. Multimorbidity patterns and their relationship to mortality in the US older adult population. PLoS ONE 2021, 16, e0245053. [Google Scholar] [CrossRef] [PubMed]
- Lu, H.; Uddin, S. A disease network-based recommender system framework for predictive risk modelling of chronic diseases and their comorbidities. Appl. Intell. 2022, 52, 10330–10340. [Google Scholar] [CrossRef]
- Ongwere, T.; Cantor, G.; Martin, S.R.; Shih, P.C.; Clawson, J.; Connelly, K. Design hotspots for care of discordant chronic comorbidities: Patients’ perspectives. In Proceedings of the 10th Nordic Conference on Human-Computer Interaction, Oslo, Norway, 29 September–3 October 2018; pp. 571–583. [Google Scholar]
- Ongwere, T.; Cantor, G.S.; Clawson, J.; Shih, P.C.; Connelly, K. Design and care for discordant chronic comorbidities: A comparison of healthcare providers’ perspectives. In Proceedings of the 14th EAI International Conference on Pervasive Computing Technologies for Healthcare, online, 6–8 October 2020; pp. 133–145. [Google Scholar]
- Ongwere, T.; Stolterman, E.; Shih, P.C.; James, C.; Connelly, K. Translating a DC 3 Model into a Conceptual Tool (DCCs Ecosystem): A Case Study with a Design Team. In Proceedings of the International Conference on Pervasive Computing Technologies for Healthcare, Virtual, 6–8 December 2021; Springer International Publishing: Cham, Switzerland; pp. 381–397. [Google Scholar]
- Ongwere, T.; Stolterman, E.; Shih, P.C.; James, C.; Connelly, K. Design for Discordant Chronic Comorbidities (DCCs): A DC3 Model. In Proceedings of the International Conference on Pervasive Computing Technologies for Healthcare, Virtual Event, 6–8 December 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 346–362. [Google Scholar]
- Kilic, A.; Dochtermann, D.; Padman, R.; Miller, J.K.; Dubrawski, A. Using machine learning to improve risk prediction in durable left ventricular assist devices. PLoS ONE 2021, 16, e0247866. [Google Scholar] [CrossRef]
- Nuryani, N.; Ling, S.S.H.; Nguyen, H.T. Electrocardiographic signals and swarm-based support vector machine for hypoglycemia detection. Ann. Biomed. Eng. 2012, 40, 934–945. [Google Scholar] [CrossRef]
- Mitchell, E.G.; Heitkemper, E.M.; Burgermaster, M.; Levine, M.E.; Miao, Y.; Hwang, M.L.; Desai, P.M.; Cassells, A.; Tobin, J.N.; Tabak, E.G.; et al. From reflection to action: Combining machine learning with expert knowledge for nutrition goal recommendations. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; pp. 1–17. [Google Scholar]
- Kavakiotis, I.; Tsave, O.; Salifoglou, A.; Maglaveras, N.; Vlahavas, I.; Chouvarda, I. Machine learning and data mining methods in diabetes research. Comput. Struct. Biotechnol. J. 2017, 15, 104–116. [Google Scholar] [CrossRef] [PubMed]
- Shah, A.M.; Muhammad, W.; Lee, K.; Naqvi, R.A. Examining different factors in web-based patients’ decision-making process: Systematic review on digital platforms for clinical decision support system. Int. J. Environ. Res. Public Health 2021, 18, 11226. [Google Scholar] [CrossRef]
- Lysaght, T.; Lim, H.Y.; Xafis, V.; Ngiam, K.Y. AI-assisted decision-making in healthcare. Asian Bioeth. Rev. 2019, 11, 299–314. [Google Scholar] [CrossRef]
- Hu, Y.-J.; Ku, T.; Jan, R.; Wang, K.; Tseng, Y.; Yang, S. Decision tree-based learning to predict patient controlled analgesia consumption and readjustment. BMC Med. Inform. Decis. Mak. 2012, 12, 131. [Google Scholar] [CrossRef]
- Rush, K.L.; Hatt, L.; Janke, R.; Burton, L.; Ferrier, M.; Tetrault, M. The efficacy of telehealth delivered educational approaches for patients with chronic diseases: A systematic review. Patient Educ. Couns. 2018, 101, 1310–1321. [Google Scholar] [CrossRef]
- Milette, K.; Thombs, B.D.; Maiorino, K.; Nielson, W.R.; Körner, A.; Peláez, S. Challenges and strategies for coping with scleroderma: Implications for a scleroderma-specific self-management program. Disabil. Rehabil. 2019, 41, 2506–2515. [Google Scholar] [CrossRef]
- Palmer, R.; Enderby, P.; Paterson, G. Using computers to enable self-management of aphasia therapy exercises for word finding: The patient and carer perspective. Int. J. Lang. Commun. Disord. 2013, 48, 508–521. [Google Scholar] [CrossRef] [PubMed]
- Wang, Z.; Huang, H.; Cui, L.; Chen, J.; An, J.; Duan, H.; Ge, H.; Deng, N. Using natural language processing techniques to provide personalized educational materials for chronic disease patients in China: Development and assessment of a knowledge-based health recommender system. JMIR Med. Inform. 2020, 8, e17642. [Google Scholar] [CrossRef] [PubMed]
- Woldaregay, A.Z.; Årsand, E.; Botsis, T.; Albers, D.; Mamykina, L.; Hartvigsen, G. Data-driven blood glucose pattern classification and anomalies detection: Machine-learning applications in type 1 diabetes. J. Med. Internet Res. 2019, 21, e11030. [Google Scholar] [CrossRef]
- Bzdok, D.; Krzywinski, M.; Altman, N. Machine learning: Supervised methods. Nat. Methods 2018, 15, 5. [Google Scholar] [CrossRef] [PubMed]
- Singh, S.; Gupta, A.; Baraheem, S.S.; Nguyen, T.V. Multi-Output Career Prediction: Dataset, Method, and Benchmark Suite. In Proceedings of the 2023 57th Annual Conference on Information Sciences and Systems (CISS), Baltimore, MD, USA, 22–24 March 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
- Liu, K.; Dou, Y.; Zhao, Y.; Ding, X.; Hu, X.; Zhang, R.; Ding, K.; Chen, C.; Peng, H.; Shu, K.; et al. Bond: Benchmarking unsupervised outlier node detection on static attributed graphs. Adv. Neural Inf. Process. Syst. 2022, 35, 27021–27035. [Google Scholar]
- Kumar, P.; Garg, S.; Garg, A. Assessment of anxiety, depression and stress using machine learning models. Procedia Comput. Sci. 2020, 171, 1989–1998. [Google Scholar] [CrossRef]
- Seckler, H.; Szwabiński, J.; Metzler, R. Machine-Learning Solutions for the Analysis of Single-Particle Diffusion Trajectories. J. Phys. Chem. Lett. 2023, 14, 7910–7923. [Google Scholar] [CrossRef]
- Zand, J. Multimodal Probabilistic Reasoning for Prediction and Coordination Problems in Machine Learning. Ph.D. Thesis, University of Oxford, Oxford, UK, 2021. [Google Scholar]
- Challen, R.; Denny, J.; Pitt, M.; Gompels, L.; Edwards, T.; Tsaneva-Atanasova, K. Artificial intelligence, bias and clinical safety. BMJ Qual. Saf. 2019, 28, 231–237. [Google Scholar] [CrossRef]
- Angelopoulos, A.N.; Bates, S. A gentle introduction to conformal prediction and distribution-free uncertainty quantification. arXiv 2021, arXiv:2107.07511. [Google Scholar]
- Hunt, N.C.; Scheetz, A.M. Using MTurk to distribute a survey or experiment: Methodological considerations. J. Inf. Syst. 2019, 33, 43–65. [Google Scholar] [CrossRef]
- Ivanisevic, J.; Benton, H.P.; Rinehart, D.; Epstein, A.; Kurczy, M.E.; Boska, M.D.; Gendelman, H.E.; Siuzdak, G. An interactive cluster heat map to visualize and explore multidimensional metabolomic data. Metabolomics 2015, 11, 1029–1034. [Google Scholar] [CrossRef]
- Ghosh, M.; Raihan, M.M.S.; Raihan, M.; Akter, L.; Bairagi, A.K.; Alshamrani, S.S.; Masud, M. A Comparative Analysis of Machine Learning Algorithms to Predict Liver Disease. Intell. Autom. Soft Comput. 2021, 30, 917–928. [Google Scholar] [CrossRef]
- Prastyo, P.H.; Paramartha, I.G.Y.; Pakpahan, M.S.M.; Ardiyanto, I. Predicting breast cancer: A comparative analysis of machine learning algorithms. Proceeding Int. Conf. Sci. Eng. 2020, 3, 455–459. [Google Scholar] [CrossRef]
- Parmar, A.; Katariya, R.; Patel, V. A review on random forest: An ensemble classifier. In Proceedings of the International Conference on Intelligent Data Communication Technologies and Internet of Things (ICICI) 2018, Coimbatore, India, 7–8 August 2018; Springer International Publishing: Cham, Switzerland, 2019; pp. 758–763. [Google Scholar]
- Mathew, R.B.; Varghese, S.; Joy, S.E.; Alex, S.S. Chatbot for disease prediction and treatment recommendation using machine learning. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 851–856. [Google Scholar]
- Assegie, T.A.; Suresh, T.; Purushothaman, R.; Ganesan, S.; Kumar, N.K. Early Prediction of Gestational Diabetes with Parameter-Tuned K-Nearest Neighbor Classifier. J. Robot. Control (JRC) 2023, 4, 452–457. [Google Scholar]
- Pettit, R.W.; Fullem, R.; Cheng, C.; Amos, C.I. Artificial intelligence, machine learning, and deep learning for clinical outcome prediction. Emerg. Top. Life Sci. 2021, 5, 729–745. [Google Scholar] [CrossRef] [PubMed]
- Vo, T.H.; Nguyen, N.T.K.; Le, N.Q.K. Improved prediction of drug-drug interactions using ensemble deep neural networks. Med. Drug Discov. 2023, 17, 100149. [Google Scholar] [CrossRef]
- Guo, C.; Zhang, J.; Liu, Y.; Xie, Y.; Han, Z.; Yu, J. Recursion enhanced random forest with an improved linear model (RERF-ILM) for heart disease detection on the internet of medical things platform. IEEE Access 2020, 8, 59247–59256. [Google Scholar] [CrossRef]
- Siddalingappa, R.; Kanagaraj, S. K-nearest-neighbor algorithm to predict the survival time and classification of various stages of oral cancer: A machine learning approach. F1000Research 2022, 11, 70. [Google Scholar] [CrossRef]
- Salem, H.; Shams, M.Y.; Elzeki, O.M.; Abd Elfattah, M.; Al-Amri, J.F.; Elnazer, S. Fine-tuning fuzzy KNN classifier based on uncertainty membership for the medical diagnosis of diabetes. Appl. Sci. 2022, 12, 950. [Google Scholar] [CrossRef]
- Walker, K.W.; Jiang, Z. Application of adaptive boosting (AdaBoost) in demand-driven acquisition (DDA) prediction: A machine-learning approach. J. Acad. Librariansh. 2019, 45, 203–212. [Google Scholar] [CrossRef]
- Hatwell, J.; Gaber, M.M.; Atif Azad, R.M. Ada-WHIPS: Explaining AdaBoost classification with applications in the health sciences. BMC Med. Inform. Decis. Mak. 2020, 20, 250. [Google Scholar] [CrossRef] [PubMed]
- Nayak, D.R.; Dash, R.; Majhi, B. Brain MR image classification using two-dimensional discrete wavelet transform and AdaBoost with random forests. Neurocomputing 2016, 177, 188–197. [Google Scholar] [CrossRef]
- Islam, M.S.; Hasan, M.M.; Wang, X.; Germack, H.D.; Noor-E-Alam, M. A systematic review on healthcare analytics: Application and theoretical perspective of data mining. Healthcare 2018, 6, 54. [Google Scholar] [CrossRef]
- Jiang, D.; Wu, Z.; Hsieh, C.Y.; Chen, G.; Liao, B.; Wang, Z.; Shen, C.; Cao, D.; Wu, J.; Hou, T. Could graph neural networks learn better molecular representation for drug discovery? A comparison study of descriptor-based and graph-based models. J. Cheminformatics 2021, 13, 12. [Google Scholar] [CrossRef] [PubMed]
- Liu, Q.; Zhang, M.; He, Y.; Zhang, L.; Zou, J.; Yan, Y.; Guo, Y. Predicting the risk of incident type 2 diabetes mellitus in chinese elderly using machine learning techniques. J. Pers. Med. 2022, 12, 905. [Google Scholar] [CrossRef]
- Shtar, G.; Rokach, L.; Shapira, B. Detecting drug-drug interactions using artificial neural networks and classic graph similarity measures. PLoS ONE 2019, 14, e0219796. [Google Scholar] [CrossRef]
- Kramer, O.; Kramer, O. Scikit-learn. In Machine Learning for Evolution Strategies; Springer: Cham, Switzerland, 2016; pp. 45–53. [Google Scholar]
- Adeniyi, D.A.; Wei, Z.; Yongquan, Y. Automated web usage data mining and recommendation system using K-Nearest Neighbor (KNN) classification method. Appl. Comput. Informatics 2016, 12, 90–108. [Google Scholar] [CrossRef]
- González, S.; García, S.; Ser, J.D.; Rokach, L.; Herrera, F. A practical tutorial on bagging and boosting based ensembles for machine learning: Algorithms, software tools, performance study, practical perspectives and opportunities. Inf. Fusion 2020, 64, 205–237. [Google Scholar] [CrossRef]
Category | Attribute | Number of Participant |
---|---|---|
Experience | Knowledgeable | 162 |
Very knowledgeable | 64 | |
Diseases Treated | Type 2 diabetes (T2D) | 226 |
T2D & one other disease | 62 | |
T2D & two other diseases | 109 | |
T2D & more than two diseases | 55 | |
Primary Specialization | Family medicine | 112 |
Endocrinologist | 14 | |
General internal medicine | 63 | |
General pediatrics | 32 | |
Gynecology | 8 | |
Nephrologist | 12 | |
Other | 23 |
Diabetes | Arthritis | Depression | Cost | Weight | Interaction | Medicine |
---|---|---|---|---|---|---|
1 | 0 | 0 | 1 | 1 | 0 | Metformin |
1 | 0 | 0 | 1 | 1 | 0 | Victoza |
1 | 0 | 0 | 1 | 1 | 0 | Metformin |
1 | 0 | 0 | 1 | 1 | 0 | Victoza |
1 | 0 | 0 | 1 | 1 | 0 | Victoza |
1 | 0 | 0 | 1 | 1 | 0 | Metformin |
1 | 0 | 0 | 1 | 0 | 1 | Victoza |
1 | 0 | 0 | 1 | 0 | 1 | Metformin |
1 | 0 | 0 | 1 | 0 | 1 | Victoza |
1 | 0 | 0 | 1 | 0 | 1 | Metformin |
1 | 0 | 0 | 1 | 0 | 1 | Metformin |
1 | 0 | 0 | 1 | 0 | 1 | Metformin |
Diabetes | Arthritis | Depression | Cost | Weight | Interaction | Medicine |
---|---|---|---|---|---|---|
1 | 0 | 0 | 1 | 1 | 0 | Metformin |
1 | 0 | 0 | 1 | 1 | 0 | Victoza |
1 | 0 | 0 | 1 | 1 | 0 | Metformin |
1 | 0 | 0 | 1 | 1 | 0 | Victoza |
1 | 0 | 0 | 1 | 1 | 0 | Victoza |
1 | 0 | 0 | 1 | 1 | 0 | Metformin |
- | - | - | - | - | - | - |
- | - | - | - | - | - | - |
1 | 1 | 1 | 1 | 1 | 1 | V + H + Es |
1 | 1 | 1 | 1 | 1 | 1 | M + H + Es |
1 | 1 | 1 | 1 | 1 | 1 | Em + I + S |
1 | 1 | 1 | 1 | 1 | 1 | Em + H + Es |
1 | 1 | 1 | 1 | 1 | 1 | M + I + Es |
Classification | Number of Targets | Targets Cardinality |
---|---|---|
Multiclass | 1 | >2 |
Multi-label | >1 | 2 (0 or 1) |
Multiclass and multi-output | >1 | >2 |
Diabetes | Arthritis | Depression | Cost | Weight | Interaction | Med1 | Med2 | Med3 |
---|---|---|---|---|---|---|---|---|
1 | 0 | 0 | 1 | 1 | 0 | Metformin | None | None |
1 | 0 | 0 | 1 | 1 | 0 | Victoza | None | None |
1 | 0 | 0 | 1 | 1 | 0 | Metformin | None | None |
1 | 0 | 0 | 1 | 1 | 0 | Victoza | None | None |
1 | 0 | 0 | 1 | 1 | 0 | Victoza | None | None |
1 | 0 | 0 | 1 | 1 | 0 | Metformin | None | None |
1 | 1 | 1 | 1 | 1 | 1 | V | H | Es |
1 | 1 | 1 | 1 | 1 | 1 | M | H | Es |
1 | 1 | 1 | 1 | 1 | 1 | Em | I | S |
1 | 1 | 1 | 1 | 1 | 1 | Em | H | Es |
1 | 1 | 1 | 1 | 1 | 1 | M | I | Es |
Algorithm/Model | Iteration 1 | Iteration 2 | Iteration 3 |
---|---|---|---|
Random Forest | 65.5% | 83.5% | 93.3% |
KNN | 60.2% | 67.4% | 78.5% |
AdaBoost | 35.8% | 36.4% | 67.3% |
XGBoost | 53.8% | 62.4% | 76.4% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sharma, I.P.; Nguyen, T.V.; Singh, S.A.; Ongwere, T. Predicting an Optimal Medication/Prescription Regimen for Patient Discordant Chronic Comorbidities Using Multi-Output Models. Information 2024, 15, 31. https://doi.org/10.3390/info15010031
Sharma IP, Nguyen TV, Singh SA, Ongwere T. Predicting an Optimal Medication/Prescription Regimen for Patient Discordant Chronic Comorbidities Using Multi-Output Models. Information. 2024; 15(1):31. https://doi.org/10.3390/info15010031
Chicago/Turabian StyleSharma, Ichchha Pradeep, Tam V. Nguyen, Shruti Ajay Singh, and Tom Ongwere. 2024. "Predicting an Optimal Medication/Prescription Regimen for Patient Discordant Chronic Comorbidities Using Multi-Output Models" Information 15, no. 1: 31. https://doi.org/10.3390/info15010031
APA StyleSharma, I. P., Nguyen, T. V., Singh, S. A., & Ongwere, T. (2024). Predicting an Optimal Medication/Prescription Regimen for Patient Discordant Chronic Comorbidities Using Multi-Output Models. Information, 15(1), 31. https://doi.org/10.3390/info15010031