Next Article in Journal
Small Ultrasound-Based Corrosion Sensor for Intraday Corrosion Rate Estimation
Previous Article in Journal
Dynamic Design of a Quad-Stable Piezoelectric Energy Harvester via Bifurcation Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Blood Glucose Concentration after Short-Acting Insulin Injection Using Discontinuous Injection Records

1
School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China
2
Key Laboratory of Trustworthy Distributed Computing and Service, Ministry of Education, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(21), 8454; https://doi.org/10.3390/s22218454
Submission received: 19 September 2022 / Revised: 25 October 2022 / Accepted: 31 October 2022 / Published: 3 November 2022
(This article belongs to the Section Biosensors)

Abstract

:
Diabetes is an increasingly common disease that poses an immense challenge to public health. Hyperglycemia is also a common complication in clinical patients in the intensive care unit, increasing the rate of infection and mortality. The accurate and real-time prediction of blood glucose concentrations after each short-acting insulin injection has great clinical significance and is the basis of all intelligent blood glucose control systems. Most previous prediction methods require long-term continuous blood glucose records from specific patients to train the prediction models, resulting in these methods not being used in clinical practice. In this study, we construct 13 deep neural networks with different architectures to atomically predict blood glucose concentrations after arbitrary independent insulin injections without requiring continuous historical records of any patient. Using our proposed models, the best root mean square error of the prediction results reaches 15.82 mg/dL, and 99.5% of the predictions are clinically acceptable, which is more accurate than previously proposed blood glucose prediction methods. Through the re-validation of the models, we demonstrate the clinical practicability and universal accuracy of our proposed prediction method.

1. Introduction

Diabetes mellitus is an increasingly common chronic disease, which is mainly characterized by hyperglycemia and often induces multiple complications [1]. Hyperglycemia is also commonly encountered in the intensive care unit (ICU). Stress hyperglycemia (SH) often occurs in critically ill patients, even without a past diagnosis of diabetes [2,3,4]. Numerous observational studies demonstrate that elevated blood glucose is significantly associated with increased ICU mortality [5,6,7,8,9,10], and the aggressive correction of hyperglycemia with exogenous insulin can reduce the morbidity and mortality of ICU patients [11,12,13,14,15,16]. Continuously tracking blood glucose values is the best way to understand the effects of insulin on patients’ blood glucose concentrations. Point sample test and continuous glucose monitoring (CGM) are currently the most commonly used blood glucose detecting methods [17]. The point sample test requires patients to draw blood with a lancet several times daily. Such discrete tests are prone to miss hyperglycemic or hypoglycemic events. The CGM system uses a microsensor inserted beneath the skin and held for a period to assess glucose levels in tissue fluids. However, limitations in detection accuracy and sensor lifetime make it still necessary for patients to order point sample tests for continuous calibration. The inconvenience, discontinuity, and limited accuracy of these two blood glucose detection methods have brought great hidden dangers to patients in clinical practice. Therefore, accurately predicting the blood glucose concentrations after exogenous insulin injections is of great significance for maintaining the stability of blood glucose, preventing the side effect of hypoglycemia, and reducing the mortality of patients in clinical practice. In addition, the accurate prediction of insulin efficacy is also a necessary basis for the development of automated medical technologies, such as intelligent blood-glucose-monitoring devices, automatic insulin-delivery devices, and intelligent patient-monitoring systems.
The deep neural network is an important branch of artificial intelligence. Related concepts and algorithms have been proposed since the 1960s [18,19,20,21,22,23]. With the improvement of algorithms and the development of computer hardware, deep neural networks have flourished and been widely used in the past 20 years. In 2006, Hinton proposed deep belief nets (DBN) [24], which used a greedy unsupervised training method to solve the problem of handwritten digit recognition and achieved good results. In 2011, researchers from Microsoft proposed a pre-trained deep neural network hidden Markov model hybrid architecture (DNN-HMM) [25] that greatly reduced the error rate of speech recognition. Alex Krizhevsky proposed AlexNet in 2012 [26], reducing the Top5 error rate of ImageNet classification from 26% to 15%. Google proposed FaceNet in 2014 [27], achieving 99.63% test accuracy on the Labeled Faces in the Wild (LFW) dataset. Microsoft built a 152-layer residual neural network(ResNet) [28] in 2015, reducing the classification error rate of ImageNet to 3.57%. In 2016, AlphaGo [29], which integrates Monte Carlo simulation and value and policy networks, defeated the world Go champion. There are great expectations on how this technology can be applied to healthcare and clinical practice [30,31,32]. In recent years, the deep neural network has also shown its feasibility in many medical applications, such as automatic diagnosis [33,34,35], medical image interpretation and processing [36,37,38,39], risk prediction [36,40,41], and toxicity prediction [42,43].
In the past five years, studies on predicting blood glucose concentrations have also been reported [44,45,46,47]. It has been observed that convolutional neural network (CNN), recursive neural network (RNN), and the variations of these network models are commonly used in prediction studies, and the prediction results with low root mean square error (RMSE) can be obtained. A study by K. Li et al. [48] presented a deep learning algorithm for glucose prediction using a multi-layer convolutional recurrent neural network (CRNN). The model was primarily trained on a simulated dataset of 10 cases generated from the UVA/Padova simulator and a clinical dataset from 10 real patient cases, each containing continuous glucose readings, insulin bolus injections, and carbohydrate data for six months. Their predictive model had leading accuracy for real patient cases (RMSE = 21.07 mg/dL for 30 min and RMSE = 33.27 mg/dL for 60 min on average). S. Mirshekarian et al. [49] constructed a recursive neural network model that uses long short-term memory (LSTM) units to train a physiological model of blood glucose. The LSTM network was trained and evaluated separately on the dataset of five different patients, containing approximately 400 days of records on each patient’s blood glucose values, insulin doses, and dietary intake. The networks trained separately for each specific patient could achieve competitive predicting accuracy (RMSE = 21.4 mg/dL for 30 min, RMSE = 38.0 mg/dL for 60 min on average). H.V. Dudukcu et al. [50] created a consolidated model using gated recurrence unit (GRU) to predict blood glucose with the OhioT1DM dataset. The RMSE values obtained in experimental studies suggested that the consolidated model trained using all patients’ data provides better results than the individual model trained specifically for each patient.
Although there have been many related studies using deep neural networks to predict blood glucose concentrations (shown in Appendix A Table A1), most of these studies have limitations in the scope of application. Most of the proposed deep neural networks are trained using continuous recording data from diabetic patients [45,46,47,48,49,50,51,52], including the blood glucose values recorded by CGM and other recordings of some correlative events. Since the network training requires continuous records lasting tens or even hundreds of days, the acquisition of the datasets is very difficult. Some studies [48,49,52] also require patients to record multiple daily information manually for hundreds of consecutive days, such as insulin injection events, dietary events, exercise events, and psychological stress, resulting in many missing and erroneous values in the dataset with fewer usable data. For the above reasons, the networks are always trained with datasets only from a small number of clinical patients [45,47,48,49,50,51,52] (mostly below 30 patients) or even with simulated dataset from virtual patients [44,47,48]. The networks trained using datasets from a few patients can only simulate the physiological environment of a few specific people, and have very limited accuracy in predicting for patients that are not involved in the training data, which leads to the difficulty of predicting blood glucose after insulin injection with deep neural networks, and cannot be widely applied in clinical practice for the patients without continuous data records.
In this study, we try to explore whether the three commonly used deep learning regression models, including fully connected neural network (FCN), progressive neural network (CNN), and long short-term memory network (LSTM), could predict the effect of short-term insulin doses on blood glucose concentrations without the need for continuous past records. We extract independent discontinuous short-acting insulin-injection events occurring in the ICU from the Medical Information Mart for Intensive Care-IV (MIMIC-IV) database [53], and match each insulin injection record to the glucose values that triggered and related to the insulin treatment. Moreover, we also integrate some clinical characteristics that have been proven to be associated with insulin efficacy as predictive references, including the patient’s age, gender, weight, ethnicity, blood pressure, and renal function indexes [54,55,56,57,58,59,60]. Using these atomic records of discrete insulin injection events to train deep neural networks, and taking these demographic characteristics and basic examination results of ICU patients as the network input, we construct various deep neural networks with different architectures to predict the blood glucose concentrations of ICU patients about 4 h after short-acting insulin injections. We re-validate the predictive accuracy of the network models using the K-fold test and the Chi-square test to demonstrate the universal accuracy of our proposed networks.

2. Proposed Method

2.1. Dataset Extraction and Integration

We extracted our dataset from the MIMIC-IV database version 1.0. MIMIC-IV is an extensive, freely available database containing real medical information during hospital stays and ICU stays for patients admitted to a tertiary academic medical center in Boston, MA, USA. The data include the patient demographics, bedside monitoring of vitals, administration of medications, results of laboratory measurements, as well as diagnoses, procedures, and prescriptions for billing.
The primary objectives of our data extraction are as follows:
  • Extract discontinuous short-acting insulin bolus injections occurring in the ICU from the electronic medical records in the MIMIC-IV database.
  • Match the blood glucose values before insulin injection and around the peak time of insulin efficacy for each insulin injection event.
  • Query and match the relevant characteristic values for each insulin injection event as the prediction basis.
  • Remove all missing values and possible erroneous values.
For that purpose, we formulated data filtering and integration rules based on data analysis, clinician recommendations, and physiological and pharmacological standards. The overall process of data extraction is shown in Figure 1.
Firstly, we extracted the records of short-acting insulin injection events (containing regular insulin, Humalog, and Novolog) from the table mimic_icu.INPUTEVENTS. After removing erroneous and incomplete data, we obtained 234,358 bolus injection records of short-acting insulin from 24,750 patients and 31,920 ICU stays (step A in Figure 1). Then we extracted the blood glucose values measured by laboratory chemistry analyzers from table mimic_hosp.LABEVENTS, and the blood glucose values measured by bedside fingerstick glucometers from table mimic_icu. CHARTEVENTS (4,450,933 in total). Posteriorly, glucose data from patients who did not receive any insulin injection were excluded (step B in Figure 1). After removing duplicates, errors, and outliers, we obtained a total of 1,298,679 blood glucose test results (step C in Figure 1). We matched the occurrence time of all blood glucose readings and insulin injection events to the ICU admission records from table mimic_icu.ICUSTAYS, and all events that did not occur in the ICU were removed (step D in Figure 1).
Secondly, to predict the effect of insulin doses on blood glucose concentrations, we tried to match patients’ blood glucose values before and after insulin injection (glc_before and glc_after) for each insulin injection event according to time (step E in Figure 1). We calculated the time interval between each insulin injection and the closest blood glucose detection before and after the injection event (timediff_before and timediff_after).
Figure 2a shows the density distribution of timediff_before. A total of 89.36% of the insulin injection events had a blood glucose record within 90 min before the insulin injection. Therefore, we selected the blood glucose value closest to the insulin injection event as glc_before from all blood glucose records within 90 min before and 5 min after the insulin injection. If there were no blood glucose records within 90 min before insulin injection, we would extend the backward search time and select the closest blood glucose results within 10 min after insulin injection as glc_before.
Figure 2b shows the density distribution of timediff_after. The mean value of timediff_after was 3 h 43 min, and the median was 3 h 59 min. A total of 92.29% of insulin injection events have blood glucose records within 6 h after the insulin injection. Considering that the peak effect of short-acting insulin is about 4 h after injection, we selected the blood glucose value closest to 4 h and within 6 h after insulin injection as glc_after. Finally, we obtained 174,280 insulin injection events that were successfully matched to glc_before and glc_after.
In addition, to further improve the prediction accuracy of the insulin effect, we extracted many related characteristics from the MIMIC-IV database (step F in Figure 1). We obtained the patients’ weight and nutritional intake from table mimic_icu.INPUTEVENTS, patients’ ethnicity from table mimic_core.ADMISSIONS, patients’ gender and age from table mimic_core.PATIENTS, the average systolic and diastolic blood pressure (sbp and dbp) within two hours around insulin injection from table mimic_icu.CHARTEVENTS, and the values of creatinine and blood urea nitrogen (bun) from table mimic_hosp.LABEVENTS. Since the total amount of patient sugar intake could not be accurately measured according to the records from the MIMIC-IV database, we removed all the insulin injection records with nutrient solutions ingesting between the two blood glucose measurements before and after insulin injection. We also removed all the insulin injection records with glc_before less than glc_after to avoid the unknown nutrient intake affecting our prediction of the insulin efficacy. Finally, after removing all records with missing values and outliers, we obtained 86,833 insulin injection records as the dataset for predicting the effect of insulin doses on patients’ blood glucose concentrations (step G in Figure 1).

2.2. Neural Network Training and Hyperparameter Tuning

For data preprocessing, we performed one-hot encoding for two discrete input characteristics (gender and ethnicity). For other numerical input characteristics, we performed min-max normalization and added the square and root of the normalized values into the input. We ended up with a 31-dimensional input dataset for all deep neural networks.
We trained three commonly used numerical regression neural networks for glucose prediction, including FCN, CNN and LSTM. We trained all the network models with different architectures using 80% of our dataset as the same training set. To explore the model architectures with high prediction accuracy, we generated network architectures with different depths and widths for these three kinds of neural network models.
We first tried to train FCNs with different numbers of fully connected layers (FL). When the number of network layers is less than 10, the network prediction accuracy improves with the increase in the network depth, indicating that there is still room for improvement. The prediction accuracy of the 10-layer and 11-layer networks is very close, the accuracy of 12-layer networks has a notable increase, and the 13-layer networks show a decrease in the testing accuracy and have an overfitting tendency. Through experiments, we found that CNN and LSTM networks have similar rules in the number of network layers and the prediction accuracy, so most of the layer numbers we tried were between 10 and 12. We also performed similar experiments to decide the width of the network layers, and we chose the layer size that improves prediction accuracy without causing overfitting while minimizing the computational cost. We tried to use different numbers of convolution layers (CL) and pooling layers (PL) in CNN. For the LSTM networks, we experimented with different numbers of LSTM layers and dropout layers. The complete architectures of the 13 networks with the highest prediction accuracy (RMSE below 16.90 mg/dL) are shown in the Appendix A in Table A2.
The kernel initializer for all Dense layers was RandomNormal, and for all Conv1D layers and LSTM layers was GlorotUniform. The bias initializer for all layers was Zeros. ReLU was used as the activation function of all the network layers. We used the Adam optimizer to update the weights of the networks. The learning rate of the Adam optimizer was 0.001, the beta1 was 0.9, and the beta2 was 0.999. Mean square error (MSE) was used as the loss function in the training process. Therefore, the goal of all the network training processes is to generate F that minimizes the MSE:
a r g m i n 1 n i = 1 n ( F x i y i ) 2
The training epochs of deep neural networks are also important for prediction accuracy. To prevent insufficient training, we did not use the setting of earlystop. We tried different training epochs in 500 epochs increments. With the increase in training epochs, the prediction accuracy of the training set will gradually increase, while the prediction accuracy of the testing set will initially increase and then decrease. The decreased testing accuracy indicates that the network is overfitting to the training set, and the network training should be stopped. The training epochs with the highest test accuracy of each model are finally picked, as shown in the Appendix A in Table A2.
All the networks were trained with the Python Keras framework on four NVIDIA TITAN Xp GPUs with 12 G memory size. To maximize the throughput of the utilization machine, we set the training batchsize to 1000. Figure 3 shows the complete process of our predicting methods.

2.3. Network Evaluation Metrics

We used the following four metrics to measure the prediction accuracy of all network models. (y represents the actual blood glucose value after insulin injection, y ^ represents the predicted blood glucose value after insulin injection, y ¯ represents the average value of y, and n represents the total number of the testing set):
  • MAE: mean absolute error, which is calculated as the sum of absolute errors divided by the sample size. Lower values indicate better model fitting and more accurate prediction.
    M A E = 1 n i = 1 n y i y ^ i
  • RMSE: root mean square error, which is the square root of the average squared errors. Lower values indicate better model fitting and more accurate prediction.
    R M S E = 1 n i = 1 n y i y ^ i 2
  • R2_Score: The coefficient of determination, denoted R 2 or r 2 , can represent the proportion of the dependent variable that is predictable from the independent variables. Higher values indicate better model fitting and more accurate prediction.
    R 2 = 1 i = 1 n y i y ^ i 2 / n i = 1 n y i y ¯ i 2 / n = 1 M S E V a r i a n c e o f y
  • EGA [61]: Clarke error grid analysis, to quantify clinical accuracy of the predicted blood glucose values by subdividing the prediction results into five zones. The prediction values located in Zone A represent accurate predictions, and values in Zone B are acceptable glucose results. Values in Zone C may prompt unnecessary corrections. Values in Zone D indicate a dangerous failure to detect and treat. Values in Zone E represent erroneous treatment. More prediction results located in Zone A and Zone B indicate better model fit.
MAE, RMSE, and R2_ Score are all common metrics for measuring the performance of regression learning models. EGA is one of the “gold standards” for determining the clinical accuracy of blood glucose prediction. These four evaluation metrics provide us with different helpful information. MAE is a linear score, meaning that all individual differences are equally weighted in the average and can visually represent the absolute prediction error. In comparison, since the errors are squared before being averaged in the calculation process of RMSE, RMSE gives relatively high weight to the larger errors and is more sensitive to outliers, which means that RMSE is most useful when significant errors are especially undesirable. Unlike MAE and RMSE for calculating offset distance, R2_ Score indicates how well the predictor variables can explain the variation in the response variable. R2_ Score is conveniently scaled between 0 and 1, independent of the actual value of the regression task, which is obviously easier to interpret and compare across different regression tasks. As a widely recognized indicator, EGA can provide a reference for the clinical reliability of our prediction network.
Our study is dedicated to predicting blood glucose concentrations after short-acting insulin injections. Since our prediction results will serve as an important reference for clinicians or intelligent monitoring devices in determining insulin doses, our prediction network must avoid large prediction errors that may affect the clinical treatment. From the perspectives of clinical accuracy, safety, and reliability, we choose RMSE, which is more sensitive to large prediction errors, as the primary metric to compare the prediction accuracy of all network models, and use the other evaluation metrics as auxiliary references. Most previous glucose prediction studies have similarly used RMSE as the primary assessment criterion [45,46,47,48,49,50,52].

2.4. Machine Learning for Comparison

Regression prediction is one of the most common applications of machine learning. We used scikit-learn to build and train five commonly used regression machine learning models: support vector regression (SVR), K-nearest neighbors regression (KNN), classification and regression tree (CART), random forest regression (RF), and extreme gradient boosting regression (XGBoost). Scikit-learn is a Python module that integrates various state-of-the-art machine-learning algorithms for solving medium-scale supervised and unsupervised problems [62].

2.5. Model Retest

In order to verify the universal effectiveness of our network models and avoid the evaluation bias caused by the single testing set, we used K-fold cross validation [63] to re-validate the prediction accuracy of each network, which is a typical method for the reliability assessment of deep neural networks.
In our study, we relied on 5-fold cross validation to retrain and retest the 13 most accurate deep neural network models. We divided the entire dataset into five non-intersecting parts on average. For each training, we take four of the data parts as the training set and the remaining one as the testing set. After training and testing each model five times, all instances of the entire dataset were tested for each network model, and we obtained the mean prediction accuracy of the five testing sets for each network model. Furthermore, we performed the Chi-squared test [64] on the predicted RMSE values of the 13 networks on the 5 testing sets to prove that the networks have universal accuracy in predicting different testing sets.

3. Experiment Results

3.1. Dataset Extracted from MIMIC-IV Database

From 244,000 short-acting insulin bolus injection events from the MIMIC-IV database, through data screening, glucose matching, and characteristic integration, we obtained 86,833 independent insulin injection events that occurred in the ICU without missing values. These insulin injection records and the relevant information come from 25,168 ICU admissions of 20,426 adult patients. We used 80% of the entries in the dataset as the training set for deep neural networks, and the remaining 20% were used as the testing set.
Each insulin injection record contains 11 items, including the patient’s blood glucose before and after insulin injection, injected insulin dose, gender, age, weight, ethnicity, creatinine, blood urine nitrogen, systolic blood pressure, and diastolic blood pressure. The blood glucose before insulin injection is the closest glucose detection result to the injection event within 90 min before and 10 min after insulin injection. The blood glucose after insulin injection is determined according to the blood glucose detection results closest to 4 h after each injection event, which is the peak time of the efficacy of short-acting insulin. Creatinine and blood urine nitrogen are the basic vital signs of ICU patients, which are the main indicators reflecting the renal function of patients. Table 1 summarizes the statistical descriptions of 11 characteristics of the entire dataset that we created for this study, including the maximum, minimum, average, standard deviation and median of 9 digital characteristics, and the distribution of 2 discrete characteristics.

3.2. Deep Neural Networks and Performance Evaluation

In contrast to previous studies on blood glucose prediction shown in Appendix A Table A1, our study is not aimed at predicting the real-time blood glucose concentrations in the daily life of diabetic patients, but focuses on predicting the clinical effect of insulin dose on blood glucose in each insulin injection event. Most previous studies [45,47,48,50,52] require the relevant records of patients from tens of consecutive days for network training, and the trained models cannot predict the blood glucose for any patient without a long-term blood glucose history. However, in our study, we regarded each insulin injection as an independent atomic event unrelated to the patients’ historical state, and used the patients’ real-time characteristics at the time of insulin injection to predict the blood glucose after the short-acting insulin injection.
We built three kinds of deep neural networks, including FCN, CNN, and LSTM, to atomically predict the ICU patients’ blood glucose concentration after the short-acting insulin injection. We tried dozens different network model architectures to explore whether these three commonly used numerical regression networks can be used for glucose prediction based on discrete insulin injection records and tried to find network models with high accuracy. The same training set was used to train all network models and the same testing set for network evaluation. The testing results of each network were evaluated using MAE, RMSE, and R2_Score, which are the common metrics to measure the prediction accuracy. Furthermore, we used Clarke error grid analysis (EGA) to evaluate the network performance, which is dedicated to measuring the accuracy of blood glucose predictions.
After experiments and comparisons, we obtained 13 network architectures with excellent prediction accuracy (RMSE less than 16.9 mg/dL). The detailed architectures of the 13 networks are shown in Appendix A Table A2, and the testing accuracy of these 13 network models is shown in Table 2. As in previous studies [45,46,47,48,49,50,52], taking RMSE as the primary evaluation criterion, the testing RMSEs of the five most accurate networks are all about 16 mg/dL, and the smallest RMSE is as low as 15.82 mg/dL. The accuracy of predicting using stochastically independent insulin injection records in this study is much better than that of previous studies [45,46,47,48,49,50,52] (RMSE more than 20 mg/dL, shown in Appendix A Table A1) predicted by continuous glucose monitoring data. Clarke error grid analysis of the prediction results of the five most accurate models shows that about 94% of the predicted values are located in Zone A of the error grid, and 5.5% are in Zone B, which means that over 99.5% of testing results could be regarded as being clinically acceptable. The Clarke error grid of the prediction results of the testing set for these five network models with the lowest RMSE is shown in Figure 4.
The testing accuracy of our proposed models demonstrates that the three most commonly used deep neural regression models (FCN, CNN, and LSTM) can accurately predict the blood glucose concentration after any atomically independent insulin-injection event without requiring any long-term relevant records. Our proposed method makes blood glucose prediction more effective, convenient and feasible in clinical practice.

3.3. Comparison with Machine Learning

To demonstrate the superiority of deep neural networks, we tried to use several machine learning models for blood glucose prediction. We trained five nonlinear regression machine learning models, including SVR, KNN, CART, RF, and XGBoost. The five models were trained using the same training set as the deep neural networks, and the same testing set was used for accuracy evaluation. Table 3 shows the prediction accuracy of the five machine learning models used for comparison. Among the five machine learning models, RF has the lowest RMSE and the highest R2_Score (RMSE = 17.07 mg/dL, R2_Score = 0.8869), and the CART model has the lowest MAE (MAE = 5.97 mg/dL). RF also performs best in EGA, with approximately 92.80% of the prediction results located in Zone A and Zone B. The comparison results show that all of the 13 individual neural networks perform better than the best machine learning models in all accuracy evaluation metrics.
It is worth discussing that the RMSE of RF is very close to our proposed deep neural network models, but its MAE is much higher. RF is an ensemble learning method based on decision trees, which integrates the prediction results of many decision trees by regressing the average prediction value. Since RF averages the predictions of multiple CART models, RF has a significant effect in reducing variance, which results in better testing RMSE and R2_Score. However, the MAE value of our RF model is as high as 8.29 mg/dL, which indicates that the RF model tends to overfit during the training process. In pursuit of the accurate prediction of some noise data and reducing the RMSE, which is sensitive to outliers, RF sacrifices the absolute prediction error of the overall dataset in the training process.

3.4. Test–Retest Reliability

We retested 13 deep neural networks using 5-fold cross validation to verify the accuracy of the network models. The 86,833 records in the dataset were randomly divided into five parts called folds, and all of the folds have instances of equal size (17,366 for each fold). Each model was trained and tested five times, each time using one part as the testing set and the other four parts as the training set. The final prediction accuracy of the 5-fold cross validation for each network model was the average test performance of the five parts, which is presented in Table 4. For all the deep neural network models proposed in this study, the RMSE values after retesting does not increase by more than 0.5 mg/dL, demonstrating that our models are generally effective in predicting the effect of insulin dose on blood glucose for ICU patients.
We performed the Chi-squared test on the RMSE values of each testing fold for these 13 networks. The RMSE values of 13 models against five testing folds and the p-value of the Chi-squared test are shown in the Appendix A in Table A3. The p-value of the Chi-squared test is greater than 0.995, indicating that there is almost no statistical difference in the prediction accuracy for the five different testing sets, which further proves that the prediction results of the proposed models are generally accurate.

4. Conclusions

In this study, we developed 13 deep neural networks with different architectures, using 10 related characteristics as the network input to atomically predict the blood glucose concentrations of ICU patients after stochastically short-acting insulin injections without requiring any long-term history records..
We extracted 86,833 short-acting insulin bolus injection records discretely occurring in the ICU from the MIMIC-IV dataset, each record containing 11 characteristics associated with insulin efficacy. Using these records as the training and testing sets, we trained 13 deep neural network models with different architectures (shown in the Appendix A in Table A2). All of these 13 networks achieve good prediction accuracy. The testing RMSEs of these 13 deep neural networks are all below 16.90 mg/dL, reaching a minimum of 15.82 mg/dL, which is superior to five machine learning regression models (shown in Table 3) and the methods proposed in previous blood glucose prediction studies (shown in Appendix A Table A1.) The predicting EGA of these 13 network models shows that more than 99.5% of the prediction results are clinically acceptable.
In order to prevent the evaluation bias of model accuracy caused by the single testing set, we retested the 13 individual network models using 5-fold cross validation, and we performed the Chi-squared test on the prediction accuracy of the five different testing sets. The retesting results demonstrate that the network architecture proposed in this study is universally accurate in predicting the effect of insulin dose on blood glucose values.
In summary, our study demonstrates that it is a feasible, effective, and accurate method to atomically predict the effect of insulin doses on glucose concentrations using deep neural networks and discontinuous insulin injection records. The blood glucose concentration predicted by the deep neural networks has reference significance in clinical drug dosage. The various deep neural network architectures proposed in this study provide a reference for subsequent studies related to automatic insulin injection and continuous glucose monitoring. Our network models also provide basic simulation environments for further study on insulin dose recommendations using deep reinforcement learning.

Author Contributions

Conceptualization, Y.Y., B.T., L.Q. and J.Y.; methodology, Y.Y. and B.T.; software, B.T.; validation, S.Z. and B.T.; formal analysis, B.T., S.Z. and J.S.; investigation, L.Q. and B.T.; data curation, B.T. and J.S.; writing—original draft preparation, B.T. and Y.Y.; writing—review and editing, B.T., Y.Y., J.Y. and L.Q.; visualization, B.T. and J.Y.; supervision, Y.Y.; project administration, Y.Y.; funding acquisition, L.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by two funds: the National Nature Science Foundation of China (No. 62072488) and the Beijing Natural Science Foundation (No. 4202064).

Institutional Review Board Statement

The dataset we used in this study is extracted from the open database MIMIC-IV v1.0. The collection of patient information and creation of the research resource was approved by the Institutional Review Board of the Beth Israel Deaconess Medical Center (protocol code 2001-P-001699/14), which granted a waiver of informed consent and approved the data-sharing initiative.

Informed Consent Statement

Patient consent was waived due to all the medical data in the MIMIC-IV database being deidentified. The Institutional Review Board of the Beth Israel Deaconess Medical Center granted a waiver of informed consent and approved the data-sharing initiative.

Data Availability Statement

The MIMIC-IV database is openly available in https://doi.org/10.13026/s6n6-xd98 (accessed on 23 December 2021).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Appendix Tables

Table A1. Summary of previous studies addressing blood glucose prediction using deep nueral networks: model type, number of participants in the cohort, mean number of monitored days per patient, methods and frequency of blood glucose monitoring, whether continuous recording is required (C), whether manual record is required (M), prediction horizon (PH), test performance of the model, citation and publish year.
Table A1. Summary of previous studies addressing blood glucose prediction using deep nueral networks: model type, number of participants in the cohort, mean number of monitored days per patient, methods and frequency of blood glucose monitoring, whether continuous recording is required (C), whether manual record is required (M), prediction horizon (PH), test performance of the model, citation and publish year.
ModelCohortDaysGlucose RecordCMPHTest PerformanceRefYear
Previous related studies:
FCN 1100 VP 2-random simulation after mealreduce BGRI 3[44]2018
LSTM 41 T2D 530by SMBG 6 60 minRMSE 7≈18.79 mg/dL[45]2019
LSTM174 patients2∼3every 5 min by CGM 8 30/60/120 minRMSE≈19.44/22.86/24.48 mg/dL[46]2019
CNN 920 VP, 16 T1D 10180every 5 min by CGM30/60 minRMSE≈19.28/31.83 mg/dL[47]2020
CRNN 1110 VP, 10 T1D180every 5 min by CGM30/60 minRMSE≈21.07/33.27 mg/dL[48]2020
LSTM5 T1D400every 5 min by CGM30/60 minRMSE≈21.4/38.0 mg/dL[49]2017
GRU 1212 T1D56every 5 min by CGM 30 minRMSE≈21.54 mg/dL[50]2021
FCN25 T1D1∼2every 5 min by CGM 30 minPRED-EGA 13, accuracy < 95%[51]2017
FCN27 patients16every 5 min by CGM75 minRMSE≈43.9 mg/dL[52]2011
Best prediction model proposed in this study:
LSTM20,426 patients-discrete record about 4hRMSE≈15.82 mg/dL-2022
1 Fully connected neural network. 2 Virtual patient. 3 Blood glucose risk index. 4 Long short-term memory network. 5 Type 2 diabetes. 6 Self-monitoring blood glucose device. 7 Root mean square error. 8 Continuous glucose monitoring devices. 9 Convolutional neural network. 10 Type 1 diabetes. 11 Convolutional recurrent neural network. 12 Gated recurrence unit. 13 Prediction-error grid analysis, an extension of the Clark error grid analysis.
Table A2. Detailed architecture of 13 different deep neural networks.
Table A2. Detailed architecture of 13 different deep neural networks.
network_id:FCN1layer number:12                        parameters number:6306945
training epochs:7000layer type:Fully Connected Layer(units) 1
network architecture:
FCL(2048)-FCL(1536)-FCL(1024)-FCL(768)-FCL(512)-FCL(384)-FCL(256)-FCL(128)-FCL(64)-FCL(32)-FCL(16)-FCL(1)
network_id:FCN2layer number:10parameters number:914905
training epochs:4000layer type:Fully Connected Layer(units)
network architecture:
FCL(1024)-FCL(512)-FCL(400)-FCL(256)-FCL(128)-FCL(72)-FCL(64)-FCL(32)-FCL(16)-FCL(1)
network_id:CNN1layer number:12parameters number:1040241
training epochs:7000layer type:Convolutional Layer(filter, kernel_size) 2, Pooling Layer(pool_size) 3,
Fully Connected Layer(units)
network architecture:
CL(16,3)-CL(64,3)-CL(128,3)-CL(256,3)-CL(512,3)-CL(256,3)-CL(128,3)-CL(64,3)-CL(16,3)-CL(8,3)-PL(3)-FCL(1)
network_id:CNN2layer number:12parameters number:5868897
training epochs:4000layer type:Convolutional Layer(filter, kernel_size), Fully Connected Layer(units)
network architecture:
CL(32,3)-CL(64,3)-CL(128,3)-CL(64,3)-FCL(2048)-FCL(1024)-FCL(512)-FCL(256)-FCL(128)-FCL(64)-FCL(16)-FCL(1)
network_id:CNN3layer number:12parameters number:1040353
training epochs:7000layer type:Convolutional Layer(filter, kernel_size),Fully Connected Layer(units)
network architecture:
CL(16,3)-CL(64,3)-CL(128,3)-CL(256,3)-CL(512,3)-CL(256,3)-CL(128,3)-CL(64,3)-CL(16,3)-CL(8,3)-CL(4,3)-FCL(1)
network_id:CNN4layer number:12parameters number:9381633
training epochs:4000layer type:Convolutional Layer(filter, kernel_size), Fully Connected Layer(units)
network architecture:
CL(16,3)-CL(64,3)-CL(128,3)-FCL(2048)-FCL(1024)-FCL(512)-FCL(256)-FCL(128)-FCL(64)-FCL(32)-FCL(16)-FCL(1)
network_id:CNN5layer number:12parameters number:7753953
training epochs:7000layer type:Convolutional Layer(filter, kernel_size),Fully Connected Layer(units)
network architecture:
CL(32,3)-CL(64,3)-CL(128,3)-CL(64,3)-FCL(2048)-FCL(1536)-FCL(768)-FCL(384)-FCL(128)-FCL(64)-FCL(16)-FCL(1)
network_id:CNN6layer number:12parameters number:1041937
training epochs:7000layer type:Convolutional Layer(filter, kernel_size), Pooling Layer(pool_size),
Fully Connected Layer(units)
network architecture:
CL(16,3)-CL(64,3)-CL(128,3)-CL(256,3)-CL(512,3)-CL(256,3)-CL(128,3)-CL(64,3)-CL(16,3)-PL(3)-FCL(32)-FCL(1)
network_id:CNN7layer number:11parameters number:1047777
training epochs:4000layer type:Convolutional Layer(filter, kernel_size),Fully Connected Layer(units)
network architecture:
CL(32,3)-CL(64,3)-CL(128,3)-CL(256,3)-CL(512,3)-CL(256,3)-CL(128,3)-CL(64,3)-CL(32,3)-CL(16,3)-FCL(1)
network_id:LSTM1layer number:12parameters number:75779817
training epochs:6000layer type:LSTM Layer(units) 4, TimeDistributed Layer(Fully Connected Layer(units))
network architecture:
LSTM(2048)-LSTM(2048)-LSTM(1024)-LSTM(1024)-LSTM(512)-LSTM(256)-LSTM(128)-LSTM(64)-LSTM(32)-LSTM(16)
-LSTM(8)-TDL(FCL(1))
network_id:LSTM2layer number:12parameters number:8523497
training epochs:6000layer type:LSTM Layer(units), Dropout Layer(dropout_rate),
TimeDistributed Layer(Fully Connected Layer(units))
network architecture:
LSTM(1024)-Dropout(0.2)-LSTM(512)-Dropout(0.2)-LSTM(256)-Dropout(0.2)-LSTM(128)-LSTM(64)-LSTM(32)-LSTM(16)
-LSTM(8)-TDL(FCL(1))
network_id:LSTM3layer number:12parameters number:33824489
training epochs:3000layer type:LSTM Layer(units), Dropout Layer(dropout_rate),
TimeDistributed Layer(Fully Connected Layer(units))
network architecture:
LSTM(2048)-Dropout(0.2)-LSTM(1024)-Dropout(0.2)-LSTM(512)-LSTM(256)-LSTM(128)-LSTM(64)-LSTM(32)-LSTM(16)
-LSTM(8)-TDL(FCL(1))
network_id:LSTM4layer number:12parameters number:44316393
training epochs:6000layer type:LSTM Layer(units), TimeDistributed Layer(Fully Connected Layer(units))
network architecture:
LSTM(2048)-LSTM(1024)-LSTM(1024)-LSTM(512)-LSTM(512)-LSTM(256)-LSTM(128)-LSTM(64)-LSTM(32)-LSTM(16)
-LSTM(8)-TDL(FCL(1))
1 “Units” represents the activation unit number for this fully connected layer (FCL). 2 “Filter” and “kernel_size” represent the number of output filters and the width of the convolution window for this convolutional layer (CL). 3 “Pool_size” represents the size of the spatial window for this pooling layer (PL). 4 “Units” represents the dimensionality of the output space for this LSTM layer.
Table A3. The prediction RMSE for the five testing folds and the results of the Chi-squared test.
Table A3. The prediction RMSE for the five testing folds and the results of the Chi-squared test.
The RMSE Value for Each Testing Fold (mg/dL)
ModelsFold1Fold2Fold3Fold4Fold5
FCN115.9616.0816.8316.4616.68
FCN215.9316.3916.8816.7416.87
CNN115.8316.1216.9516.6916.07
CNN216.0316.2516.5416.0216.16
CNN316.1216.4916.8816.1716.58
CNN416.2816.0016.6416.0915.90
CNN516.0916.0316.6416.5916.14
CNN615.8716.0416.7116.5416.35
CNN715.7916.2216.616.8316.16
LSTM116.4316.5316.6616.6016.42
LSTM216.6717.0617.4417.0617.06
LSTM316.816.8117.4117.2817.46
LSTM416.2716.7216.8616.5316.69
The results for the Chi-Squared Test
χ 2 :0.1236
Freedom degree:48
p-value:>0.995

References

  1. IDF Diabetes Atlas. Available online: https://www.diabetesatlas.org (accessed on 21 June 2022).
  2. Baker, L.; Maley, J.H.; Arévalo, A.; DeMichele, F.; Mateo-Collado, R.; Finkelstein, S.; Celi, L.A. Real-world characterization of blood glucose control and insulin use in the intensive care unit. Sci. Rep. 2020, 10, 10718. [Google Scholar] [CrossRef] [PubMed]
  3. Abdelhamid, Y.A.; Kar, P.; Finnis, M.E.; Phillips, L.K.; Plummer, M.P.; Shaw, J.E.; Horowitz, M.; Deane, A.M. Stress hyperglycaemia in critically ill patients and the subsequent risk of diabetes: A systematic review and meta-analysis. Crit. Care 2016, 20, 301. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Marik, P.E.; Bellomo, R. Stress hyperglycemia: An essential survival response! Crit. Care 2013, 17, 305. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Umpierrez, G.E.; Isaacs, S.D.; Bazargan, N.; You, X.; Thaler, L.M.; Kitabchi, A.E. Hyperglycemia: An Independent Marker of In-Hospital Mortality in Patients with Undiagnosed Diabetes. J. Clin. Endocrinol. Metab. 2002, 87, 978–982. [Google Scholar] [CrossRef]
  6. Whitcomb, B.W.; Pradhan, E.K.; Pittas, A.G.; Roghmann, M.C.; Perencevich, E.N. Impact of admission hyperglycemia on hospital mortality in various intensive care unit populations. Crit. Care Med. 2005, 33, 2772–2777. [Google Scholar] [CrossRef]
  7. Barsheshet, A.; Garty, M.; Grossman, E.; Sandach, A.; Lewis, B.S.; Gottlieb, S.; Shotan, A.; Behar, S.; Caspi, A.; Schwartz, R.; et al. Admission blood glucose level and mortality among hospitalized nondiabetic patients with heart failure. Arch. Intern. Med. 2006, 166, 1613–1619. [Google Scholar] [CrossRef] [Green Version]
  8. Preiser, J.C.; Marik, P.E. Hyperglycemia-related mortality in critically ill patients varies with admission diagnosis. Crit. Care Med. 2010, 38, 1388. [Google Scholar] [CrossRef]
  9. Viana, M.V.; Moraes, R.B.; Fabbrin, A.R.; Santos, M.F.; Gerchman, F. Assessment and treatment of hyperglycemia in critically ill patients. Rev. Bras. Ter. Intensiv. 2014, 26, 71–76. [Google Scholar] [CrossRef]
  10. Liao, W.I.; Wang, J.C.; Chang, W.C.; Hsu, C.W.; Chu, C.M.; Tsai, S.H. Usefulness of glycemic gap to predict ICU mortality in critically ill patients with diabetes. Medicine 2015, 94, e1525. [Google Scholar] [CrossRef]
  11. Van den Berghe, G.; Wouters, P.; Weekers, F.; Verwaest, C.; Bruyninckx, F.; Schetz, M.; Vlasselaers, D.; Ferdinande, P.; Lauwers, P.; Bouillon, R. Intensive Insulin Therapy in Critically Ill Patients. N. Engl. J. Med. 2001, 345, 1359–1367. [Google Scholar] [CrossRef]
  12. Malmberg, K.; Norhammar, A.; Wedel, H.; Rydén, L. Glycometabolic State at Admission: Important Risk Marker of Mortality in Conventionally Treated Patients With Diabetes Mellitus and Acute Myocardial Infarction. Circulation 1999, 99, 2626–2632. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Furnary, A.P.; Gao, G.; Grunkemeier, G.L.; Wu, Y.X.; Zerr, K.J.; Bookin, S.O.; Floten, H.S.; Starr, A. Continuous insulin infusion reduces mortality in patients with diabetes undergoing coronary artery bypass grafting. J. Thorac. Cardiovasc. Surg. 2003, 125, 1007–1021. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Van den Berghe, G.; Wilmer, A.; Hermans, G.; Meersseman, W.; Wouters, P.J.; Milants, I.; Van Wijngaerden, E.; Bobbaers, H.; Bouillon, R. Intensive Insulin Therapy in the Medical ICU. N. Engl. J. Med. 2006, 354, 449–461. [Google Scholar] [CrossRef] [Green Version]
  15. Reed, C.C.; Stewart, R.M.; Sherman, M.; Myers, J.G.; Corneille, M.G.; Larson, N.; Gerhardt, S.; Beadle, R.; Gamboa, C.; Dent, D.; et al. Intensive Insulin Protocol Improves Glucose Control and Is Associated with a Reduction in Intensive Care Unit Mortality. J. Am. Coll. Surg. 2007, 204, 1048–1055. [Google Scholar] [CrossRef] [Green Version]
  16. Vlasselaers, D.; Milants, I.; Desmet, L.; Wouters, P.J.; Vanhorebeek, I.; van den Heuvel, I.; Mesotten, D.; Casaer, M.P.; Meyfroidt, G.; Ingels, C.; et al. Intensive insulin therapy for patients in paediatric intensive care: A prospective, randomised controlled study. Lancet 2009, 373, 547–556. [Google Scholar] [CrossRef]
  17. Wang, H.C.; Lee, A.R. Recent developments in blood glucose sensors. J. Food Drug Anal. 2015, 23, 191–200. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386. [Google Scholar] [CrossRef] [Green Version]
  19. Werbos, P. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 1974. [Google Scholar]
  20. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  21. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  22. Fukushima, K.; Miyake, S. Neocognitron: A Self-Organizing Neural Network Model for a Mechanism of Visual Pattern Recognition. In Proceedings of the Competition and Cooperation in Neural Nets, Kyoto, Japan, 15–19 February 1982; pp. 267–285. [Google Scholar] [CrossRef]
  23. Waibel, A.; Hanazawa, T.; Hinton, G.; Shikano, K.; Lang, K. Phoneme recognition using time-delay neural networks. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 328–339. [Google Scholar] [CrossRef]
  24. Hinton, G.E.; Osindero, S.; Teh, Y.W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  25. Dahl, G.E.; Yu, D.; Deng, L.; Acero, A. Context-Dependent Pre-Trained Deep Neural Networks for Large-Vocabulary Speech Recognition. IEEE Trans. Audio Speech Lang. Process. 2012, 20, 30–42. [Google Scholar] [CrossRef] [Green Version]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  27. Schroff, F.; Kalenichenko, D.; Philbin, J. FaceNet: A unified embedding for face recognition and clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar] [CrossRef] [Green Version]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  29. Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; Van Den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef]
  30. Naylor, C.D. On the Prospects for a (Deep) Learning Health Care System. JAMA 2018, 320, 1099–1100. [Google Scholar] [CrossRef] [PubMed]
  31. Stead, W.W. Clinical Implications and Challenges of Artificial Intelligence and Deep Learning. JAMA 2018, 320, 1107–1108. [Google Scholar] [CrossRef] [PubMed]
  32. Hinton, G. Deep Learning—A Technology With the Potential to Transform Health Care. JAMA 2018, 320, 1101–1102. [Google Scholar] [CrossRef]
  33. Ehteshami Bejnordi, B.; Veta, M.; Johannes van Diest, P.; van Ginneken, B.; Karssemeijer, N.; Litjens, G.; van der Laak, J.A.W.M.; the CAMELYON16 Consortium. Diagnostic Assessment of Deep Learning Algorithms for Detection of Lymph Node Metastases in Women With Breast Cancer. JAMA 2017, 318, 2199–2210. [Google Scholar] [CrossRef] [Green Version]
  34. De Fauw, J.; Ledsam, J.R.; Romera-Paredes, B.; Nikolov, S.; Tomasev, N.; Blackwell, S.; Askham, H.; Glorot, X.; O’Donoghue, B.; Visentin, D.; et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 2018, 24, 1342–1350. [Google Scholar] [CrossRef]
  35. Ribeiro, A.H.; Ribeiro, M.H.; Paixão, G.M.M.; Oliveira, D.M.; Gomes, P.R.; Canazart, J.A.; Ferreira, M.P.S.; Andersson, C.R.; Macfarlane, P.W.; Meira Jr., W.; et al. Automatic diagnosis of the 12-lead ECG using a deep neural network. Nat. Commun. 2020, 11, 1760. [Google Scholar] [CrossRef] [Green Version]
  36. Lima, E.M.; Ribeiro, A.H.; Paixão, G.M.M.; Ribeiro, M.H.; Pinto-Filho, M.M.; Gomes, P.R.; Oliveira, D.M.; Sabino, E.C.; Duncan, B.B.; Giatti, L.; et al. Deep neural network-estimated electrocardiographic age as a mortality predictor. Nat. Commun. 2021, 12, 5117. [Google Scholar] [CrossRef] [PubMed]
  37. Zhang, Y.; An, L.; Xu, J.; Zhang, B.; Zheng, W.J.; Hu, M.; Tang, J.; Yue, F. Enhancing Hi-C data resolution with deep convolutional neural network HiCPlus. Nat. Commun. 2018, 9, 750. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. de Vos, B.D.; Berendsen, F.F.; Viergever, M.A.; Sokooti, H.; Staring, M.; Išgum, I. A deep learning framework for unsupervised affine and deformable image registration. Med. Image Anal. 2019, 52, 128–143. [Google Scholar] [CrossRef] [Green Version]
  39. Ahn, S.H.; Yeo, A.U.; Kim, K.H.; Kim, C.; Goh, Y.; Cho, S.; Lee, S.B.; Lim, Y.K.; Kim, H.; Shin, D.; et al. Comparative clinical evaluation of atlas and deep-learning-based auto-segmentation of organ structures in liver cancer. Radiat. Oncol. 2019, 14, 213. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Zeleznik, R.; Foldyna, B.; Eslami, P.; Weiss, J.; Alexander, I.; Taron, J.; Parmar, C.; Alvi, R.M.; Banerji, D.; Uno, M.; et al. Deep convolutional neural networks to predict cardiovascular risk from computed tomography. Nat. Commun. 2021, 12, 715. [Google Scholar] [CrossRef]
  41. Alghamdi, M.; Al-Mallah, M.; Keteyian, S.; Brawner, C.; Ehrman, J.; Sakr, S. Predicting diabetes mellitus using SMOTE and ensemble machine learning approach: The Henry Ford ExercIse Testing (FIT) project. PLoS ONE 2017, 12, e0179805. [Google Scholar] [CrossRef] [Green Version]
  42. Ibragimov, B.; Toesca, D.A.; Chang, D.T.; Yuan, Y.; Koong, A.C.; Xing, L. Automated hepatobiliary toxicity prediction after liver stereotactic body radiation therapy with deep learning-based portal vein segmentation. Neurocomputing 2020, 392, 181–188. [Google Scholar] [CrossRef]
  43. Zhen, X.; Chen, J.; Zhong, Z.; Hrycushko, B.; Zhou, L.; Jiang, S.; Albuquerque, K.; Gu, X. Deep convolutional neural network with transfer learning for rectum toxicity prediction in cervical cancer radiotherapy: A feasibility study. Phys. Med. Biol. 2017, 62, 8246–8263. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Cappon, G.; Vettoretti, M.; Marturano, F.; Facchinetti, A.; Sparacino, G. A Neural-Network-Based Approach to Personalize Insulin Bolus Calculation Using Continuous Glucose Monitoring. J. Diabetes Sci. Technol. 2018, 12, 265–272. [Google Scholar] [CrossRef]
  45. Padmapritha, T. Prediction of Blood Glucose Level by using an LSTM based Recurrent Neural networks. In Proceedings of the 2019 IEEE International Conference on Clean Energy and Energy Efficient Electronics Circuit for Sustainable Development (INCCES), Krishnankoil, India, 18–20 December 2019; pp. 1–4. [Google Scholar] [CrossRef]
  46. Song, W.; Cai, W.; Li, J.; Jiang, F.; He, S. Predicting Blood Glucose Levels with EMD and LSTM Based CGM Data. In Proceedings of the 2019 6th International Conference on Systems and Informatics (ICSAI), Shanghai, China, 2–4 November 2019; pp. 1443–1448. [Google Scholar] [CrossRef]
  47. Li, K.; Liu, C.; Zhu, T.; Herrero, P.; Georgiou, P. GluNet: A Deep Learning Framework for Accurate Glucose Forecasting. IEEE J. Biomed. Health Inform. 2020, 24, 414–423. [Google Scholar] [CrossRef]
  48. Li, K.; Daniels, J.; Liu, C.; Herrero, P.; Georgiou, P. Convolutional Recurrent Neural Networks for Glucose Prediction. IEEE J. Biomed. Health Inform. 2020, 24, 603–613. [Google Scholar] [CrossRef] [Green Version]
  49. Mirshekarian, S.; Bunescu, R.; Marling, C.; Schwartz, F. Using LSTMs to learn physiological models of blood glucose behavior. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Jeju, Korea, 11–15 July 2017; pp. 2887–2891. [Google Scholar] [CrossRef]
  50. Dudukcu, H.V.; Taskiran, M.; Yildirim, T. Consolidated or individual training: Which one is better for blood glucose prediction? In Proceedings of the 2021 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Kocaeli, Turkey, 25–27 August 2021; pp. 1–6. [Google Scholar] [CrossRef]
  51. Mhaskar, H.N.; Pereverzyev, S.V.; van der Walt, M.D. A Deep Learning Approach to Diabetic Blood Glucose Prediction. Front. Appl. Math. Stat. 2017, 3, 14. [Google Scholar] [CrossRef] [Green Version]
  52. Pappada, S.M.; Cameron, B.D.; Rosman, P.M.; Bourey, R.E.; Papadimos, T.J.; Olorunto, W.; Borst, M.J. Neural Network-Based Real-Time Prediction of Glucose in Patients with Insulin-Dependent Diabetes. Diabetes Technol. Ther. 2011, 13, 135–141. [Google Scholar] [CrossRef]
  53. Johnson, A.; Bulgarelli, L.; Pollard, T.; Horng, S.; Celi, L.A.; Mark., R. MIMIC-IV (version 1.0). PhysioNet 2021. [Google Scholar] [CrossRef]
  54. Leahy, S.; O’ Halloran, A.; O’ Leary, N.; Healy, M.; McCormack, M.; Kenny, R.; O’ Connell, J. Prevalence and correlates of diagnosed and undiagnosed type 2 diabetes mellitus and pre-diabetes in older adults: Findings from the Irish Longitudinal Study on Ageing (TILDA). Diabetes Res. Clin. Pract. 2015, 110, 241–249. [Google Scholar] [CrossRef]
  55. Alhyas, L.; McKay, A.; Majeed, A. Prevalence of Type 2 Diabetes in the States of The Co-Operation Council for the Arab States of the Gulf: A Systematic Review. PLoS ONE 2012, 7, e0040948. [Google Scholar] [CrossRef] [Green Version]
  56. Wild, S.; Roglic, G.; Green, A.; Sicree, R.; King, H. Global Prevalence of Diabetes: Estimates for the year 2000 and projections for 2030. Diabetes Care 2004, 27, 1047–1053. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Berkowitz, G.S.; Lapinski, R.H.; Wein, R.; Lee, D. Race/Ethnicity and Other Risk Factors for Gestational Diabetes. Am. J. Epidemiol. 1992, 135, 965–973. [Google Scholar] [CrossRef] [PubMed]
  58. Cheng, Y.J.; Kanaya, A.M.; Araneta, M.R.G.; Saydah, S.H.; Kahn, H.S.; Gregg, E.W.; Fujimoto, W.Y.; Imperatore, G. Prevalence of Diabetes by Race and Ethnicity in the United States, 2011–2016. JAMA 2019, 322, 2389–2398. [Google Scholar] [CrossRef]
  59. Alaveras, A.; Thomas, S.; Sagriotis, A.; Viberti, G. Promoters of progression of diabetic nephropathy: The relative roles of blood glucose and blood pressure control. Nephrol. Dial. Transplant. Off. Publ. Eur. Dial. Transpl. Assoc.-Eur. Ren. Assoc. 1997, 12 (Suppl. S2), 71–74. [Google Scholar]
  60. de Boer, M.J.; Miedema, K.; Casparie, A.F. Glycosylated haemoglobin in renal failure. Diabetologia 1980, 18, 437–440. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Clarke, W.L. The Original Clarke Error Grid Analysis (EGA). Diabetes Technol. Ther. 2005, 7, 776–779. [Google Scholar] [CrossRef] [PubMed]
  62. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  63. Bengio, Y.; Grandvalet, Y. No Unbiased Estimator of the Variance of K-Fold Cross-Validation. J. Mach. Learn. Res. 2004, 5, 1089–1105. [Google Scholar]
  64. Fisher, R.A. On the Interpretation of χ2 from Contingency Tables, and the Calculation of P. J. R. Stat. Soc. 1922, 85, 87–94. [Google Scholar] [CrossRef]
Figure 1. The overview process of dataset extraction.
Figure 1. The overview process of dataset extraction.
Sensors 22 08454 g001
Figure 2. (a) Probability density diagram of timediff_before. (b) Probability density diagram of timediff_after.
Figure 2. (a) Probability density diagram of timediff_before. (b) Probability density diagram of timediff_after.
Sensors 22 08454 g002
Figure 3. The overview process of the predicting framework.
Figure 3. The overview process of the predicting framework.
Sensors 22 08454 g003
Figure 4. (a) Clarke Error Grid Analysis of model FCN1. (b) Clarke Error Grid Analysis of model CNN1. (c) Clarke Error Grid Analysis of model CNN2. (d) Clarke Error Grid Analysis of model CNN3. (e) Clarke Error Grid Analysis of model LSTM1.
Figure 4. (a) Clarke Error Grid Analysis of model FCN1. (b) Clarke Error Grid Analysis of model CNN1. (c) Clarke Error Grid Analysis of model CNN2. (d) Clarke Error Grid Analysis of model CNN3. (e) Clarke Error Grid Analysis of model LSTM1.
Sensors 22 08454 g004
Table 1. Statistical description of the 11 characteristics in the dataset.
Table 1. Statistical description of the 11 characteristics in the dataset.
Characteristics 1MaximumMinimumMean ± Standard DeviationMedian
      Blood glucose before insulin (mg/dL)            399.0059.00198.49 ± 56.45184.00
      Blood glucose after insulin (mg/dL)398.0017.00155.08 ± 50.58146.00
      Insulin dose (unit)35.000.104.69 ± 3.604.00
      Age91.0018.0065.35 ± 13.7667.00
      Weight (kg)587.0023.9086.86 ± 25.0883.70
      Systolic blood pressure (mmHg)228.0039.00119.27 ± 19.44116.50
      Diastolic blood pressure (mmHg)156.0014.7060.2 ± 12.5958.75
      Creatinine (mg/dL)22.600.101.61 ± 1.501.10
      Blood urea nitrogen (mg/dL)283.002.0034.26 ± 26.4425.00
      Gender (num/%) 2FemaleMale
33,942/39.09%52,891/60.91%
      Ethnicity (num/%) 3American Indian, Asian, Black, Hispanic(Latino), White, Others (Unknown)
205, 2285, 8549, 3676, 58,136, 13,982/0.24%, 2.63%, 9.85%, 4.23%, 66.95%, 16.10%
The table summarizes the basic statistical characteristics of the dataset used in this study. 1 Each record contains 11 characteristics with different units of measurement. The table contains the minimum, maximum, mean, median and standard deviation of nine numerical characteristics, and the distribution of two discrete characteristics. 2 Distribution of patients’ gender for all entries in the dataset. 3 Distribution of patients’ ethnicity for all entries in the dataset, including six categories.
Table 2. The prediction accuracy of 13 deep neural networks.
Table 2. The prediction accuracy of 13 deep neural networks.
Network_idRMSE (mg/dL) 1MAE (mg/dL) 2R2_Score 3EGA (A%_B%_C%_D%_E%) 4
FCN116.195.270.898294.06%_5.48%_0.17%_0.27%_0.02%
FCN216.565.790.893594.00%_5.53%_0.13%_0.30%_0.04%
CNN116.065.560.899994.07%_5.51%_0.16%_0.22%_0.04%
CNN216.175.550.898493.97%_5.59%_0.12%_0.29%_0.03%
CNN316.245.060.897794.03%_5.53%_0.17%_0.27%_0.03%
CNN416.245.210.897694.26%_5.27%_0.10%_0.34%_0.02%
CNN516.295.090.896994.03%_5.51%_0.16%_0.28%_0.02%
CNN616.546.400.893793.71%_5.78%_0.16%_0.32%_0.03%
CNN716.295.660.897093.99%_5.59%_0.12%_0.26%_0.03%
LSTM115.825.170.902794.60%_4.95%_0.17%_0.25%_0.03%
LSTM216.827.470.890294.03%_5.49%_0.19%_0.27%_0.02%
LSTM316.8810.630.889493.95%_5.53%_0.18%_0.31%_0.03%
LSTM416.405.350.895693.95%_5.58%_0.13%_0.31%_0.03%
1 Root mean square error, the standard deviation of the prediction error, lower values indicate better model fit. 2 Mean absolute error, the arithmetic average of the absolute errors, lower values indicate better model fit. 3 The coefficient of determination, the proportion of dependent variables that can be predicted through regression relationship; larger values indicate better model fit. 4 Clarke error grid analysis; the more values that appear in Zones A and B indicate better model fit.
Table 3. The prediction accuracy of five machine learning models.
Table 3. The prediction accuracy of five machine learning models.
Model TypeRMSE (mg/dL)MAE (mg/dL)R2_ScoreEGA (A%_B%_C%_D%_E%)
SVR34.0425.430.550472.65%_25.60%_0.66%_0.97%_0.13%
KNN22.209.750.808689.02%_9.98%_0.23%_0.69%_0.08%
CART20.045.970.844192.74%_6.56%_0.26%_0.33%_0.10%
RF17.078.290.886992.80%_6.58%_0.11%_0.48%_0.03%
XGBoost30.1523.080.647277.00%_21.65%_0.26%_1.04%_0.05%
Table 4. The average prediction accuracy of 5-fold cross validation of 13 neural network models.
Table 4. The average prediction accuracy of 5-fold cross validation of 13 neural network models.
Network_idRMSE (mg/dL)MAE (mg/dL)R2_ScoreEGA (A%_B%_C%_D%_E%)
FCN116.405.340.894794.03%_5.50%_0.18%_0.27%_0.02%
FCN216.565.490.892694.21%_5.33%_0.12%_0.30%_0.04%
CNN116.335.670.895694.07%_5.44%_0.17%_0.29%_0.02%
CNN216.205.360.897394.02%_5.52%_0.16%_0.27%_0.03%
CNN316.455.710.894293.94%_5.56%_0.18%_0.28%_0.03%
CNN416.185.270.897593.97%_5.56%_0.17%_0.27%_0.02%
CNN516.305.180.896194.07%_5.47%_0.20%_0.23%_0.03%
CNN616.305.630.896194.07%_5.48%_0.16%_0.27%_0.03%
CNN716.325.580.895893.99%_5.55%_0.16%_0.27%_0.03%
LSTM116.535.390.893293.90%_5.58%_0.21%_0.28%_0.03%
LSTM217.067.480.886293.97%_5.49%_0.20%_0.31%_0.02%
LSTM317.156.340.884993.72%_5.73%_0.20%_0.29%_0.04%
LSTM416.615.260.892093.86%_5.66%_0.20%_0.25%_0.03%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tang, B.; Yuan, Y.; Yang, J.; Qiu, L.; Zhang, S.; Shi, J. Predicting Blood Glucose Concentration after Short-Acting Insulin Injection Using Discontinuous Injection Records. Sensors 2022, 22, 8454. https://doi.org/10.3390/s22218454

AMA Style

Tang B, Yuan Y, Yang J, Qiu L, Zhang S, Shi J. Predicting Blood Glucose Concentration after Short-Acting Insulin Injection Using Discontinuous Injection Records. Sensors. 2022; 22(21):8454. https://doi.org/10.3390/s22218454

Chicago/Turabian Style

Tang, Baoyu, Yuyu Yuan, Jincui Yang, Lirong Qiu, Shasha Zhang, and Jinsheng Shi. 2022. "Predicting Blood Glucose Concentration after Short-Acting Insulin Injection Using Discontinuous Injection Records" Sensors 22, no. 21: 8454. https://doi.org/10.3390/s22218454

APA Style

Tang, B., Yuan, Y., Yang, J., Qiu, L., Zhang, S., & Shi, J. (2022). Predicting Blood Glucose Concentration after Short-Acting Insulin Injection Using Discontinuous Injection Records. Sensors, 22(21), 8454. https://doi.org/10.3390/s22218454

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop