
Glucose trend prediction based on continuous glucose monitoring (CGM) data is a crucial step in the implementation of an artificial pancreas (AP). A glucose trend prediction model with high accuracy in real-time can greatly improve the glycemic control effect of the artificial pancreas and effectively prevent the occurrence of hyperglycemia and hypoglycemia. In this paper, we propose an improved wavelet transform threshold denoising algorithm for the non-linearity and non-smoothness of the original CGM data. By quantitatively comparing the mean square error (MSE) and signal-to-noise ratio (SNR) before and after the improvement, we prove that the improved wavelet transform threshold denoising algorithm can reduce the degree of distortion after the smoothing of CGM data and improve the extraction effect of CGM data features at the same time. Based on this finding, we propose a glucose trend prediction model (IWT-GRU) based on the improved wavelet transform threshold denoising algorithm and gated recurrent unit. We compared the root mean square error (RMSE), mean absolute percentage error (MAPE), and coefficient of determination (R2) of Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM), Support vector regression (SVR), Gated Recurrent Unit (GRU) and IWT-GRU on the original CGM monitoring data of 80 patients for 7 consecutive days with different prediction horizon (PH). The results showed that the IWT-GRU model outperformed the other four models. At PH = 45 min, the RMSE was 0.5537 mmol/L, MAPE was 2.2147%, R2 was 0.989 and the average runtime was only 37.2 seconds. Finally, we analyze the limitations of this study and provide an outlook on the future direction of blood glucose trend prediction.
Citation: Tao Yang, Qicheng Yang, Yibo Zhou, Chuanbiao Wen. Glucose trend prediction model based on improved wavelet transform and gated recurrent unit[J]. Mathematical Biosciences and Engineering, 2023, 20(9): 17037-17056. doi: 10.3934/mbe.2023760
[1] | Lingmin Lin, Kailai Liu, Huan Feng, Jing Li, Hengle Chen, Tao Zhang, Boyun Xue, Jiarui Si . Glucose trajectory prediction by deep learning for personal home care of type 2 diabetes mellitus: modelling and applying. Mathematical Biosciences and Engineering, 2022, 19(10): 10096-10107. doi: 10.3934/mbe.2022472 |
[2] | Huawei Jiang, Tao Guo, Zhen Yang, Like Zhao . Deep reinforcement learning algorithm for solving material emergency dispatching problem. Mathematical Biosciences and Engineering, 2022, 19(11): 10864-10881. doi: 10.3934/mbe.2022508 |
[3] | Xiaotong Ji, Dan Liu, Ping Xiong . Multi-model fusion short-term power load forecasting based on improved WOA optimization. Mathematical Biosciences and Engineering, 2022, 19(12): 13399-13420. doi: 10.3934/mbe.2022627 |
[4] | Hongli Niu, Kunliang Xu . A hybrid model combining variational mode decomposition and an attention-GRU network for stock price index forecasting. Mathematical Biosciences and Engineering, 2020, 17(6): 7151-7166. doi: 10.3934/mbe.2020367 |
[5] | Zhenglong Tang, Chao Chen . Spatio-temporal information enhance graph convolutional networks: A deep learning framework for ride-hailing demand prediction. Mathematical Biosciences and Engineering, 2024, 21(2): 2542-2567. doi: 10.3934/mbe.2024112 |
[6] | Weijiu Liu . A mathematical model for the robust blood glucose tracking. Mathematical Biosciences and Engineering, 2019, 16(2): 759-781. doi: 10.3934/mbe.2019036 |
[7] | Lin Feng, Ziren Chen, Harold A. Lay Jr., Khaled Furati, Abdul Khaliq . Data driven time-varying SEIR-LSTM/GRU algorithms to track the spread of COVID-19. Mathematical Biosciences and Engineering, 2022, 19(9): 8935-8962. doi: 10.3934/mbe.2022415 |
[8] | Ping Wang, Qiaoyan Sun, Yuxin Qiao, Lili Liu, Xiang Han, Xiangguang Chen . Online prediction of total sugar content and optimal control of glucose feed rate during chlortetracycline fermentation based on soft sensor modeling. Mathematical Biosciences and Engineering, 2022, 19(10): 10687-10709. doi: 10.3934/mbe.2022500 |
[9] | Hugo Flores-Arguedas, Marcos A. Capistrán . Bayesian analysis of Glucose dynamics during the Oral Glucose Tolerance Test (OGTT). Mathematical Biosciences and Engineering, 2021, 18(4): 4628-4647. doi: 10.3934/mbe.2021235 |
[10] | Lihe Liang, Jinying Cui, Juanjuan Zhao, Yan Qiang, Qianqian Yang . Ultra-short-term forecasting model of power load based on fusion of power spectral density and Morlet wavelet. Mathematical Biosciences and Engineering, 2024, 21(2): 3391-3421. doi: 10.3934/mbe.2024150 |
Glucose trend prediction based on continuous glucose monitoring (CGM) data is a crucial step in the implementation of an artificial pancreas (AP). A glucose trend prediction model with high accuracy in real-time can greatly improve the glycemic control effect of the artificial pancreas and effectively prevent the occurrence of hyperglycemia and hypoglycemia. In this paper, we propose an improved wavelet transform threshold denoising algorithm for the non-linearity and non-smoothness of the original CGM data. By quantitatively comparing the mean square error (MSE) and signal-to-noise ratio (SNR) before and after the improvement, we prove that the improved wavelet transform threshold denoising algorithm can reduce the degree of distortion after the smoothing of CGM data and improve the extraction effect of CGM data features at the same time. Based on this finding, we propose a glucose trend prediction model (IWT-GRU) based on the improved wavelet transform threshold denoising algorithm and gated recurrent unit. We compared the root mean square error (RMSE), mean absolute percentage error (MAPE), and coefficient of determination (R2) of Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM), Support vector regression (SVR), Gated Recurrent Unit (GRU) and IWT-GRU on the original CGM monitoring data of 80 patients for 7 consecutive days with different prediction horizon (PH). The results showed that the IWT-GRU model outperformed the other four models. At PH = 45 min, the RMSE was 0.5537 mmol/L, MAPE was 2.2147%, R2 was 0.989 and the average runtime was only 37.2 seconds. Finally, we analyze the limitations of this study and provide an outlook on the future direction of blood glucose trend prediction.
Diabetes mellitus (DM) is an endocrine metabolic disease caused by the interaction of genetic, immune and environmental factors that lead to a defect in insulin secretion or a decrease in insulin sensitivity of the cells, resulting in disorders of blood glucose, carbohydrate, lipid and protein metabolism [1]. According to the World Health Organization, diabetes is one of the leading non-communicable diseases with the highest mortality rate worldwide [2]. It is one of the most common non-communicable diseases worldwide. Over the past 30 years, the prevalence of diabetes has increased annually and by 2021, there were approximately 537 million adults with diabetes worldwide [3]. According to projections by the International Diabetes Federation in 2021, the total number of diabetics worldwide will increase to 643 million by 2030 and to 783 million by 2045 [4]. The American Diabetes Association (ADA) classifies DM as type 1 diabetes, type 2 diabetes, gestational diabetes and specific types of diabetes [5]. Of these, type 1 diabetes accounts for about 5% and type 2 diabetes for 90% to 95% [6]. The main characteristics of DM are high fasting and postprandial glucose, and the common clinical manifestations are mainly polyhydramnios, polyuria, polyphagia and weight loss. Long-term elevated glucose can damage the body's cardiovascular and nervous systems, which can lead to chronic complications such as diabetic cardiovascular disease, diabetic nephropathy, diabetic retinopathy and peripheral neuropathy [7]. Similarly, recurrent episodes of hypoglycemia can easily damage the brain function of patients, accelerating the process of dementia and even brain death [8]. The disease has not yet been cured. The only way to control the progression of diabetes is through early diagnosis and timely treatment, which can be achieved by managing diet [9], reasonable insulin injections [10], active monitoring of blood glucose levels [11], etc., to maintain blood glucose levels within normal limits [12].
The artificial pancreas (also known as the closed-loop insulin delivery system) has been one of the breakthroughs pursued in the treatment of diabetes since the 19th century [13]. It consists of three main components: a continuous glucose monitoring device (CGM), an insulin pump (IP) and a closed-loop control algorithm (CA), which automatically monitors the blood glucose level of the diabetic patient through the CGM and controls the insulin pump to automatically deliver the right amount of insulin to the patient through the closed-loop control algorithm [14]. The closed-loop control algorithm is the most central and critical part of the artificial pancreas function [15]. However, most current CGM devices monitor glucose levels in interstitial fluid [16], which has a delay in real-time glucose levels in the blood, and there is a physiological lag, which requires closed-loop control algorithms to accurately predict glucose levels. With the continuous development of glucose prediction technology, numerous scholars have started to explore the performance of different types of models in glucose prediction [17,18,19,20,21,22,23,24]. Based on the published literature, two types of glucose prediction methods can be classified: physiology-based glucose prediction and data-driven glucose prediction [25,26]. The physiologically based blood glucose prediction model is a continuous dynamic model that relies on the basic physiological mechanism of the human body. However, since the regulation of the human body environment is a dynamic process, this prediction method depends on the physiological regulation mechanism of the individual. However, the blood glucose changes between different patients have significant variability, so it is difficult for the overall physiological model to guarantee the accuracy of different individuals. Moreover, due to the excessive factors to be considered in the construction of physiological models, the cost of modeling is high, and the practical application is low at present.
In recent years, with the widespread use of CGM, data-driven blood glucose prediction models have become a new hot topic of research, which usually use previous continuous blood glucose time series data to predict future blood glucose trends at some prediction horizon (PH), often using regression, machine learning and deep learning methods to build prediction models. In 2015, Georga et al. [27] used random forest algorithm and RReliefF to rank feature sets to obtain feature sets of blood glucose level, insulin dose, dietary intake, and physical activity, which were combined with regression models for blood glucose prediction. In 2019, Idriss et al. [28] proposed a deep learning neural network model (LSTMNN) based on a Long Short Term Memory Network (LSTM) layer and two hidden layers to predict blood glucose levels in diabetic patients, which predicted blood glucose levels with a mean RMSE of 12.38 mg/dL. In 2020, Teng et al. [29] proposed a prediction model using the GRU method to predict blood glucose time series data with RMSE and MAPE of 0.7612 mmol/L and 7.3427%, respectively, for a PH = 60 min. In 2021, Karim et al. [20] trained an artificial neural network with the bolus and basal insulin dosing and timing, baseline glucose levels, maximum glucose infusion rate and total carbohydrate content as parameters and predicted an RMSE of 1.72 mmol/L at PH = 120 min. Cichosz et al. [30,31] proposed that adding penalty weights could improve the clinical performance of CGM prediction models. In 2022, Koutny and Mayo [22] proposed a new low-complexity, interpretable method for blood glucose prediction using Meta-Differential Evolution to determine the parameters of the predictor, extending the prediction time by 5 min with a relative error of less than 30% for 95.8% of predicted blood glucose levels in all patients. In 2023, Yang et al. [23] proposed a short-term prediction method for blood glucose based on the Patients' Time Multiple Attention Mechanism for Diabetes (PBGTAM), with overall root mean square error of prediction, prediction accuracy and clinical accuracy of 20.57, 84.35 and 85.18%, respectively.
Due to the non-linearity and non-smoothness of CGM data, direct use of raw data to build prediction models often fails to meet the accuracy requirements. More researchers have applied relevant methods of signal analysis to pre-process CGM data in order to reduce its non-stationarity and improve the prediction accuracy. Rabby et al. [32] and Xu et al. [21] used Kalman filtering to smooth the CGM data in 2021 and 2022, respectively, but the Kalman filtering method can only process linear processes and the signal noise must obey a Gaussian distribution, and the smoothing effect of the Kalman filtering method on the data will be reduced when the nonlinear characteristics of the actual system are slightly stronger or the noise characteristics deviate from the Gaussian distribution. Carrillo-Moreno et al. [33] used a sliding average filter to deal with noise in CGM data. The advantage of this method is that it is simple to operate, but the size of the sliding window will have a large impact on the smoothing effect, and this filtering method will lose some of its accuracy when the real value within the sliding window changes a lot, and the filtering result is close to the average expectation of the real value. The larger the window is, the more serious the lag phenomenon is. The wavelet transform threshold denoising algorithm is not only able to retain more information about the signal at high frequencies, but also the signal at low frequencies is almost invisible to noise interference, and has a better smoothing effect for non-linear non-smooth signals. In 2022, Liang [34] and Peng [35] used the traditional wavelet transform threshold denoising algorithm to smooth CGM data, but did not study in depth the way the threshold of wavelet transform was selected and the influence caused by the threshold function on the smoothing effect of CGM data.
Therefore, in order to further improve the prediction effect of the blood glucose trend prediction model, this paper improves the wavelet transform threshold selection method and threshold function, improves the data smoothing effect while retaining the original characteristics of CGM data to the greatest extent so that the smoothed CGM data can be input to the GRU model to effectively improve the accuracy of blood glucose prediction and combines the characteristics of the fast running speed of the GRU model to build a blood glucose trend prediction model with high real-time capability and high accuracy. The rest of the paper is structured as follows: Section 2 introduces the data sources and relevant theoretical background, including the basic principles of wavelet transform and GRU. Section 3 describes in detail the improved method of using the wavelet transform threshold denoising algorithm and compares the SNR and MSE of the algorithm before and after the improvement applied to CGM data. Section 4 provides a detailed description of the data pre-processing steps and the parameters of the IWT-GRU model and the remaining control models. Section 5 compares the strengths and weaknesses of the IWT-GRU model and the other four models by means of model performance measures and verifies the actual performance of the models using Clarke Error Grid Analysis (CEGA). Section 6 concludes the research presented in this paper and provides an outlook on future developments in the field of glucose trend prediction.
The wavelet transform inherits the idea of localization of the short-time Fourier transform, while overcoming the disadvantage of the window size not varying with frequency, providing a "time-frequency" window that changes with frequency, making it an ideal tool for time-frequency analysis and processing of signals [36]. The wavelet transform is highly de-data-dependent, concentrating the energy of the signal on some of the larger wavelet coefficients in the wavelet domain, while the white noise contained in the orthogonal basis of the transform is white noise, the wavelet transform distributes the noise energy over most of the unfolded coefficients in the wavelet domain. Therefore, the wavelet coefficients with larger amplitude after wavelet transform are generally dominated by the signal, while the coefficients with smaller amplitude are largely noise. Therefore, the coefficients with larger amplitude can be retained and the coefficients with smaller amplitude can be set to zero to obtain the estimated wavelet coefficients, and then the signal can be reconstructed to achieve the purpose of removing noise. Take the three-layer wavelet transform as an example, as shown in Figure 1 The signal is decomposed into three layers to obtain four wavelet coefficients (A3, D3, D2, D1). Approximation is the low frequency component of the signal and A1 is the approximate wavelet coefficient. Detail is the high frequency component of the signal and D1, D2 and D3 are the detail wavelet coefficients. After the three-layer decomposition, the amplitude of the wavelet coefficients of the original signal is larger than that of the noise and the number of signal wavelet coefficients is much smaller than that of the noise wavelet coefficients [37].
The first step of wavelet threshold denoising is to select a suitable wavelet basis function and the number of decomposition layers to carry out wavelet decomposition of the noisy signal. The commonly used wavelets are mainly dbN wavelet, symN wavelet, coifN wavelet, Haar wavelet, biorNr.Nd wavelet, etc. The number of decomposition layers is generally selected as 3−5 layers. The second step is to select a suitable threshold λ and threshold function. If the selected threshold is too large, some of the original signal information will be lost, resulting in distortion of the reconstructed signal; if the selected threshold is too small, too much noise will be retained, resulting in poor denoising effect. The common threshold functions are mainly the hard threshold function (Eq (1)), soft threshold function (Eq (2)) and Garrote threshold function (Eq (3)). The third step is to perform a discrete wavelet transform on the processed wavelet coefficients to obtain the reconstructed denoised signal.
wi,k={wi,k|wi,k|≥λ0|wi,k|<λ | (1) |
wi,k={[sgn(wi,k)](|wi,k|−λ)|wi,k|≥λ0|wi,k|<λ | (2) |
wi,k={wi,k−λ2wi,k|wi,k|≥λ0|wi,k|<λ | (3) |
The flow of the wavelet transform threshold denoising algorithm is shown in Figure 2.
Artificial Neural Networks (ANNs) were inspired by the neural networks in the biological brain. ANNs consist of a large number of neural nodes that model the complex relationships between input data and output signals through connections between nerves [38]. In general, an ANN consists of an input layer, a hidden layer, and an output layer, and the parameters of data modelling are the structural information of the nerve and the weight values between the two nodes connected. In 1982, Hopfiled [39] first proposed Recurrent Neural Networks (RNNs), which, compared to traditional artificial neural networks, also connect neurons in a single layer to form a "loop" structure between neurons in a single layer, enabling the sharing of weights between layers, and the recurrent neural network transmits information through the "loop" structure, realizing the information sharing between layers. It can be used to model sequential data such as text statements, time sequences and biosignal sequences. If the input vector is X=(x0,x1,x2…,xp−1)T, then the states of the layers at time t are
Ht=fhh(WhhHt−1+WxhXt) | (4) |
Yt=fhy(WhyHt) | (5) |
Where fhh and fhy are the activation functions from the hidden layer to the hidden layer and from the hidden layer to the output layer, respectively, and are generally Sigmoid functions or tanh functions; Wxh, Whh, and Why are connection weight matrices between input layer and hidden layer, hidden layer and hidden layer, and hidden layer and output layer, respectively. The structure of the RNN is shown in Figure 3(a).
Since the basic RNN uses the back-propagation through time (BPTT) algorithm for parameter tuning, there is the problem of gradient disappearance or gradient explosion, and the basic recurrent neural network is unable to remember information with long time steps. Based on this, two variants of the basic RNN emerged: the LSTM and the GRU, whose variations are shown schematically in Figure 3(d).
The LSTM uses long and short memory units to replace the activation functions in the basic RNN hidden layer. In addition to the original activation functions, the LSTM contains input gates, forgetting gates and output gates made up of Sigmoid functions, as well as a series of multiplication and addition operations. The input gate controls new information, Xt, merged into long-term memory, the forgetting gate controls the extent to which past unit states Ct−1. The input gate and the forgetting gate together determine the extent to which the cell state (Ct) updates, and the output gate is based on the cell state Ct and controls the hidden layer of Ht. The Sigmoid function transforms the data obtained from the linear transformation of the hidden layer information at the previous time point and the current input information into a numerical output between 0 and 1, which controls the degree of information memory on the one hand and solves the gradient disappearance and gradient explosion problems caused by the BP algorithm on the other. The internal structure of the LSTM is shown in Figure 3(b).
LSTM is widely used in the prediction of time series problems because of its good prediction performance, but its complex internal structure also leads to the slowdown of model training speed. GRU is a variant of LSTM, which does not have an explicit cell state, but uses a reset gate to implement the role of the forgetting gate and input gate in LSTM, and an update gate to control the update of the hidden layer state. This internal structure of GRU makes it inherit the advantages of LSTM on the one hand, and reduce the parameters required for model training on the other hand, thus reducing the model training time. The GRU hidden layer output is calculated as follows:
Rt=σ(WrxXt+WrhHt−1) | (6) |
Zt=σ(WzxXt+WzhHt−1) | (7) |
−Ht=tanh(WshRtHt−1+WsxXt) | (8) |
Ht=(1−Zt)Ht−1+Zt−Ht | (9) |
where Xt is the input vector at moment t, Rt is the reset guys vector at moment t, Zt is the update gate vector at time t, Ht is the hidden layer output vector at time t, −Ht is the updated candidate vector and Wsh, Wsx, Wzx, Wzh, Wrx and Wrh are the weight matrices in each connection vector, respectively. σ is the Sigmoid function. The internal structure of the GRU is shown in Figure 3(c).
The data for this study was obtained from a silicon-based kinetic glucose monitor that acquired patients' blood glucose data at 5-minute intervals, and raw CGM data from 80 patients were selected over 7 consecutive days. Table 1 presents baseline data for patients. Data are represented as means ± standard deviation. Figure 4 shows the blood glucose change curve (2016 blood glucose values) for a patient over a consecutive 7-day period.
Index | Sample size (n) | Statistic |
Gender | Male (54) | / |
Female (26) | / | |
Age Disease duration Height (cm) |
80 80 80 |
56.144 ± 8.333 8.5 ± 5.94 162.4 ± 8.26 |
Weight (kg) | 80 | 63.55 ± 10.82 |
BMI (kg·m-2) | 80 | 29.953 ± 3.621 |
The main wavelet denoising thresholds currently in use are based on unbiased likelihood estimation (rigsure), fixed thresholds (sqtwolog), heursure thresholds (heursure) and minimaxi thresholds (minimaxi). The choice of a suitable threshold is important for subsequent signal processing. If the threshold is chosen too small, there is a residual noise removal, and if the threshold is chosen too large, part of the signal is filtered out. Since the wavelet coefficient of noise decreases with the increase of decomposition layers, the threshold value of different decomposition layers should decrease with the increase of decomposition layers when denoising the signal. In this paper, we propose the threshold layer selection method, and the threshold layer selection equation is:
λ=σn√2lnN/ln(e+2j). | (10) |
In the above equation, N is the number of corresponding wavelet transform coefficients, j is the corresponding number of layers of the wavelet transform and σn is the variance contained in the noise, and since the noise variance is not known in practical use, it can be estimated from the wavelet decomposition coefficients of the first layer, i.e.,
σn=median(wi,k)/0.6745 | (11) |
The three traditional thresholding functions are often used in signal processing, but each of them has certain shortcomings. When using hard thresholding, the signal is better marginalized, but the function itself is discontinuous and may lead to unwanted oscillations in the reconstructed signal; soft thresholding has good continuity and the denoising result is smoother, but when the wavelet coefficients are large, there is a fixed deviation between the estimated wavelet coefficients and the original coefficients, resulting in a loss of the high frequency part of the signal, leading to the approximation degree of the reconstructed signal being poor compared with the original signal, which easily produces distortion. The Garrote thresholding function combines the advantages of soft and hard thresholding functions, with good signal continuity, and the fixed deviation tends to zero when the wavelet coefficients are large, effectively overcoming the shortcomings of soft and hard thresholding functions. However, it ignores the characteristics of noise decreases with increasing scale under wavelet transform, and for mutating signals, its denoising effect is still insufficient. The three traditional thresholding functions are shown in Figure 5.
As the three types of traditional threshold functions mentioned above all have their own shortcomings, a new threshold function is constructed under the premise of fully considering the respective advantages of these three types of functions, with the specific expression shown in Eq (12). The improved image of the threshold function is shown in Figure 6. By introducing the adjustment factors α and β, it retains the continuity of the soft threshold function and can rapidly approximate the hard threshold function, and the approximation speed of the function can be reasonably controlled by controlling the values of α and β, which is more applicable to different signals.
wi,k={[sgn(wi,k)](|wi,k|−λ/β√|wi,k|α−|λ|α+1)|wi,k|≥λ0|wi,k|<λ | (12) |
In this paper, the CGM data of a patient was randomly selected from 80 patients, and three traditional threshold functions and the improved threshold function were adopted to denoise the signal. dB10 wavelets were selected to do 3-layer wavelet decomposition on the CGM signal, and the adjustment factors α and β were chosen as 0.5 and 2 respectively. The traditional three threshold functions adopt fixed threshold, and the improved threshold function adopts hierarchical threshold selection method. The signal-noise ratio (SNR) and mean square error (MSE) are used to quantify and compare their denoising effects. The SNR is an important measure of the denoising effect of the signal, and the higher the SNR, the better the signal extraction effect. The smaller the MSE between the denoised signal and the original signal, the smaller the distortion, the more complete the signal extraction and the better the denoising effect. The equations for SNR and MSE are shown in Eqs (13) and (14), where s(i) represents the original signal and y(i) represents the denoised signal.
SNR=10log(∑ni=1s2(i)/∑ni=1(y(i)−s(i)2)) | (13) |
MSE=√(∑ni=1((y(i)−s(i)2))/n | (14) |
When α and β are selected as 0.5 and 2 respectively, the model SNR and mean square error can obtain the best values. The relationship between α and β values and SNR and mean square error is shown in Figures 7 and 8. The effect of denoising the CGM data by the four methods is shown in Figure 9 and the comparison of SNR and MSE is shown in Table 1. From Table 2, it can be seen that the improved wavelet transform threshold denoising algorithm has a SNR of 24.8135 and an MSE of 0.2309, and the denoising effect is better than the three traditional methods.
Hard threshold | Soft threshold | Garrote threshold | Improved threshold | |
SNR | 23.3562 | 24.5873 | 24.5369 | 24.8135 |
MSE | 0.3087 | 0.2642 | 0.2955 | 0.2309 |
As the Dynamic Glucose Monitor from Silicon Motion has a data storage function, it does not have individual missing values if not for sensor damage, so there is no need to fill in the data and exclude user data that has not been continuously monitored for 7 days.
Due to the non-linearity and non-smoothness of the raw CGM data, the prediction accuracy of the model training directly using the raw data often fails to achieve good results, so the data needs to be smoothed. In this paper, the improved wavelet transform threshold denoising algorithm described in Section 3 is used to smooth the raw CGM data.
As the neural network model frameworks all have adapted input data scales, different input sequence lengths have a significant impact on the prediction results. Therefore, to better train the model, a linear normalization method is first used to scale the data to the (0, 1) interval. The normalized data is then divided using the data sliding window method, thus converting the dataset into a supervised learning type of data sample and enabling flexible transformation of data lengths. By adjusting the size of different windows, the model can use feature time series data with a previous time length of 5*m (min) to predict future blood glucose fluctuations with a total time length of 5*n (min).
In this paper, the RNN, LSTM, SVR, GRU and IWT-GRU algorithms are selected for modelling and training respectively, in which, except for the IWT-GRU model, the data smoothing algorithms of all the models adopt the general wavelet threshold denoising algorithm, i.e., the CGM data are smoothed using fixed threshold and soft threshold functions and the IWT-GRU model adopts the improved wavelet transform proposed in this paper. For the threshold denoising algorithm for data smoothing, wavelet basis is selected as sym8 and the number of decomposition layers is 3. In addition, the IWT-GRU model chooses MAE as the loss function, the learning rate is 0.0001, the optimizer chooses RMSProp, the number of samples included in each training is 64 and the maximum number of iterations is 200. The structure and parameters of the RNN, LSTM and GRU models are the same as those of IWT-GRU, and the SVR uses radial basis neural network as the kernel, C = 10, gamma = 0.1. The overall construction process of the blood glucose trend prediction model in this paper is shown in Figure 10.
In this paper, root mean square error (RMSE), mean absolute percentage error (MAPE), coefficient of determination (R2) and running time are used as the evaluation metrics for model performance. The running environment of the code is Windows 10 operating system, and the CPU parameters are i5-11260H@2.60GHz. RMSE, MAPE and R2 are calculated as shown in the following equations:
RMSE=√1n∑ni=1(xi− xi)2 | (15) |
MAPE=100n∑ni=1|xi− xixi| | (16) |
R2=1−RMSE2∑ni=1(xi−−xi)2 | (17) |
where n is the number of samples and xi is the actual value of the ith sample, the xi is the predicted value of the ith sample and −xi is the mean value of the original data. In statistics, the smaller the value of RMSE and MAPE, the better the model predicts. R2 is the ratio of the total regression sum of squares to the total deviation sum of squares, ranging from 0 to 1. Closer to 1 means a more significant fit of the model and better model prediction.
Tables 3−5 show the mean prediction performance of the IWT-GRU model for different prediction horizon when the length of the input data sequence is 30, 60 and 90 min, respectively. From Tables 3−5, it can be seen that for different input data sequence lengths, the model performance is better when the prediction horizon is 45 min than for the other prediction horizon. As the length of the input data sequence increases, the performance of the model at PH = 45 min shows a higher and then lower state. Overall, the IWT-GRU model performs best when the input sequence data length is 60 min and the prediction horizon is 45 min. Table 6 shows the comparison of the mean prediction results of the five models at an input sequence data length of 60 min and PH = 45 min. From Table 6, it can be seen that the GRU model performs best in terms of performance and running time under the same pre-processing method, while the performance of the IWT-GRU model processed with the improved wavelet transform threshold denoising algorithm has a significant improvement over the performance of the conventional GRU model, and the running time is only 0.9 seconds slower than that of the GRU model.
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15 min | 0.8489 | 4.1325 | 0.895 |
30 min 45 min 60 min |
0.6416 0.5766 0.9754 |
2.8462 2.5674 4.3012 |
0.905 0.956 0.872 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15 min | 0.8753 | 4.1358 | 0.899 |
30 min 45 min 60 min |
0.5984 0.5537 0.6833 |
2.7763 2.2147 4.0258 |
0.924 0.989 0.898 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15min | 0.8887 | 4.1401 | 0.857 |
30min 45min 60min |
0.6529 0.5896 0.9824 |
2.9563 2.3411 4.5986 |
0.899 0.966 0.870 |
Model | RMSE (mmol/L) | MAPE (%) | R2 | Running time (s) |
RNN | 1.2126 | 3.8743 | 0.854 | 45.1 |
LSTM SVR GRU IWT-GRU |
1.1533 1.3754 0.8423 0.5537 |
3.9866 4.2574 3.7549 2.2147 |
0.833 0.872 0.897 0.989 |
48.6 28.9 36.3 37.2 |
In order to demonstrate the validity and accuracy of the model in a clinical setting, the Clarke Error Grid Analysis (CEGA), which is commonly used in clinical practice, was chosen to assess the difference between actual and predicted blood glucose values. This method has a target blood glucose control value of 3.9−10.0 mmol/L, with interventions required below 3.9 mmol/L or above 13.3 mmol/L. After converting the blood glucose values to mg/dL, a Clark error analysis grid was drawn using the CGM-value (without filtering) and the model predicted values as the horizontal and vertical axes of the grid, respectively. Zone A is the clinically accurate zone, representing ± 20% difference from the reference value; zone B is the clinically acceptable zone, located above or below zone A, representing blood glucose values above 20% of the range relative to the reference value; zones C, D and E are the clinically incorrect zones. The data is not desirable. The more blood glucose values falling in zones A and B, the closer the model prediction is to the measured value, and the higher the prediction accuracy is. Figure 11(a)−(e) is the results of CEGA for predicting the change of blood glucose in the next 60 min for a patient using SVR, LSTM, RNN, IWT-GRU and GRU models respectively. The prediction results of the five algorithms accounted for 84.38, 92.64, 89.33, 98.69 and 95.26% in region A, respectively, which shows that the IWT-GRU model proposed in this paper also has the best prediction accuracy compared to other models in actual clinical outcomes.
In this paper, an improved wavelet transform threshold denoising algorithm was proposed to smooth CGM data for its non-linearity and non-smoothness, and a novel blood glucose trend prediction model was constructed by combining GRU. The performance of the algorithm was validated on CGM data from 80 diabetic patients, and by comparing the RMSE, MAPE and R2of SVR, RNN, LSTM, GRU and IWT-GRU models, the IWT-GRU model proposed in this paper was proved to have the best prediction performance under the same parameter conditions. Meanwhile, the clinical performance of the model was verified by Clarke Error Grid Analysis (CEGA), which improved the prediction accuracy of the model with reduced running time and can provide a reference for the research of artificial pancreas closed-loop control algorithm.
However, there are some limitations in this study, first, the prediction effect of data collected by different CGM devices was not compared in this paper, and second, the diabetic patients selected in this paper were all type 2 diabetic patients, and the model needs to be optimized by CGM data from other types of diabetic patients in the future. For the improved wavelet transform threshold denoising algorithm, the values of moderators α and β were selected based on mathematical experience in this paper. In the future, we need to further explore in depth what is the relationship between the values of different moderators α and β and different prediction horizon, how their combination affects the prediction performance of the model, and the IWT-GRU model can be added with features such as exercise, diet and medication status for multi-factor prediction. The model's predictive accuracy can be further improved.
In conclusion, all experiments are for practical clinical application, and the future glucose trend prediction model based on CGM data will still be optimized in the direction of high performance and high efficiency. Meanwhile, with the gradual development of non-contact blood glucose monitoring and non-invasive blood glucose monitoring technologies, the use of a wider range of physiological parameters and data assimilation techniques to establish algorithmic models for multimodal sensor data fusion will be another new research direction for blood glucose trend prediction in the future.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
The study was supported by the Key Research and Development Support Plan of Chengdu Science and Technology Bureau (2019-YF09-00185-SN): Demonstration and promotion of diabetes graded linkage management technology and Application based on TCM cloud health platform, and Joint Innovation Fundation of JIICM (2021IR020): Study on remote MDT diagnosis, treatment and management mode of Type 2 diabetes mellitus with Integrated Traditional Chinese and Western Medicine.
The authors declare there is no conflict of interest.
[1] | T. T. Zhou, The Discovery and Mechanism of Anti-T2DM Lead Structure Based on Pancreatic β Cell Function Improvement/Liver Gluconeogenic Inhibition Strategy (in Chinese), Ph.D thesis, University of Chinese Academy of Sciences (Shanghai Institute of Materia Medica, CAS), 2017. Available from: https://kns.cnki.net/kcms/detail/detail.aspx?FileName = 1017820618.nh & DbName = CDFD2018. |
[2] |
R. Williams, S. Karuranga, B. Malanda, P. Saeedi, A. Basit, S. Besancon, et al., Global and regional estimates and projections of diabetes-related health expenditure: Results from the International Diabetes Federation Diabetes Atlas, 9th edition, Diabetes Res. Clin. Pract., 162 (2020), 108072. https://doi.org/10.1016/j.diabres.2020.108072 doi: 10.1016/j.diabres.2020.108072
![]() |
[3] |
M. Khan, M. J. Hashim, J. K. King, R. D. Govender, H. Mustafa, J. Al Kaabi, Epidemiology of type 2 diabetes - global burden of disease and forecasted trends, J. Epidemiol. Global Health, 10 (2020), 107-111. https://doi.org/10.2991/jegh.k.191028.001 doi: 10.2991/jegh.k.191028.001
![]() |
[4] |
N. H. Cho, J. E. Shaw, S. Karuranga, Y. Huang, R. F. J. Da, A. W. Ohlrogge, et al., IDF diabetes atlas: global estimates of diabetes prevalence for 2017 and projections for 2045, Diabetes Res. Clin. Pract., 138 (2018), 271-281. https://doi.org/10.1016/j.diabres.2018.02.023 doi: 10.1016/j.diabres.2018.02.023
![]() |
[5] |
Z. Luo, G. Fabre, V. G. Rodwin, Meeting the challenge of diabetes in China, Int. J. Health Policy Manage., 9 (2020), 47-52. https://doi.org/10.15171/ijhpm.2019.80 doi: 10.15171/ijhpm.2019.80
![]() |
[6] |
N. A. Elsayed, G. Aleppo, V. R. Aroda, R. R. Bannuru, F. M. Brown, D. Bruemmer, et al., 2. classification and diagnosis of diabetes: standards of care in diabetes-2023, Diabetes Care, 46 (2023), S19-S40. https://doi.org/10.2337/dc23-S002 doi: 10.2337/dc23-S002
![]() |
[7] |
H. Sun, P. Saeedi, S. Karuranga, M. Pinkepank, K. Ogurtsova, B. B. Duncan, et al., IDF Diabetes Atlas: global, regional and country-level diabetes prevalence estimates for 2021 and projections for 2045, Diabetes Res. Clin. Pract., 183 (2022), 109119. https://doi.org/10.1016/j.diabres.2021.109119 doi: 10.1016/j.diabres.2021.109119
![]() |
[8] |
J. Osorio, Severe hypoglycemia associated with risk of vascular events and death, Nat. Rev. Cardiol., 7 (2010), 666. https://doi.org/10.1038/nrcardio.2010.176 doi: 10.1038/nrcardio.2010.176
![]() |
[9] | S. J. Dicken, R. L. Batterham, The role of diet quality in mediating the association between ultra-processed food intake, obesity and health-related outcomes: a review of prospective cohort studies, Nutrients, 14 (2021). https://doi.org/10.3390/nu14010023 |
[10] |
A. Consoli, G. Formoso, Patient perceptions of insulin therapy in diabetes self-management with insulin injection devices, Acta Diabetol., 60 (2023), 705-710. https://doi.org/10.1007/s00592-023-02054-7 doi: 10.1007/s00592-023-02054-7
![]() |
[11] |
S. Reddy, C. C. Wu, A. Jose, J. L. Hsieh, S. D. Rautela, Personalized virtual care using continuous glucose monitoring in adults with type 2 diabetes treated with less intensive therapies, Clin. Diabetes, 41 (2023), 452-457. https://doi.org/10.2337/cd22-0128 doi: 10.2337/cd22-0128
![]() |
[12] | A. T. Reenberg, T. K. S. Ritschel, B. Dammann, J. B. Jørgensen, High-performance uncertainty quantification in large-scale virtual clinical trials of closed-loop diabetes treatment, in 2022 American Control Conference (ACC), (2022), 1367-1372. https://doi.org/10.23919/ACC53348.2022.9867234 |
[13] |
J. Huang, A. M. Yeung, A. Y. Dubord, H. Wolpert, P. G. Jacobs, W. A. Lee, et al., Diabetes technology meeting 2022, J. Diabetes Sci. Technol., 17 (2023), 550757959. https://doi.org/10.1177/19322968221148743 doi: 10.1177/19322968221148743
![]() |
[14] |
D. L. Rodriguez-Sarmiento, F. Leon-Vargas, M. Garcia-Jaramillo, Artificial pancreas systems: experiences from concept to commercialisation, Expert Rev. Med. Devices, 19 (2022), 877-894. https://doi.org/10.1080/17434440.2022.2150546 doi: 10.1080/17434440.2022.2150546
![]() |
[15] |
S. L. Kang, Y. N. Hwang, J. Y. Kwon, S. M. Kim, Effectiveness and safety of a model predictive control (MPC) algorithm for an artificial pancreas system in outpatients with type 1 diabetes (T1D): systematic review and meta-analysis, Diabetol. Metab. Syndr., 14 (2022), 187. https://doi.org/10.1186/s13098-022-00962-2 doi: 10.1186/s13098-022-00962-2
![]() |
[16] | L. N. Zhang, T. Y. Li, L. X. Guo, Q. Pan, Clinical progress and future prospect of continuous glucose monitoring (in Chinese), Chin. J. Clin. Healthcare, 25 (2022), 303-309. Available from: https://kns.cnki.net/kcms/detail/detail.aspx?FileName = LZBJ202203003 & DbName = CJFQ2022. |
[17] |
J. P. Anderson, J. R. Parikh, D. K. Shenfeld, V. Ivanov, C. Marks, B. W. Church, et al., Reverse engineering and evaluation of prediction models for progression to type 2 diabetes: an application of machine learning using electronic health records, J. Diabetes Sci. Technol., 10 (2015), 6-18. https://doi.org/10.1177/1932296815620200 doi: 10.1177/1932296815620200
![]() |
[18] |
K. Saiti, M. Macas, L. Lhotska, K. Stechova, P. Pithova, Ensemble methods in combination with compartment models for blood glucose level prediction in type 1 diabetes mellitus, Comput. Methods Programs Biomed., 196 (2020), 105628. https://doi.org/10.1016/j.cmpb.2020.105628 doi: 10.1016/j.cmpb.2020.105628
![]() |
[19] |
F. Tena, O. Garnica, J. Lanchares, J. I. Hidalgo, Ensemble models of cutting-edge deep neural networks for blood glucose prediction in patients with diabetes, Sensors, 21 (2021), 7090. https://doi.org/10.3390/s21217090 doi: 10.3390/s21217090
![]() |
[20] |
R. Karim, I. Vassanyi, I. Kosa, Improved methods for mid-term blood glucose level prediction using dietary and insulin logs, Medicina, 57 (2021), 676. https://doi.org/10.3390/medicina57070676 doi: 10.3390/medicina57070676
![]() |
[21] |
H. Xu, S. Bao, X. Zhang, S. Liu, W. Jing, Y. Ji, Blood glucose prediction method based on particle swarm optimization and model fusion, Diagnostics, 12 (2022), 3062. https://doi.org/10.3390/diagnostics12123062 doi: 10.3390/diagnostics12123062
![]() |
[22] |
T. Koutny, M. Mayo, Predicting glucose level with an adapted branch predictor, Comput. Biol. Med., 145 (2022), 105388. https://doi.org/10.1016/j.compbiomed.2022.105388 doi: 10.1016/j.compbiomed.2022.105388
![]() |
[23] |
G. Yang, S. Liu, Y. Li, L. He, Short-term prediction method of blood glucose based on temporal multi-head attention mechanism for diabetic patients, Biomed. Signal Process. Control, 82 (2023), 104552. https://doi.org/10.1016/j.bspc.2022.104552 doi: 10.1016/j.bspc.2022.104552
![]() |
[24] |
Z. Nie, M. Rong, K. Li, Blood glucose prediction based on imaging photoplethysmography in combination with machine learning, Biomed. Signal Process. Control, 79 (2023), 104179. https://doi.org/https://doi.org/10.1016/j.bspc.2022.104179 doi: 10.1016/j.bspc.2022.104179
![]() |
[25] | S. Oviedo, J. Vehi, R. Calm, J. Armengol, A review of personalized blood glucose prediction strategies for T1DM patients, Int. J. Numer. Methods Biomed. Eng., 33 (2017). https://doi.org/10.1002/cnm.2833 |
[26] |
V. Felizardo, N. M. Garcia, N. Pombo, I. Megdiche, Data-based algorithms and models using diabetics real data for blood glucose and hypoglycaemia prediction - a systematic literature review, Artif. Intell. Med., 118 (2021), 102120. https://doi.org/10.1016/j.artmed.2021.102120 doi: 10.1016/j.artmed.2021.102120
![]() |
[27] |
E. I. Georga, V. C. Protopappas, D. Polyzos, D. I. Fotiadis, Evaluation of short-term predictors of glucose concentration in type 1 diabetes combining feature ranking with regression models, Med. Biol. Eng. Comput., 53 (2015), 1305-1318. https://doi.org/10.1007/s11517-015-1263-1 doi: 10.1007/s11517-015-1263-1
![]() |
[28] | T. E. Idriss, A. Idri, I. Abnane, Z. Bakkoury, Predicting blood glucose using an LSTM neural network, in 2019 Federated Conference on Computer Science and Information Systems (FedCSIS), (2019), 35-41. https://doi.org/10.15439/2019F159 |
[29] | J. L. Teng, Z. J. Rong, Y. Xu, B. B. Dan, Study on blood glucose prediction method based on GRU network (in Chinese), Comput. Appl. Software, 37 (2020), 107-112. Available from: https://kns.cnki.net/kcms/detail/detail.aspx?FileName = JYRJ202010018 & DbName = CJFQ2020. |
[30] |
S. L. Cichosz, T. Kronborg, M. H. Jensen, O. Hejlesen, Penalty weighted glucose prediction models could lead to better clinically usage, Comput. Biol. Med., 138 (2021), 104865. https://doi.org/10.1016/j.compbiomed.2021.104865 doi: 10.1016/j.compbiomed.2021.104865
![]() |
[31] |
S. L. Cichosz, M. H. Jensen, O. Hejlesen, Short-term prediction of future continuous glucose monitoring readings in type 1 diabetes: development and validation of a neural network regression model, Int. J. Med. Inf., 151 (2021), 104472. https://doi.org/10.1016/j.ijmedinf.2021.104472 doi: 10.1016/j.ijmedinf.2021.104472
![]() |
[32] |
M. F. Rabby, Y. Tu, M. I. Hossen, I. Lee, A. S. Maida, X. Hei, Stacked LSTM based deep recurrent neural network with kalman smoothing for blood glucose prediction, BMC Med. Inf. Decis. Making, 21 (2021), 101. https://doi.org/10.1186/s12911-021-01462-5 doi: 10.1186/s12911-021-01462-5
![]() |
[33] |
J. Carrillo-Moreno, C. Pérez-Gandía, R. Sendra-Arranz, G. García-Sáez, M. E. Hernando, Á. Gutiérrez, Long short-term memory neural network for glucose prediction, Neural Comput. Appl., 33 (2021), 4191-4203. https://doi.org/10.1007/s00521-020-05248-0 doi: 10.1007/s00521-020-05248-0
![]() |
[34] | C. Liang, Study on Methods of Blood Glucose Trend Prediction Based on Time Series Data (in Chinese), Master's thesis, Guilin University of Electronic Technology, 2022. Available from: https://kns.cnki.net/kcms/detail/detail.aspx?FileName = 1022783447.nh & DbName = CMFD2023. |
[35] | X. L. Peng, Blood Glucose Prediction and Hypoglycemia Warning Evaluation Based on LSTM-GRU Model (in Chinese), Master's thesis, School of Henan University, 2022. Available from: https://kns.cnki.net/kcms/detail/detail.aspx?FileName = 1022688198.nh & DbName = CMFD2023. |
[36] |
F. Uesugi, Novel image processing method inspired by wavelet transform, Micron, 168 (2023), 103442. https://doi.org/10.1016/j.micron.2023.103442 doi: 10.1016/j.micron.2023.103442
![]() |
[37] | J. E. Oh, W. T. Kim, H. J. Sim, A. B. Abu, H. J. Lee, J. Y. Lee, Fault diagnosis using wavelet transform method for random signals, J. Korean Soc. Precis. Eng., 22 (2005), 80-89. Available from: https://www.dbpia.co.kr/Journal/articleDetail?nodeId = NODE00855112. |
[38] |
J. Zhao, P. Xu, X. Liu, X. Ji, M. Li, D. Sooranna, et al., Application of machine learning methods for the development of antidiabetic drugs, Curr. Pharm. Des., 28 (2022), 260-271. https://doi.org/10.2174/1381612827666210622104428 doi: 10.2174/1381612827666210622104428
![]() |
[39] |
J. J. Hopfield, Neural networks and physical systems with emergent collective computational abilities, PNAS, 79 (1982), 2554-2558. https://doi.org/10.1073/pnas.79.8.2554 doi: 10.1073/pnas.79.8.2554
![]() |
1. | Rubi Thapa, Sachin Poudel, Katarzyna Krukiewicz, Anil Kunwar, A topical review on AI-interlinked biodomain sensors for multi-purpose applications, 2024, 227, 02632241, 114123, 10.1016/j.measurement.2024.114123 | |
2. | Tomasz Wasilewski, Wojciech Kamysz, Jacek Gębicki, AI-Assisted Detection of Biomarkers by Sensors and Biosensors for Early Diagnosis and Monitoring, 2024, 14, 2079-6374, 356, 10.3390/bios14070356 |
Index | Sample size (n) | Statistic |
Gender | Male (54) | / |
Female (26) | / | |
Age Disease duration Height (cm) |
80 80 80 |
56.144 ± 8.333 8.5 ± 5.94 162.4 ± 8.26 |
Weight (kg) | 80 | 63.55 ± 10.82 |
BMI (kg·m-2) | 80 | 29.953 ± 3.621 |
Hard threshold | Soft threshold | Garrote threshold | Improved threshold | |
SNR | 23.3562 | 24.5873 | 24.5369 | 24.8135 |
MSE | 0.3087 | 0.2642 | 0.2955 | 0.2309 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15 min | 0.8489 | 4.1325 | 0.895 |
30 min 45 min 60 min |
0.6416 0.5766 0.9754 |
2.8462 2.5674 4.3012 |
0.905 0.956 0.872 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15 min | 0.8753 | 4.1358 | 0.899 |
30 min 45 min 60 min |
0.5984 0.5537 0.6833 |
2.7763 2.2147 4.0258 |
0.924 0.989 0.898 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15min | 0.8887 | 4.1401 | 0.857 |
30min 45min 60min |
0.6529 0.5896 0.9824 |
2.9563 2.3411 4.5986 |
0.899 0.966 0.870 |
Model | RMSE (mmol/L) | MAPE (%) | R2 | Running time (s) |
RNN | 1.2126 | 3.8743 | 0.854 | 45.1 |
LSTM SVR GRU IWT-GRU |
1.1533 1.3754 0.8423 0.5537 |
3.9866 4.2574 3.7549 2.2147 |
0.833 0.872 0.897 0.989 |
48.6 28.9 36.3 37.2 |
Index | Sample size (n) | Statistic |
Gender | Male (54) | / |
Female (26) | / | |
Age Disease duration Height (cm) |
80 80 80 |
56.144 ± 8.333 8.5 ± 5.94 162.4 ± 8.26 |
Weight (kg) | 80 | 63.55 ± 10.82 |
BMI (kg·m-2) | 80 | 29.953 ± 3.621 |
Hard threshold | Soft threshold | Garrote threshold | Improved threshold | |
SNR | 23.3562 | 24.5873 | 24.5369 | 24.8135 |
MSE | 0.3087 | 0.2642 | 0.2955 | 0.2309 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15 min | 0.8489 | 4.1325 | 0.895 |
30 min 45 min 60 min |
0.6416 0.5766 0.9754 |
2.8462 2.5674 4.3012 |
0.905 0.956 0.872 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15 min | 0.8753 | 4.1358 | 0.899 |
30 min 45 min 60 min |
0.5984 0.5537 0.6833 |
2.7763 2.2147 4.0258 |
0.924 0.989 0.898 |
PH | RMSE (mmol/L) | MAPE (%) | R2 |
15min | 0.8887 | 4.1401 | 0.857 |
30min 45min 60min |
0.6529 0.5896 0.9824 |
2.9563 2.3411 4.5986 |
0.899 0.966 0.870 |
Model | RMSE (mmol/L) | MAPE (%) | R2 | Running time (s) |
RNN | 1.2126 | 3.8743 | 0.854 | 45.1 |
LSTM SVR GRU IWT-GRU |
1.1533 1.3754 0.8423 0.5537 |
3.9866 4.2574 3.7549 2.2147 |
0.833 0.872 0.897 0.989 |
48.6 28.9 36.3 37.2 |