Research article Special Issues

Predictive modeling based on small data in clinical medicine: RBF-based additive input-doubling method

  • The paper considers the problem of handling short sets of medical data. Effectively solving this problem will provide the ability to solve numerous classification and regression tasks in case of limited data in health decision support systems. Many similar tasks arise in various fields of medicine. The authors improved the regression method of data analysis based on artificial neural networks by introducing additional elements into the formula for calculating the output signal of the existing RBF-based input-doubling method. This improvement provides averaging of the result, which is typical for ensemble methods, and allows compensating for the errors of different signs of the predicted values. These two advantages make it possible to significantly increase the accuracy of the methods of this class. It should be noted that the duration of the training algorithm of the advanced method remains the same as for existing method. Experimental modeling was performed using a real short medical data. The regression task in rheumatology was solved based on only 77 observations. The optimal parameters of the method, which provide the highest prediction accuracy based on MAE and RMSE, were selected experimentally. A comparison of its efficiency with other methods of this class has been performed. The highest accuracy of the proposed RBF-based additive input-doubling method among the considered ones is established. The method can be modified by using other nonlinear artificial intelligence tools to implement its training and application algorithms and such methods can be applied in various fields of medicine.

    Citation: Ivan Izonin, Roman Tkachenko, Ivanna Dronyuk, Pavlo Tkachenko, Michal Gregus, Mariia Rashkevych. Predictive modeling based on small data in clinical medicine: RBF-based additive input-doubling method[J]. Mathematical Biosciences and Engineering, 2021, 18(3): 2599-2613. doi: 10.3934/mbe.2021132

    Related Papers:

    [1] Yuqing Qian, Tingting Shang, Fei Guo, Chunliang Wang, Zhiming Cui, Yijie Ding, Hongjie Wu . Identification of DNA-binding protein based multiple kernel model. Mathematical Biosciences and Engineering, 2023, 20(7): 13149-13170. doi: 10.3934/mbe.2023586
    [2] Wenjun Xu, Zihao Zhao, Hongwei Zhang, Minglei Hu, Ning Yang, Hui Wang, Chao Wang, Jun Jiao, Lichuan Gu . Deep neural learning based protein function prediction. Mathematical Biosciences and Engineering, 2022, 19(3): 2471-2488. doi: 10.3934/mbe.2022114
    [3] Yuquan Meng, Manjunath Rajagopal, Gowtham Kuntumalla, Ricardo Toro, Hanyang Zhao, Ho Chan Chang, Sreenath Sundar, Srinivasa Salapaka, Nenad Miljkovic, Placid Ferreira, Sanjiv Sinha, Chenhui Shao . Multi-objective optimization of peel and shear strengths in ultrasonic metal welding using machine learning-based response surface methodology. Mathematical Biosciences and Engineering, 2020, 17(6): 7411-7427. doi: 10.3934/mbe.2020379
    [4] Yutian Cai, Ting Wang, Shaohua Wang . A deep neural network-based smart error measurement method for fiscal accounting data. Mathematical Biosciences and Engineering, 2023, 20(6): 10866-10882. doi: 10.3934/mbe.2023482
    [5] Sungwon Kim, Meysam Alizamir, Youngmin Seo, Salim Heddam, Il-Moon Chung, Young-Oh Kim, Ozgur Kisi, Vijay P. Singh . Estimating the incubated river water quality indicator based on machine learning and deep learning paradigms: BOD5 Prediction. Mathematical Biosciences and Engineering, 2022, 19(12): 12744-12773. doi: 10.3934/mbe.2022595
    [6] Yueming Zhou, Junchao Yang, Amr Tolba, Fayez Alqahtani, Xin Qi, Yu Shen . A Data-Driven Intelligent Management Scheme for Digital Industrial Aquaculture based on Multi-object Deep Neural Network. Mathematical Biosciences and Engineering, 2023, 20(6): 10428-10443. doi: 10.3934/mbe.2023458
    [7] Eric Ke Wang, Nie Zhe, Yueping Li, Zuodong Liang, Xun Zhang, Juntao Yu, Yunming Ye . A sparse deep learning model for privacy attack on remote sensing images. Mathematical Biosciences and Engineering, 2019, 16(3): 1300-1312. doi: 10.3934/mbe.2019063
    [8] Ivan Izonin, Nataliya Shakhovska . Special issue: Informatics & data-driven medicine. Mathematical Biosciences and Engineering, 2021, 18(5): 6430-6433. doi: 10.3934/mbe.2021319
    [9] Jing Zhang, Qianqian Wang, Ding Lang, Yuguang Xu, Hong-an Li, Xuewen Li . Research on user recruitment algorithms based on user trajectory prediction with sparse mobile crowd sensing. Mathematical Biosciences and Engineering, 2023, 20(7): 11998-12023. doi: 10.3934/mbe.2023533
    [10] Xianfang Wang, Qimeng Li, Yifeng Liu, Zhiyong Du, Ruixia Jin . Drug repositioning of COVID-19 based on mixed graph network and ion channel. Mathematical Biosciences and Engineering, 2022, 19(4): 3269-3284. doi: 10.3934/mbe.2022151
  • The paper considers the problem of handling short sets of medical data. Effectively solving this problem will provide the ability to solve numerous classification and regression tasks in case of limited data in health decision support systems. Many similar tasks arise in various fields of medicine. The authors improved the regression method of data analysis based on artificial neural networks by introducing additional elements into the formula for calculating the output signal of the existing RBF-based input-doubling method. This improvement provides averaging of the result, which is typical for ensemble methods, and allows compensating for the errors of different signs of the predicted values. These two advantages make it possible to significantly increase the accuracy of the methods of this class. It should be noted that the duration of the training algorithm of the advanced method remains the same as for existing method. Experimental modeling was performed using a real short medical data. The regression task in rheumatology was solved based on only 77 observations. The optimal parameters of the method, which provide the highest prediction accuracy based on MAE and RMSE, were selected experimentally. A comparison of its efficiency with other methods of this class has been performed. The highest accuracy of the proposed RBF-based additive input-doubling method among the considered ones is established. The method can be modified by using other nonlinear artificial intelligence tools to implement its training and application algorithms and such methods can be applied in various fields of medicine.



    Modern development of digital technologies stimulates the development of various spheres of human activity, including medicine. In particular, the advent of many micro devices in phones or other small gadgets, biochips, and other specialized information collection and transmission tools have created potential opportunities for the development of real-time tracking systems designed to monitor a patient's health. If we talk about monitoring one person, then there is a small data approach [1]. Over time, such data are combined with others to form Big data [2], based on which experts conduct intellectual analysis to develop medical decision-making systems for support on diagnosis or treatment.

    There are many other sources of small data [3,4]. In particular, in countries with underdeveloped medical and IT (Information Technologies) areas, various patient tests are recorded in small electronic documents of the doctor or in personal paper cards of patients. Grouping and combining such data into large datasets for further intellectual analysis requires considerable material, time and human resources [5]. Furthermore, there is no absolute confidence in the effectiveness of the analysis of such data. There are a number of problems regarding the accuracy and reliability of such a dataset, the absence of gaps, noise and anomalies in it, and so on. Therefore, there is a very important task to develop a methodology of predictive analytics based on limited data processing in various fields of medicine.

    The task in such formulation is not at all simple. Existing artificial intelligence (AI) tools do not provide sufficient prediction accuracy in processing short data sets [6,7,8]. There are a number of reasons for this. The first is overfitting, which is very typical of methods in this class [9]. Another problem is the very large impact of anomalies and omissions in the already limited amount of available information intended for processing [10]. Furthermore, the use of artificial intelligence methods to process short data sets does not provide a sufficient level of generalization, which affects the prediction accuracy. In turn, this raises concerns about the reliability of the result and the possibility of some generalization of the method, in particular on other data. Despite this, such task does not disappear and its number grows every year, but the available methodology for solving it is quite limited.

    In [11], the authors investigated the development of a small data paradigm and its benefits in medicine. This theoretical work presents a number of very strong arguments in favor of the development of this area. The authors divided them into two main groups: scientific and applied. Each of these reasons is very well argued, and supported by real cases from medicine. This work was the impetus for developing of a new methodology for processing short data sets using artificial intelligence tools.

    If we talk about the practical aspect of the problem, then one of the newest works in the direction of artificial intelligence methods for processing short data sets is a book [9]. Here are a number of machine learning methods for intelligent analysis in the case of limited data sets. Among the shortcomings of the work should be noted the description of only existing methods, as well as their ineffectiveness in the case of processing very short sets of medical data (starting from 10 vectors).

    The review paper [12] is devoted to the problem of applying machine learning methods in order to improve medical services by reducing the necessary costs [13]. The authors studied the existing neural network tools to solve this task. They selected more than 3000 articles from various scientometric databases and conducted a detailed analysis of the application of various artificial neural networks in medical decision-making processes. The analysis showed the widespread use of this artificial intelligence tool in solving various medical problems, particularly in the United States

    The "Biomedical Data Mining Using RBF Neural Networks" chapter of the book [14] describes the use of the Radial Basis Function Artificial Neural Network (RBF ANN) to improve the accuracy of cancer diagnosis. This task has a great practical importance for both patients and physicians. RBF ANN showed high accuracy and, consequently, the possibility of its practical application in various medical decision support systems.

    In [15], the authors investigated the possibility of using RBF ANN to predict the onset of Parkinson's disease tremors in humans. The main practical value of this task is the reduction of battery power used for the Deep Brain Stimulation (DBS). The use of RBF ANN in this case provides accurate recognition of tremors and determines when it is necessary to carry out electrical activity in target brain areas DBS. The forecasting accuracy was more than 80%. In order to reduce the computational costs for the RBF ANN application, the authors proposed an improved forecasting method based on the additional use of particle swarm optimization. According to experimental studies presented in this paper, this approach has paid off.

    Another successful example of the use of RBF ANN to solve medical tasks is a study [16]. The authors developed an adaptive back stepping controller based on this neural network type for medical devices to treat heart rhythm disorders. The main purpose of such development is to increase the productivity of the latter. Since this is a question of human life, such developments must be reliable, stable and effective. The authors provided all these conditions using the Lyapunov function. Experimental results in four patients showed a very high accuracy of the prognosis. Nevertheless, the authors' development requires more extensive experimental research.

    In [17,18], new kernel functions were developed to increase the accuracy of RBF ANN. In [17], the cosine measure was proposed and a significant increase in the accuracy of neural networks of this type based on three different data sets was demonstrated. In [18], the authors used weighted cosine distance and also showed a significant increase in the accuracy of solving a medical problem, namely the prediction of leukemia. These developments can serve as a basis for improving all the methods analyzed above in the case of processing both large and small data sets. However, only a small number of studies have been devoted to the problem of processing short data sets based on RBF ANN.

    The authors of [19,20] presented a new method of processing short data sets. It is based on the use of the classic RBF ANN. As the training algorithm of RBF neural network involves random initial initialization of the weights, and possible random selection of rbf-centers, the method demonstrates different results after each run. The idea of the method from [19,20] is to repeatedly run a neural network of this type with the optimal parameters of its operation to select the most accurate result "method of multiple runs". In the process of obtaining it, it is possible to fix the weights of this artificial intelligence tool, which provides the possibility of reproducing the results of its work.

    In [21], a new RBF-based input-doubling method is introduced. The main idea of the method is to extend the short data set and apply the author's prediction procedures using RBF ANN. This approach demonstrates the improvement of the prediction accuracy when processing short data sets with a satisfactory time of the training procedure (given that we are talking about small data). This work is a continuation of the research begun in [21]. The aim of this paper is to increase the prediction accuracy with satisfactory indicators of the duration of the training procedure of the RBF ANN.

    Our main contribution in this paper can be summarized as follows:

    1) We have improved the existing RBF-based input-doubling method by added additional elements in the formula for calculating the output signal of the method, which improves its prediction accuracy;

    2) We have shown that the adding of additional elements in the formula for calculating the predicted value of the developed method allows to compensate for the errors of various signs of the predicted values, which improves the prediction accuracy of methods of this class;

    3) We have established the highest prediction accuracy of the proposed RBF-based additive input-doubling method with pre-selected values of the optimal parameters of its operation compare to other methods of this class.

    The remainder of this paper is organized as follows: the second section provides basic information about RBF ANN as the basis of the improved method, and describes the general approach, training and application procedures of the improved method. The third section of this paper describes the procedures for selecting the optimal parameters of the proposed RBF-based additive input-doubling method, the results of its work, as well as comparing its effectiveness with other methods in this class. Conclusions and prospects for further research are presented in the conclusion section.

    In general, the classical interpretation of the RBF neural network involves the sequential connection of three layers, the first of which is the input, the weights of which fix the coordinates of the centers of radial functions. Most often, when choosing the coordinates of the centers use the principle of randomization, or they are placed in the centers of clusters [22,23]. In the latter case, Kohonen's self-organizing maps are sometimes used as the input layer [24]. The radial elements of the second layer reproduce Gaussian functions (rarely other variants of functions) from the corresponding scalar products of the weights on the components of the input vectors. Instead of scalar products, Euclidean distances from the coordinates of the input vectors to the coordinates of the RBF centers are often used [25]. The third layer of elements forms a given response surface by a combination (often linear) of the corresponding Gaussian functions (Figure 1). The theoretical substantiation of efficiency of application of neural networks of this type follows from conclusions of the well-known Cover's theorem about the increase of probability of linear separability of images by a transformation of a difficult problem of classification in the space of higher dimension [26].

    Figure 1.  RBF topology.

    Let consider the reflection of the hidden layer. Each hidden item is associated with a function ϕ. Each of these functions has a combined input and generates the value of the activity supplied to the output. The set of activity values of all hidden elements determines the vector on which the input vector is displayed: ϕ(x)=[ϕ(x1),]ϕ(x2),...,ϕ(xM), where М is a the number of hidden elements, and х is input vector.

    The connections of the hidden layer element determine the center of the radial function for the hidden element. The input for each element is equal to the Euclidean norm: netj=XWj=[ni=1(xiwij)2], where n is the number of input elements.

    A variety of activation functions, such as the Gaussian function, is used for hidden elements ϕ(r)exp[r2] or function ϕ(r)c2+r2.

    Here are the main tasks that need to be solved in the process of setting up the RBF ANN [21]:

    –to choose of the method of forming the coordinates of the rbf-centers and the type of basic functions;

    –to set the number of rbf-centers and the parameter of basic functions, which specifies the magnitude of their scope (smooth factor);

    –to train the output ANN layer.

    In practice, two main methods are used to specify the coordinates of the centers: the choice as the centers of coordinates of the inputs of the training sample obtained randomly (from the general training sample), or the location of the centers of basic functions in the centers of clusters [27]. To implement the latter option as a shaper of the input signals of radial functions in this paper, the kNN algorithm is used. Details of the practical implementation of RBF ANN, which is used in this paper, algorithms for its training and application are given in [28].

    Let the sample of independent variables be presented in the form of a matrix with dimension n×m in the form:

    A=[ai,j], (1)

    where i=1,n;j=1,m, n is the number of observations, m the number of independent variables of each observation.

    Given the known output yi of each of n observations, we can form a new matrix ˉA whose dimension is n×(m+1) in the next form:

    ˉA=[ai,j;yi]. (2)

    Algorithmic implementation of the training procedure of the RBF-based additive input-doubling method involves the following steps:

    For each of the i -th observations, where i=1,n, we form a matrix Bi whose dimension is n×2m in the next form:

    Bi=[ai,jak,l], (3)

    where i is a fixed value, j=1,m;k=1,n;l=1,m.

    The output for Eq (3) is formed as the difference of outputs between each new current point yi and all other points of the training data set yk in Eq (3):

    zi=(yiyk). (4)

    Taking into account the values of outputs calculated based on Eq (4), we form a matrix ˉBi whose dimension is n×(2m+1) :

    ˉBi=[ai,jak,l;zi]. (5)

    The final extended matrix of the training data whose dimension is n2×(2m+1) is formed by the following way:

    ˉB=[ˉB1ˉB2...ˉBn]. (6)

    Matrix Eq (6) is the basis of the training procedure of the RBF ANN.

    Application mode of the improved method.

    Let's consider the vector up0 the dimension of which is m, where p0 is fixed, for which (vector) the value of the output signal is unknown. Based on it, in contrast to the existing method [21], we form two matrices C1 і C2 whose dimension is n×2m :

    C1=[up0 ai], (7)
    C2=[ai up0], (8)

    where ai - all vectors from the initial training data set from Eq (1), i=1,n

    Based on matrices Eqs (7) and (8), taking into account the unknown value of the outputs t(1)i,t(2)i, we obtain extended matrices that will contain a set of independent attributes with the corresponding dependent attribute:

    ¯C1=[up0 ai;t(1)i], (9)
    ¯C2=[ai up0;t(2)i], (10)

    Now it is possible to build the response surfaces t(1)i,t(2)i for ¯C1, ¯C2 by applying pre-trained RBF ANN.

    The final required signal y of the improved method, in contrast to the basic method [21], is calculated according to the expression:

    ypredni=1yin+ni=1t(1)i2nni=1t(2)i2n. (11)

    It should be noted that in Eq (11), based on ni=1t(1)i2nni=1t(2)i2n the principle of averaging, which is characteristic of the work of ensemble methods, is implemented. Furthermore, the use of both these terms is in contrast to the existing method, which uses only ni=1t(1)i2n in Eq (11), provides partial mutual compensation of random errors of different signs in equality Eq (11). These two circumnutates significantly increase the accuracy of the output signal based on the improved method.

    The section provides information about the dataset used for modeling the improved method, a description of the procedures for the selection of the optimal parameters of its work, the results and a comparison with other methods of this class.

    A short dataset in one of the branches of medicine was chosen to model the work of the improved method. The authors in [19] used two datasets to predict the compressive strength of trabecular bone (CS in MPa). One of these datasets is real, and the other one is artificial. Each of them contains 77 observations. The actual dataset is compiled based on a review of patients with osteoarthritis. The practical value of this task is the need to predict the strength of the load on the bone when performing exercise in the elderly suffering from osteoarthritis. The purpose of accurate prognosis is to avoid physical injuries from overload, which occur quite often during this disease in the elderly.

    The list of attributes and their main characteristics are given in Table 1.

    Table 1.  Dataset.
    Characteristic Maximum value Average value Minimum value
    Patients' age 87, 00 64, 80 41, 70
    Patients' gender 1, 00 - 0, 00
    Tissue porosity (BV/TV), 43, 50 26, 22 2, 10
    Trabecular thickness factor (tb.th) 419, 00 259, 69 154, 00
    Structure model index (SMI) 2, 10 0, 65 0, 04
    CS (in MPa) 28, 80 16, 40 1, 90

     | Show Table
    DownLoad: CSV

    During modeling, the real dataset was randomly divided into three data sets: training, validation and test. The ratio of this division: 70, 10 and 20%, respectively.

    One of the important steps in applying of RBF ANN is the choice of the number of clusters of the kNN algorithm. Effective selection of this parameter significantly affects the accuracy of the method. We used Mean Absolute Error Eq (12) and Root Mean Square Error Eq (13) as performance indicators.

    MAE=1nni=1(yiypredi)2. (12)
    MAE=ni=1(yiypredi)2n. (13)

    The experiment to select the optimal value of the investigated parameter based on Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) was as follows. Both existing and improved methods were run 100 times when changing the number of clusters of the kNN algorithm from 15 to 100. Other parameters of the method were the same (number of epochs RBF ANN-200 [21], smooth factor equal 0.4., Mean Square Error (MSE) loss function, 100 iterations for k-nn method). It should be noted that the RBF centers at each run were the same for both methods.

    From the obtained results, the maximum value, the minimum value and the average value of MAE and RMSE errors were selected. The results of this experiment are summarized in Figure 2.

    Figure 2.  MAE and RMSE values for the test mode of existing and designed methods when changing the number of clusters [15,100] of k-nn methods for 100 runs of both methods: a) minimal values; b) maximum values; c) average values.

    From all graphs in Figure 2 it is clear that:

    - as the number of clusters of the kNN algorithm increases, the errors of MAE and RMSE increase for both methods;

    - the increasing of MAE and RMSE errors of the improved method with increasing the number of clusters k of the kNN algorithm is significantly higher in comparison with the existing method (except for small values of k);

    - the lowest value of the error of the improved method was obtained for 15 clusters in all three cases (minimum, maximum and average value);

    - the errors of the existing method at this value of the studied parameter (k = 15) are higher in comparison with the improved method both in the case of maximum and minimum values and in the case of the average value at 100 runs of both methods.

    Taking all this into account, the value of the number of clusters of the kNN algorithm equal to 15.

    One of the important parameters of the classical iterative RBF ANN, which is the basis of the existing and improved methods, is the smooth factor. The value of this parameter significantly affects the operation of the neural network of this type and, as a consequence - the operation of existing and improved methods.

    In this work, a study was conducted on the selection of the optimal value of the smooth factor (σ). The experiment was as follows. Both existing and improved methods were run 100 times (one experiment) and for each experiment, we have changed the smooth factor values from 0.1 to 0.9 with a step of 0.1. Other parameters of the method were the same (number of epochs RBF ANN-200 [21], MSE loss function, 15 clusters, 100 iterations for k-nn method).

    From the obtained results, the maximum value, the minimum value and the average value of MAE and RMSE errors for 100 runs of the method were selected. The results of this experiment are summarized in Figure 3.

    Figure 3.  MAE and RMSE values for the test mode of existing and designed methods when changing smooth factor σ for 10 runs of both methods: a) minimal values; b) maximum values; c) average values.

    All graphs from Figure 2 clearly show the following:

    - the local maximum at small values of σ, (σ < 0.2) for both studied methods indicates the need to choose much larger values of this parameter for the practical implementation of both methods;

    - as the value of the smooth factor increased, the errors of MAE and RMSE decreased for both methods;

    - at a value of σ = 0.5, both methods show very similar results, although the improved method shows a smaller value of both errors;

    - in the interval 0.7 < σ < 0.9, both errors for both studied methods acquire saturation stages; this statement is true for the minimum, maximum and average values among 100 runs of both methods;

    - in the interval 0.7 < σ < 0.9, the improved method demonstrates a significant increase in the prediction accuracy compared to the existing one;

    - the smallest value of errors for the improved method is received at the value of smooth factor σ = 0.8 both in case of the maximum and minimum values and in case of average value at its 100 runs in comparison with the existing method.

    Therefore, the value of the smooth factor σ = 0.8 was used for further research.

    Based on the selected optimal parameters of the improved method the following results which are summarized in Table 2 are received.

    Table 2.  Results of the RBF-based additive input-doubling method.
    MAE RMSE Training time (seconds)
    Train mode
    6.150 6.237 13.906
    Test mode
    5.089 6.057 -

     | Show Table
    DownLoad: CSV

    It should be noted that the minimum values obtained at 100 runs of the method are selected as the results of its work. This is justified by the possibility of fixing the synaptic weights, which provides all the necessary conditions for the full reproduction of the results of the method.

    The comparison of the efficiency of the improved method was performed with the existing method, as well as the classical iterative RBF ANN, which is the basis of the work of both studied methods. The comparison was based on both errors as well as the duration of the training procedure. The results of this study are summarized in Figure 4.

    Figure 4.  Performance indicators for train and test modes of all investigated methods: a) MAE values; b) RMSE values; c) training time (in seconds).

    As can be seen from Figure 4, the highest operating errors in both training and application modes were obtained using the classical iterative RBF ANN. More accurate results were obtained using the existing RBF-based input-doubling method. However, the duration of the training procedure is increased by 5 seconds, which is about 50% of the duration of the classical RBF ANN, which is the basis of its work. This is due to two reasons: 1) the quadratic increase of the training sample according to (3)-(5)-(6); 2) doubling the number of input features according to (3) and as a consequence of the expansion of the RBF ANN's input layer, which is the basis of this method.

    If we analyze the effectiveness of the proposed method, we can say the following:

    - the improved method provides a significant increase of the prediction accuracy based on both errors compared to the existing method;

    - the training errors for both existing and improved methods will be the same (or very close) due to the same procedure for the formation of the training sample according to (3)-(5)-(6);

    - the duration of the training procedure will also be the same (or very close to the duration of the training procedure of the existing method), as explained in the previous case;

    - the application time of the improved method will be longer than the application time of the existing method, due to the additional procedures and as a consequence of changing the search formula for the output signal of the method. However, since these are small and very small data sets, this shortcoming is not significant.

    The development of modern medicine is accompanied by a large amount of diverse information about the object of observation. There are many applications in the field of medicine that require the use of artificial intelligence tools to obtain a faster or more accurate solution for quality diagnosis or treatment. In some cases, limited data sets significantly reduce the efficiency or simply make it impossible to apply artificial intelligence tools for the intellectual analysis of such information. This paper considers the current problem of regression analysis via artificial neural networks based on short sets of medical data. The authors proposed an improved RBF-based input-doubling method by introducing additional elements into the formula for calculating the output signal of the method. This improves prediction accuracy while the duration of the training procedure remains unchanged. The effectiveness of the advanced method was tested using a real short set of medical data, which contained only 77 observations. It is experimentally established that the additional elements in the formula for calculating the predicted value of the developed method allows to compensate for the errors of different signs of the temporary predicted values, which improves the accuracy of methods of this class as a whole. The selection of the optimal parameters of the developed method is carried out. The highest prediction accuracy of the proposed RBF-based additive input-doubling method is established by comparing the efficiency of its work with other methods of this class. The accuracy of the improved method is increased by 1% based on both MAE and RMSE compared to the basic method. The training time (13.9 seconds) of both methods is the same. Among the disadvantages of the studied methods should be noted significant time delays of their work. That is why future studies will consider non-iterative variants of RBF ANN [29] and will model the operation of the method based on them. This should provide a significant reduction of the training time of the method while maintaining, and even increasing the accuracy of its work.

    The National Research Foundation of Ukraine funds this study from the state budget of Ukraine within the project "Decision support system for modelling the spread of viral infections" (№ 2020.01 / 0025).

    The authors declare that they have no conflict of interests.



    [1] D. I. Christine, M. Thinyane, Small Data approaches provide nuance and context to health datasets, Available from: http://theconversation.com/small-data-approaches-provide-nuance-and-context-to-health-datasets-121972.
    [2] N. Melnykova, N. Shakhovska, M. G. ml, V. Melnykov, Using big data for formalization the patient's personalized data, Proc. Comp. Scien., 155 (2019), 624–629. doi: 10.1016/j.procs.2019.08.088
    [3] E. K. Wang, F. Wang, R. P. Sun, X. Liu, A new privacy attack network for remote sensing images classification with small training samples, Math. Biosci. Eng. 16 (2019), 4456–4476. doi: 10.3934/mbe.2019222
    [4] T. Mao, L. Yu, Y. Zhang, L. Zhou, Modified mahalanobis-taguchi system based on proper orthogonal decomposition for high-dimensional-small-sample-size data classification, Math. Biosci. Eng. 18 (2021), 426–444. doi: 10.3934/mbe.2021023
    [5] L. Mochurad, M. Yatskiv, Simulation of a human operator's response to stressors under production conditions, CEUR-WS, 2753 (2020), 156–169.
    [6] V. Kotsovsky, F. Geche, A. Batyuk, On the computational complexity of learning bithreshold neural units and networks, Adv. Intel. Syst. Comp., 1020 (2019), 189–202.
    [7] V. Kotsovsky, F. Geche, A. Batyuk, Finite generalization of the offline spectral learning, in 2018 IEEE 2nd Intern. Conf. on Data Stream Mining Processing (DSMP), 2018,356–360.
    [8] S. Fedushko, M. G. ml, T. Ustyianovych, Medical card data imputation and patient psychological and behavioral profile construction, Proc. Comp. Scien., 160 (2019), 354–361. doi: 10.1016/j.procs.2019.11.080
    [9] S. Huang, H. Deng, Data Analytics A Small Data Approach, 1st edition, Routledge & CRC Press, 2021.
    [10] T. Hovorushchenko, A Herts, Y. Hnatchuk, Concept of intelligent decision support system in the legal regulation of the surrogate motherhood, CEUR-WS, 2488 (2019), 57–68.
    [11] E.B. Hekler, P. Klasnja, G. Chevance G, N. M. Golaszewski, D. Lewis, I. Sim, Why we need a small data paradigm, BMC Med., 17 (2019), 133. doi: 10.1186/s12916-019-1366-x
    [12] N. Shahid, T. Rappon, W. Berta, Applications of artificial neural networks in health care organizational decision-making: a scoping review, PLOS ONE, 14 (2019), e0212356. doi: 10.1371/journal.pone.0212356
    [13] S. Kaczor, N. Kryvinska, It is all about services-fundamentals, drivers, and business models, J. Serv. Sci. Res., 5 (2013), 125–154.
    [14] J. Wang, Data Warehousing and Mining: Concepts, Methodologies, Tools, and Applications, IGI Global, 2008.
    [15] D. Wu, K. Warwick, Z. Ma, M. N. Gasson, J. G. Burgess, S. Pan, et. al., Prediction of parkinson's disease tremor onset using a radial basis function neural network based on particle swarm optimization, Int. J. Neural. Syst., 20 (2010), 109–116. doi: 10.1142/S0129065710002292
    [16] M. E. Karar, Robust RBF neural network‑based backstepping controller for implantable cardiac pacemakers, I. J. Adap. Cont. Sign. Proc., 32 (2018), 1040–1051. doi: 10.1002/acs.2884
    [17] W. Aftab, M. Moinuddin, M. S. Shaikh, A novel kernel for RBF based neural networks, Abs. Appl. Anal., 2 (2014), 176253
    [18] M. Moinuddin, I. Naseem, W. Aftab, S. A. Bencherif, A. Memic, A weighted cosine RBF neural networks, J. Mol. Biol. Biotech., 2 (2017), 2.
    [19] T. Shaikhina, N. A. Khovanova, Handling limited datasets with neural networks in medical applications: a small-data approach, Artifl. Intel. Med., 75 (2017), 51–63. doi: 10.1016/j.artmed.2016.12.003
    [20] T. Shaikhina, D. Lowe, S. Daga, D. Briggs, R. Higgins, N. Khovanova, Machine learning for predictive modelling based on small data in biomedical engineering, IFAC-PapersOnLine, 48 (2015), 469–474.
    [21] I. Izonin, R. Tkachenko, S. Fedushko, D. Koziy, K. Zub, O, Vovk, RBF-based input doubling method for small medical data processing, Adv. Intell. Syst. Comput., 2021.
    [22] Y. Bodyanskiy, I. Pliss, A. Deineko, Multilayer radial-basis function network and its learning, IEEE Int. Conf. Comp. Sci. Inf. Tech., (2020), 92–95.
    [23] S. Babichev, J. Škvor, Technique of gene expression profiles extraction based on the complex use of clustering and classification methods, Diagnostics, 10 (2020), 584.
    [24] T. Kohonen, Essentials of the self-organizing map, Neur. Netw., 37 (2013), 52–65. doi: 10.1016/j.neunet.2012.09.018
    [25] S. Subbotin, Radial-basis function neural network synthesis on the basis of decision tree, Opt. Mem. Neur. Networks, 29 (2020), 7–18. doi: 10.3103/S1060992X20010051
    [26] F. Samuelson, D. G. Brown, Application of cover's theorem to the evaluation of the performance of CI observers, Int. Joint Conf. Neur. Netw., (2011), 1020–1026.
    [27] Ye. V. Bodyanskiy, A. O. Deineko, Ya. V. Kutsenko, On-line kernel clustering based on the general regression neural network and kohonen's self-organizing map, Aut. Control Comp. Sci., 51 (2017), 55–62. doi: 10.3103/S0146411617010023
    [28] M. Deshp, Using neural networks for regression: radial basis function networks, Available from: https://pythonmachinelearning.pro/using-neural-networks-for-regression-radial-basis-function-networks/.
    [29] R. Tkachenko, H. Kutucu, I. Izonin, A. Doroshenko, Y. Tsymbal, Non-iterative neural-like predictor for solar energy in libya, CEUR-WS, 2105 (2018), 35–45.
  • This article has been cited by:

    1. Ivan Izonin, Roman Tkachenko, Nataliya Shakhovska, Nataliia Lotoshynska, The Additive Input-Doubling Method Based on the SVR with Nonlinear Kernels: Small Data Approach, 2021, 13, 2073-8994, 612, 10.3390/sym13040612
    2. Lesia Mochurad, Natalia Kryvinska, Parallelization of Finding the Current Coordinates of the Lidar Based on the Genetic Algorithm and OpenMP Technology, 2021, 13, 2073-8994, 666, 10.3390/sym13040666
    3. Alireza Mohammadi, Dmytro Chumachenko, Tetyana Chumachenko, 2021, Machine Learning Model of COVID-19 Forecasting in Ukraine Based on the Linear Regression, 978-1-6654-4296-1, 149, 10.1109/ELIT53502.2021.9501122
    4. Natalya Shakhovska, Vitaliy Yakovyna, Valentyna Chopyak, A new hybrid ensemble machine-learning model for severity risk assessment and post-COVID prediction system, 2022, 19, 1551-0018, 6102, 10.3934/mbe.2022285
    5. Vladyslav Kotsovsky, Anatoliy Batyuk, Volodymyr Voityshyn, 2021, On the Size of Weights for Bithreshold Neurons and Networks, 978-1-6654-4257-2, 13, 10.1109/CSIT52700.2021.9648833
    6. Nataliia Dotsenko, Dmytro Chumachenko, Igor Chumachenko, 2021, Formation of a project team using an intelligent logic-combinatorial approach, 978-1-6654-4257-2, 206, 10.1109/CSIT52700.2021.9648786
    7. Larysa Zomchak, Maryna Nehrey, 2022, Chapter 59, 978-3-031-04808-1, 645, 10.1007/978-3-031-04809-8_59
    8. Jarosław Bilski, Bartosz Kowalczyk, 2023, Chapter 1, 978-3-031-23491-0, 3, 10.1007/978-3-031-23492-7_1
    9. Lindsay C. Todman, Alex Bush, Amelia S.C. Hood, ‘Small Data’ for big insights in ecology, 2023, 01695347, 10.1016/j.tree.2023.01.015
    10. Dmytro Chumachenko, Ievgen Meniailov, Kseniia Bazilevych, Tetyana Chumachenko, Sergiy Yakovlev, On intelligent agent-based simulation of COVID-19 epidemic process in Ukraine, 2022, 198, 18770509, 706, 10.1016/j.procs.2021.12.310
    11. Hossein Saberi, Ehsan Esmaeilnezhad, Hyoung Jin Choi, Artificial Neural Network to Forecast Enhanced Oil Recovery Using Hydrolyzed Polyacrylamide in Sandstone and Carbonate Reservoirs, 2021, 13, 2073-4360, 2606, 10.3390/polym13162606
    12. Kaixin Liu, Fumin Wang, Yuxiang He, Yi Liu, Jianguo Yang, Yuan Yao, Data-Augmented Manifold Learning Thermography for Defect Detection and Evaluation of Polymer Composites, 2022, 15, 2073-4360, 173, 10.3390/polym15010173
    13. Vladyslav Kotsovsky, Anatoliy Batyuk, 2021, Decision List-Based Representation of Bithreshold Functions, 978-1-6654-4257-2, 21, 10.1109/CSIT52700.2021.9648689
    14. Zongliang Guo, Sikai Lin, Runze Suo, Xinming Zhang, An Offline Weighted-Bagging Data-Driven Evolutionary Algorithm with Data Generation Based on Clustering, 2023, 11, 2227-7390, 431, 10.3390/math11020431
    15. Ivan Izonin, Nataliya Shakhovska, Special issue: Informatics & data-driven medicine, 2021, 18, 1551-0018, 6430, 10.3934/mbe.2021319
    16. Vladyslav Kotsovsky, Anatoliy Batyuk, 2022, Feed-forward Neural Network Classifiers with Bithreshold-like Activations, 979-8-3503-3431-9, 9, 10.1109/CSIT56902.2022.10000739
    17. Dmytro Chumachenko, Kseniia Bazilevych, Ievgen Meniailov, Sergiy Yakovlev, Tetyana Chumachenko, 2021, Simulation of COVID-19 Dynamics using Ridge Regression, 978-1-6654-0618-5, 163, 10.1109/AICT52120.2021.9628991
    18. Dmytro Chumachenko, Tetyana Chumachenko, Ievgen Meniailov, Olena Muradyan, Grigoriy Zholtkevych, 2021, Forecasting of COVID-19 Epidemic Process by Lasso Regression, 978-1-6654-2652-7, 80, 10.1109/UkrMiCo52950.2021.9716621
    19. Vitaliy Yakovyna, Natalya Shakhovska, Software failure time series prediction with RBF, GRNN, and LSTM neural networks, 2022, 207, 18770509, 837, 10.1016/j.procs.2022.09.139
    20. Liang-Sian Lin, Susan C Hu, Yao-San Lin, Der-Chiang Li, Liang-Ren Siao, A new approach to generating virtual samples to enhance classification accuracy with small data—a case of bladder cancer, 2022, 19, 1551-0018, 6204, 10.3934/mbe.2022290
    21. Junbo Qiu, Xin Yin, Yucong Pan, Xinyu Wang, Min Zhang, Prediction of Uniaxial Compressive Strength in Rocks Based on Extreme Learning Machine Improved with Metaheuristic Algorithm, 2022, 10, 2227-7390, 3490, 10.3390/math10193490
    22. Nataliia Melnykova, Nataliya Shakhovska, Volodymyr Melnykov, Kateryna Melnykova, Khrystyna Lishchuk-Yakymovych, Personalized Data Analysis Approach for Assessing Necessary Hospital Bed-Days Built on Condition Space and Hierarchical Predictor, 2021, 5, 2504-2289, 37, 10.3390/bdcc5030037
    23. Hongliang Gao, Lang Xiong, Research on a hybrid controller combining RBF neural network supervisory control and expert PID in motor load system control, 2022, 14, 1687-8132, 168781322211099, 10.1177/16878132221109994
    24. Elias Dritsas, Maria Trigka, Machine Learning Techniques for Chronic Kidney Disease Risk Prediction, 2022, 6, 2504-2289, 98, 10.3390/bdcc6030098
    25. Yaroslav Tolstyak, Valentyna Chopyak, Myroslav Havryliuk, An investigation of the primary immunosuppressive therapy's influence on kidney transplant survival at one month after transplantation, 2023, 09663274, 101832, 10.1016/j.trim.2023.101832
    26. Lesia Mochurad, Yaroslav Hladun, Yevgen Zasoba, Michal Gregus, An Approach for Opening Doors with a Mobile Robot Using Machine Learning Methods, 2023, 7, 2504-2289, 69, 10.3390/bdcc7020069
    27. Laith Alzubaidi, Jinshuai Bai, Aiman Al-Sabaawi, Jose Santamaría, A. S. Albahri, Bashar Sami Nayyef Al-dabbagh, Mohammed A. Fadhel, Mohamed Manoufali, Jinglan Zhang, Ali H. Al-Timemy, Ye Duan, Amjed Abdullah, Laith Farhan, Yi Lu, Ashish Gupta, Felix Albu, Amin Abbosh, Yuantong Gu, A survey on deep learning tools dealing with data scarcity: definitions, challenges, solutions, tips, and applications, 2023, 10, 2196-1115, 10.1186/s40537-023-00727-2
    28. Anusuya KRİSHNAN, Kennedyraj MARİAFRANCİS, Analyzing the Impact of Augmentation Techniques on Deep Learning Models for Deceptive Review Detection: A Comparative Study, 2023, 3, 2757-7422, 96, 10.54569/aair.1329048
    29. Dmytro Chumachenko, Sergiy Yakovlev, Artificial Intelligence Algorithms for Healthcare, 2024, 17, 1999-4893, 105, 10.3390/a17030105
    30. Dmytro Chumachenko, Tetyana Chumachenko, Ievgen Meniailov, Olena Muradyan, Grigoriy Zholtkevych, 2023, Chapter 30, 978-3-031-35466-3, 503, 10.1007/978-3-031-35467-0_30
    31. Dmytro Chumachenko, Pavlo Pyrohov, 2022, Estimation of the Migration Impact on COVID-19 Dynamics in Slovakia by Machine Learning: Simulation Study during Russian War in Ukraine, 979-8-3503-9891-5, 383, 10.1109/PICST57299.2022.10238479
    32. Viacheslav Kovtun, Krzysztof Grochla, Vyacheslav Kharchenko, Mohd Anul Haq, Andriy Semenov, Stochastic forecasting of variable small data as a basis for analyzing an early stage of a cyber epidemic, 2023, 13, 2045-2322, 10.1038/s41598-023-49007-2
    33. Dmytro Chumachenko, Adam Wojciechowski, Sergiy Yakovlev, 2023, Chapter 19, 978-3-031-36200-2, 227, 10.1007/978-3-031-36201-9_19
    34. Viacheslav Kovtun, Elena Zaitseva, Vitaly Levashenko, Krzysztof Grochla, Oksana Kovtun, Small Stochastic Data Compactification Concept Justified in the Entropy Basis, 2023, 25, 1099-4300, 1567, 10.3390/e25121567
    35. Yue Ma, Mingming Guo, Ye Tian, Jialing Le, Recent advances and prospects in hypersonic inlet design and intelligent optimization, 2024, 146, 12709638, 108953, 10.1016/j.ast.2024.108953
    36. Xingjiang Xu, Predicting the Risk of Chronic Kidney Disease Using Machine Learning, 2023, 1556-5068, 10.2139/ssrn.4636627
    37. Inyong Jeong, Yeongmin Kim, Nam-Jun Cho, Hyo-Wook Gil, Hwamin Lee, A Novel Method for Medical Predictive Models in Small Data Using Out-of-Distribution Data and Transfer Learning, 2024, 12, 2227-7390, 237, 10.3390/math12020237
    38. Tiko Iyamu, Wandisa Nyikana, Activity Theory View of Big Data Architectural Design for Enterprises, 2024, 9, 2468-4376, 29581, 10.55267/iadt.07.15494
    39. Dmytro Chumachenko, Tetiana Dudkina, Sergiy Yakovlev, Tetyana Chumachenko, Fei Hu, Effective Utilization of Data for Predicting COVID-19 Dynamics: An Exploration through Machine Learning Models, 2023, 2023, 1687-6423, 1, 10.1155/2023/9962100
    40. Dmytro Chumachenko, 2024, Chapter 2, 978-3-031-59130-3, 27, 10.1007/978-3-031-59131-0_2
    41. Daobao Luo, Xin Hu, Wujun Ji, Construction of battery charge state prediction model for new energy electric vehicles, 2024, 119, 00457906, 109561, 10.1016/j.compeleceng.2024.109561
    42. Mykola Butkevych, Olha Manakova, Dmytro Chumachenko, 2024, Forecasting of Salmonellosis Dynamics with LSTM Deep Learning Model, 979-8-3315-2056-4, 213, 10.1109/TCSET64720.2024.10755553
    43. Mykola Stakhiv, Predicting the Duration of Treatment Using Personalized Medical Data, 2024, 9, 25240382, 146, 10.23939/acps2024.02.146
    44. Pallavi V. Baviskar, Vidya A. Nemade, Vishal V. Mahale, 2025, Chapter 19, 978-981-97-8668-8, 245, 10.1007/978-981-97-8669-5_19
    45. Dmytro Chumachenko, 2025, Chapter 26, 978-3-031-80934-7, 555, 10.1007/978-3-031-80935-4_26
  • Reader Comments
  • © 2021 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(13341) PDF downloads(162) Cited by(45)

Figures and Tables

Figures(4)  /  Tables(2)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog