
In the medical field, researchers are often unable to obtain the sufficient samples in a short period of time necessary to build a stable data-driven forecasting model used to classify a new disease. To address the problem of small data learning, many studies have demonstrated that generating virtual samples intended to augment the amount of training data is an effective approach, as it helps to improve forecasting models with small datasets. One of the most popular methods used in these studies is the mega-trend-diffusion (MTD) technique, which is widely used in various fields. The effectiveness of the MTD technique depends on the degree of data diffusion. However, data diffusion is seriously affected by extreme values. In addition, the MTD method only considers data fitted using a unimodal triangular membership function. However, in fact, data may come from multiple distributions in the real world. Therefore, considering the fact that data comes from multi-distributions, in this paper, a distance-based mega-trend-diffusion (DB-MTD) technique is proposed to appropriately estimate the degree of data diffusion with less impacts from extreme values. In the proposed method, it is assumed that the data is fitted by the triangular and trapezoidal membership functions to generate virtual samples. In addition, a possibility evaluation mechanism is proposed to measure the applicability of the virtual samples. In our experiment, two bladder cancer datasets are used to verify the effectiveness of the proposed DB-MTD method. The experimental results demonstrated that the proposed method outperforms other VSG techniques in classification and regression items for small bladder cancer datasets.
Citation: Liang-Sian Lin, Susan C Hu, Yao-San Lin, Der-Chiang Li, Liang-Ren Siao. A new approach to generating virtual samples to enhance classification accuracy with small data—a case of bladder cancer[J]. Mathematical Biosciences and Engineering, 2022, 19(6): 6204-6233. doi: 10.3934/mbe.2022290
[1] | Eric Ke Wang, Fan Wang, Ruipei Sun, Xi Liu . A new privacy attack network for remote sensing images classification with small training samples. Mathematical Biosciences and Engineering, 2019, 16(5): 4456-4476. doi: 10.3934/mbe.2019222 |
[2] | Eric Ke Wang, Nie Zhe, Yueping Li, Zuodong Liang, Xun Zhang, Juntao Yu, Yunming Ye . A sparse deep learning model for privacy attack on remote sensing images. Mathematical Biosciences and Engineering, 2019, 16(3): 1300-1312. doi: 10.3934/mbe.2019063 |
[3] | Ji-Ming Wu, Wang-Ren Qiu, Zi Liu, Zhao-Chun Xu, Shou-Hua Zhang . Integrative approach for classifying male tumors based on DNA methylation 450K data. Mathematical Biosciences and Engineering, 2023, 20(11): 19133-19151. doi: 10.3934/mbe.2023845 |
[4] | Liang-Sian Lin, Chen-Huan Kao, Yi-Jie Li, Hao-Hsuan Chen, Hung-Yu Chen . Improved support vector machine classification for imbalanced medical datasets by novel hybrid sampling combining modified mega-trend-diffusion and bagging extreme learning machine model. Mathematical Biosciences and Engineering, 2023, 20(10): 17672-17701. doi: 10.3934/mbe.2023786 |
[5] | Jie Bai, Heru Xue, Xinhua Jiang, Yanqing Zhou . Classification and recognition of milk somatic cell images based on PolyLoss and PCAM-Reset50. Mathematical Biosciences and Engineering, 2023, 20(5): 9423-9442. doi: 10.3934/mbe.2023414 |
[6] | Gang Chen, Binjie Hou, Tiangang Lei . A new Monte Carlo sampling method based on Gaussian Mixture Model for imbalanced data classification. Mathematical Biosciences and Engineering, 2023, 20(10): 17866-17885. doi: 10.3934/mbe.2023794 |
[7] | Mahmoud El-Morshedy, Zubair Ahmad, Elsayed tag-Eldin, Zahra Almaspoor, Mohamed S. Eliwa, Zahoor Iqbal . A new statistical approach for modeling the bladder cancer and leukemia patients data sets: Case studies in the medical sector. Mathematical Biosciences and Engineering, 2022, 19(10): 10474-10492. doi: 10.3934/mbe.2022490 |
[8] | Ting Mao, Lanting Yu, Yueyi Zhang, Li Zhou . Modified Mahalanobis-Taguchi System based on proper orthogonal decomposition for high-dimensional-small-sample-size data classification. Mathematical Biosciences and Engineering, 2021, 18(1): 426-444. doi: 10.3934/mbe.2021023 |
[9] | Xinyan Chen, Zhaohui Jiang, Qile Tai, Chunshan Shen, Yuan Rao, Wu Zhang . Construction of a photosynthetic rate prediction model for greenhouse strawberries with distributed regulation of light environment. Mathematical Biosciences and Engineering, 2022, 19(12): 12774-12791. doi: 10.3934/mbe.2022596 |
[10] | Luca Bertelli, Frédéric Gibou . Fast two dimensional to three dimensional registration of fluoroscopy and CT-scans using Octrees on segmentation maps. Mathematical Biosciences and Engineering, 2012, 9(3): 527-537. doi: 10.3934/mbe.2012.9.527 |
In the medical field, researchers are often unable to obtain the sufficient samples in a short period of time necessary to build a stable data-driven forecasting model used to classify a new disease. To address the problem of small data learning, many studies have demonstrated that generating virtual samples intended to augment the amount of training data is an effective approach, as it helps to improve forecasting models with small datasets. One of the most popular methods used in these studies is the mega-trend-diffusion (MTD) technique, which is widely used in various fields. The effectiveness of the MTD technique depends on the degree of data diffusion. However, data diffusion is seriously affected by extreme values. In addition, the MTD method only considers data fitted using a unimodal triangular membership function. However, in fact, data may come from multiple distributions in the real world. Therefore, considering the fact that data comes from multi-distributions, in this paper, a distance-based mega-trend-diffusion (DB-MTD) technique is proposed to appropriately estimate the degree of data diffusion with less impacts from extreme values. In the proposed method, it is assumed that the data is fitted by the triangular and trapezoidal membership functions to generate virtual samples. In addition, a possibility evaluation mechanism is proposed to measure the applicability of the virtual samples. In our experiment, two bladder cancer datasets are used to verify the effectiveness of the proposed DB-MTD method. The experimental results demonstrated that the proposed method outperforms other VSG techniques in classification and regression items for small bladder cancer datasets.
Bladder cancer is a common cancer in the human urinary system. Some bladder cancer studies have reported that this cancer is related to tumor suppressor genes and oncogene genes, such as multidrug resistance (MDR), topoisomerase II (Topo II), Rb, epidermal growth factor receptor (EGFR), HER-2 (Neu), c-ErbB-3, c-ErbB-4, Cyclin A, Cyclin D1, P16, Cdc 2, Bcl-2, Bax, etc. [1,2,3,4,5,6,7]. These genes can be used to diagnose bladder cancer [1,4,6] and analyze the effects of X-ray treatment among bladder cancer patients [5,7]. For example, higher Cyclin D1 values indicate that patients are in the early stage of bladder cancer [2]. The over-expression values of EGFR, Neu, c-ErbB-3, c-ErbB-4 may reflect the treatment outcomes of Cobalt-60 (Co-60) radiation therapy on bladder cancer [3]. These studies often suffer from a small sample problem because it is difficult to obtain a sufficient number of gene expression profiles from patients due to the need for a costly genome sequencing procedure [8,9]. However, for researchers who adopt data-driven models as forecasting tools to predict cancer, it is necessary to collect sufficient data and find representative samples from which to build data-driven models with good learning performance [10,11,12]. With a limited number of gene profiles, traditional machine learning methods cannot construct data-driven models that will provide accurate predictions of bladder cancer. For this reason, small data learning has become an important challenge to bladder cancer prediction with small datasets.
Previous research approaches to the problem of small data learning can be divided into two types: The first proposes a better learning model for different data sets. For example, Mao et al. [13] proposed a modified Mahalanobis-Taguchi System to extract important information for the purpose of improving classification accuracy for a high-dimensional small dataset. Izonin et al. [14] proposed the input-doubling method to improve prediction performance of the compressive strength of trabecular bone with only 77 observations. Due to data diversity, it is difficult to apply the same model to different datasets. Researchers must fine-tune the parameters of data-driven models to improve the learning of models using small data. The second approach is a virtual sample generation (VSG) method used to expand the quantity of data. VSG methods have been developed in various fields, such as diagnostic classification of tumors [1,4,6,15], prediction of the treatment effects of radiotherapy on tumors [5,7], and engineering applications [16,17,18]. These methods based on the concept of VSG are typically classified into three types: resampling-based VSG, model-based VSG, and diffusion-based VSG. The bootstrap method and the Synthetic Minority Over-sampling Technique (SMOTE) [19] typify resampling-based VSG methods. The bootstrap method replaces the original sample with a synthetic sample by re-sampling the original samples [20]. The SMOTE method draws out a neighboring example of the original data as a new example to fill gaps between data. The neighboring example is calculated as shown in the following equation: xsynthetic= xi+random(0,1)⋅(xj−xi),∀i≠j, where xsynthetic is a synthetic virtual sample different from xi, and xi is resampled data and xj is the neighboring data of xi, e.g., Figure 1.
Many studies have combined bootstrap methods with data-driven models to achieve good learning results [21,22,23]. For example, Lee et al. [21] created samples using the combined bootstrap method to construct a support vector regression (SVR), when only a small number of samples from the health care system were used to predict oscillometric measurement of blood pressure (BP). La Rocca and Perna [23] proposed a combination of extreme learning machine (ELM) models with bootstrap samples to improve the learning accuracy of small data sets. These experimental results showed that the proposed method led to a significant improvement in the performance of small data learning. The problem with bootstrapping, however, is that it can repeatedly overlearn the same sample, thus creating an overfitting problem. In this regard, Lee et al. [21] applied a bootstrap to a SVR to reduce the probability of overfitting. Unfortunately, most machine learning models are not immune to this problem. From the perspective of augmenting data diversity, some scholars suggest that by treating data that approximates the original sample as a virtual sample, the model can learn diverse data and reduce the incidence of overfitting.
Model-based VSG methods depend on machine learning models or artificial neural networks to find the relevance between input attributes and the output targets of small data in high dimensional space. The characteristics of original data can be learned by generating virtual samples in high dimensional space. Cho et al. [24] randomly generated virtual samples and selected virtual samples that were similar to the original examples using a multilayer perceptron neural network. Huang and Moraga [25] suggested the use of the diffusion-neural-network (DNN), which utilizes information diffusion and fuzzy theory to derive artificial samples in DNNs. In the field of image recognition, the generative adversarial net (GAN) model [26], which is specifically designed to generate synthetic images. The concept of the GAN is to generate a fake image that approximates a true image to increase the amount of training data. However, the GAN model is computationally intensive and may suffer from mode collapses due to low image variability when the true image data has an extremely skewed distribution [27].
Diffusion-based VSG methods are represented based on information derived from original data. The method usually assumes that data comes from a mathematical function or a specific probability distribution to generate virtual samples. The most representative method is mega-trend-diffusion (MTD) method [28]. Based on a triangular membership function (MF), the MTD method is to estimate domain of small data and generate virtual samples within the domain. Methods based on MTD method have been developed to improve the prediction performance of data-driven models with small datasets in practical applications [29,30,31,32]. For example, Majid et al. [30] applied the MTD technique to perform estimation of the data range in minority class data. Their experimental results showed that the generated virtual samples improved the classification accuracy of four data-driven models (k-nearest neighbor, support vector machines, Naïve Bayes, and random forest) for the prediction of human breast and colon cancers.
As an alternative to the MTD method, Yang et al. [33] proposed the VSG method based on a Gaussian distribution (GD), and virtual samples were generated within the estimated data domain. However, in the real world, data may be fitted using multiple distributions. The MTD and GD methods did not consider generating virtual samples based on multiple distributions. For this reason, based on multiple triangular MFs, Zhu et al. [31] proposed the multi-distribution MTD (MD-MTD) to generate virtual samples intended to improve the learning accuracies of models of purified terephthalic acid production. In addition, the MD-MTD method has been applied to enhance the prediction of an applicant risk assessment of an early online lending platform [32]. In addition to the assumption of multiple triangular MFs, Wang et al. [34] assumed that the data came from a mixture of normal distributions to create a large amount of data from a similar process and fuse data using a transfer learning method. The distribution of the fused data can be inferred using a nonparametric method. The estimated distribution is used to generate virtual samples to improve prediction in other processes. Although their method achieved good prediction results, it is more complex than the MD-MTD method. In addition, virtual samples generated using their method follow symmetric normal distributions. However, in fact, it is more appropriate to consider using asymmetric distributions to generate virtual samples for small datasets [31,32]. Therefore, based on the assumption of multi-distributions, in this paper, a distance-based mega-trend-diffusion (DB-MTD) technique is proposed to avoid over-diffusion of data due to extreme values and estimate the central locations (CLs) of multi-distributions. The detailed motivation for this work is explained in the following section.
The traditional MTD method uses the midpoint of the data as an estimation of the CL of the data and estimates the skewness of the data by the amount of data greater or less than CL. The concept of data diffusion in MTD is that when the data is more skewed, the data domain is more diffused. However, when estimating the skewness of data in terms of data quantity, there is a problem: a small data set with about the same amount of data on both sides of the CL is estimated to be an un-skewed distribution for the reasons illustrated using Scenario A and Scenario B in Figure 2(a). Scenario A considers the amount of data used to estimate the degree of skewness of the data. As a result, the degree of skewness on the left side of the data is not significant enough to increase the spread of the data range. Scenario B considers the distance between the data and the CL to estimate the degree of skewness, and the data on the left is farther away from the CL, which means that the skewness between the left and the CL is greater, and the data distribution is close to the left skewed distribution. Therefore, the data in the left-hand area has a larger diffusion of data, meaning that more potential information is hidden in this region. Thus, more virtual samples can be explored in that domain. On the other hand, the MTD method assumes the data is fitted in a unimodal triangular MF, and it ignores the possibility that the data may come from a multi-modal triangular MF. This is illustrated in Scenarios C and D in Figure 2(b). The data in Scenario C is fitted in two triangular MFs, where the distance between the two MFs is relatively large. Therefore, it is regarded as multiple triangular MFs. The data in Scenario D indicates that when two triangular MFs have overlapping regions, this means that the similarity between the data is high, and it also means that potential information is hidden in the overlapping region. Therefore, in this paper, the two triangular MFs are fused into a trapezoidal MF.
As mentioned above, data diffusion depends on the degree of data skewness. When extreme values exist in a small dataset, the estimated skewness is excessively left or right, which results in over-diffusion of data. Most virtual samples generated in an over-extended domain are unsuitable and cannot effectively improve the learning of models. To address this problem, the MD-MTD method can be used to modify the estimation of data skewness in order to avoid excessive data diffusion. However, because data skewness in MD-MTD is still determined by the amount of data, the problem where different small datasets result in the same data diffusion shown in Scenario A and Scenario B still exists and thus needs to be solved. In addition, although the MD-MTD method estimates the CL of multi-distributions, it does not consider multiple CLs for multi-distributions. Therefore, in this paper, a distance-based mega-trend-diffusion (DB-MTD) method is proposed to extend the data range and reduce the influences of extreme values when there are small datasets. In our method, data skewness is determined by the distances between examples. When most of the examples are farther from the CL, this means that a significant amount of information is hidden between these examples and the CL, and vice versa. In addition, in this paper, a plausibility assessment mechanism (PAM) is proposed to select appropriate virtual samples generated by multiple triangular MFs or one trapezoidal MF.
This study validates the proposed DB-MTD approach using one bladder cancer classification case from Liao [6] to diagnose bladder cancer and a regression case from Tsai et al. [7] to predict the effects of treatment of bladder cancer patients with radiotherapy. In this study, a data-driven model using a back-propagation neural network (BPNN) was used to compare the learning performance between the proposed DB-MTD method and other VSG methods. The experimental results of the two bladder cancer datasets showed that the learning performance of the proposed method outperformed the other VSG methods. A paired-test was also used to verify whether there were statistically significant differences in the BPNN model for the two bladder cancer datasets.
The remainder of this study is organized as follows: Section 2 is a review of the literature on VSG methods, as well as a brief introduction to the learning tools used in the experimental validation. Section 3 explains the details of the proposed DB-MTD. Section 4 is the description of the two bladder cancer datasets and provides the experimental results. The conclusions are given in Section 5.
Many studies have attempted to improve the learning performance of small datasets by generating virtual samples that been added into the original dataset to expand the amount of training data [24,25,28,33,31]. The generation of virtual samples has been verified as an effective strategy by which to extract meaningful information from a small dataset and increase the forecasting accuracy of models. Three related studies on virtual sample generation are reviewed in this section.
Huang and Moraga [25] proposed the DNN method, which applies a diffusion function to generate virtual samples on both sides of a given data point. In their approach, the data point is designated as the midpoint of a fuzzy normal function. After determining the possibility of n midpoints, the DNN approach then carried out symmetric diffusion on both sides of the midpoint to derive 2n virtual samples from Eqs (1)–(5). In Eqs (1) and (2), u and v are virtual samples of the input and output derived by using the original data; hset is the diffusion coefficient obtained from Eq (3); "r" represents the linear correlation coefficient of the input x and the output y, and ψ(r) represents the membership function value of the linear correlation coefficient r, as shown in Eq (5). After obtaining the corresponding values of the fuzzy normal membership function using Eqs (4), (1) and (2) are applied to produce virtual samples, u and v, after which the possibilities of the virtual samples are entered into a neural network for training, as shown in Figure 3.
u=x±√−2h2setlnψ(r) | (1) |
hset={0.6841(b−a), forn=50.5404(b−a), forn=60.4482(b−a), forn=70.3839(b−a), forn=82.6851(b−a)n−1, forn⩾9, wherea=min1⩽i⩽n{xi}, b=max1⩽i⩽n{xi} | (2) |
v=y±√−2h2setlnψ(r) | (3) |
poss(x,u)=1hset√2πe[−(x−u)22h2] | (4) |
ψ(r)=ψ(0.9+m×10−2)↦0.9...9⏟m9s,∀r∈{0.91,0.92,...,0.99},m=1,...,9 | (5) |
Li et al. [28] proposed the mega-trend-diffusion (MTD) method to estimate the triangular membership function used to generate virtual samples. This method assumes that data follows the assumptions below:
1) Each variable is independent of the other.
2) Each variable is fitted into a unimodal distribution.
The virtual sample generation of data range can be estimated using Eqs (6) and (7).
a=uset−SkewL×√−2׈s2xNL×ln(φ) | (6) |
b=uset+SkewU×√−2׈s2xNU×ln(φ) | (7) |
where uset=(max+min)/2 is the midpoint, and the coefficients of skewness are defined as SkewL=NLNL+NU and SkewU=NUNL+NU, which refers to the degree of diffusion on two sides of uset; √−2 ׈s2xNL× ln(φ) and √−2 ׈s2xNU× ln(φ) are defined as the proportion of the degree of diffusion, and φ=10−20 and ˆs2x represent the sample variance, where the virtual sample is generated randomly between interval [a, b], as shown in Figure 4(a).
Zhu et al. [31] presented the multi-distributions MTD method to avoid over-diffusion of small data. They modified the MTD method to avoid extreme values to influence estimations of data skewness and CL. In their paper, the data skewness and CL are modified as:
SkewL=NLNL+NU+m | (8) |
SkewU=NUNL+NU+m | (9) |
uset={x[n2+1] , if n is odd12(x[n2]+x[n2+1]), if n is even | (10) |
where m is a shape parameter used to adjust the degree of data skewness, and x[·] is an order statistic. The m is set to one in their method. The NL (NU) is the number of values smaller (greater) than uset. The range estimation of MD-MTD method is defined by Eqs (6) and (7). Due to the adjusted skewness on the left and right side of CL, the data diffusion is reduced into a narrower domain [a, b], as shown in Figure 4(b).
The back-propagation neural network (BPNN) was adopted in the present study as the data-driven model, which is a supervised learning model applicable for classification or forecasting problems. The input and output of the training data were used to build a neural network, and the weight of each node could be adjusted based on the difference between the degree of the expected output value and the actual observed value between each layer. A mapping function was built to process the weights in the different layers, as shown in Eq (11).
Y=f (P∑i=1WiXi)+b | (11) |
The output result for the input can be expressed as Eq (11), where Xi and Y are ith input variable and one output variable, respectively; W is the weighting value between the node and other nodes in the layer; f(⋅) is an activate function for nodes; b represents the product-sum of the weights, which must be greater than an error value, so the input nodes can be transferred into other nodes between linking layers, and p is the number of inputs from other nodes. In this paper, f(⋅) is set as the sigmoid activation function to update the weights of the nodes in the jth hidden layer as follows:
fj(x)=11+e−x, 0<fj(x)<1 | (12) |
In a basic BPNN scheme, nodes are composed of so-called "layers, " as shown in Figure 5, where the input and output layers are, respectively, the attributes and target of data. A hidden layer is mainly used to process the data, and the number of layers is determined according to the data complexity. Generally speaking, a higher number of hidden layers and more nodes produce fewer learning errors, but the network scheme in the training process is more complex and has a longer convergence time.
In this paper, we propose a unique distance-based mega-trend-diffusion method, including the skewness degree, the diffusion coefficient, the selection of uset, and the number of uset. This section illustrates the proposed method in detail, including the plausibility assessment mechanism (PAM) in Subsection 3.8 used to select suitable virtual samples.
Assume that a training dataset has n examples with m-1 input attributes {Xj|j=1,2,...,m−1}, and one output target Xm, denoted as T={(→x1,→y1), (→x2,→y2),...,(→xi,→yi)}, i=1,2,⋯,n, where →xi={xi,1,xi,2,...,xi,m−1} has m-1 variables and →yi is the target of →xi. The elements of the variables are {xi,j| i=1,2,...,n;j=1,2,...,m} in the training dataset, as shown in Table 1.
NO. | Input attributes | Output | ||||
X1 | … | Xj | … | Xm−1 | Xm | |
1 | x1,1 | … | x1,j | … | x1,m−1 | x1,m |
… | … | … | … | … | … | … |
i | xi,1 | … | xi,j | … | xi,m−1 | xi,m |
… | … | … | … | … | … | … |
n | xn,1 | … | xn,j | … | xn,m−1 | xn,m |
This subsection introduces the Min-Max data normalization method and k-nearest neighbor (kNN) algorithm used for imputation of missing data.
Because the proposed method is based on the distances between examples to measure the degree of data diffusion, before performing our proposed method, a data normalization process is necessary to avoid the effects from different variable scales that affect the data diffusion measurement. In this paper, we applied the Min-Max data normalization process to transform the domain of original data xi,j into [0, 1]. The transforming formula is expressed as:
x∗i,j=xi,j−minjmaxj−minj∈[0,1] | (13) |
where x∗i,j is transformed value and maxj (minj) is the maximum(minimum) value of the Xj variable. Min-Max normalization is a linear transformation intended to maintain the relative distances between the original examples. As a result, the proposed method can be effectively performed with transformed examples for small datasets with different variable scales.
When missing values occur in the original dataset, they increase the amount of data diffusion measurement error due to an incomplete data structure. There are two types of methods that can be used to deal with missing values. One is the ignoring missing value method in which examples with missing values are deleted, and the other is the data imputation method, where missing values are replaced with plausible values. In this paper, we adopt the data imputation method to fill in missing values instead of deleting them due to the very limited amount of data. Common data imputation methods include mean imputation, imputation with distributions, kNN imputation, etc. Some studies have suggested that kNN has better imputation performance than other methods [35,36,37,38]. For example, Jadhav et al. [38] compared kNN with other methods in terms of data imputation, and kNN significantly outperformed the other methods for numerical datasets. Because the proposed method is applicable to numerical variables, we use the kNN imputation method to handle data with missing values.
The kNN is an instance-based algorithm proposed by Cove and Hart [39]. This algorithm classifies examples based on similarity measurements that are made between examples. For data imputation, the kNN is used to impute missing values by selecting the nearest neighbors of examples with missing values. In this paper, we use the Euclidean distance as the similarity measurement to select the nearest neighbors of the example, as follows:
Euclidean distance=√k∑i=1(xi,j−xi,j)2 | (14) |
where k is the number of nearest neighbors, and j represents jth data variable. We illustrate a simple example of kNN for data imputation in Figure 6. When we set k as three, the original data with a missing value has two neighbors, where the missing value can be imputed using the nearest neighbors.
In the original MTD, the number of samples occurring on two sides of uset is used to determine the degree to which the data is skewed. When the number of samples is significantly larger on one side, the MTD formula is not suitable for estimating the data distribution. In this paper, a unique MTD method is proposed based on the distance between samples. The idea of this paper is that when the sum of the distances between the samples on one side is less than that on the other side, this indicates that the sample distribution is skewed to one side.
The distance between the samples is defined by using the distance equation derived from the Minkowski Distance, as noted by Cha [40], where Distance=(d∑i=1|Pi−Qi|h)1h. Since in this paper, it is assumed that each variable is independent, only a 1-D data distribution is considered. That is, Distance=d∑i=1|Pi−Qi|, where Pi represents the ith sample point xi, and Qi is the uset of variables. The degree to which the data is skewed is considered in Eqs (18) and (19).
G=n∑i=1|xi−uset| | (15) |
GL=n∑i=1|xi−uset|=n∑i=1uset−xi,forxi<uset | (16) |
GU=n∑i=1|xi−uset|=n∑i=1xi−uset,forxi>uset | (17) |
SkewL=GLGL+GU | (18) |
SkewU=GUGL+GU | (19) |
where GL (or GU) represents the sum of the distance between the smaller (or greater) sample, as compared to uset and uset. Figure 7(a) (or (b)) shows that if SkewL is smaller (or greater) than SkewU, the data distribution is inferred as being a right (or left) skewed distribution.
In the MTD, the diffusion coefficient is set as ˆs2xNL or ˆs2xNU to fine-tune the degree of diffusion on two sides of uset. When extreme values exist in the dataset, then the variations among the samples are significant, so the sample variance ˆs2x leads to excessive effects on the estimate of the diffusion coefficient hset. In order to reduce this effect, hset is set as |ˆsx|GL or |ˆsx|GU. Therefore, in order to reduce the influence of outliers by using the distance function G on two sides of uset, we adjust the diffusion coefficient as follows: √−2 ×|ˆsx|GL× ln(φ) or √−2 ×|ˆsx|GU× ln(φ). If GL is greater than GU, then it will be skewed to the left as a small hset, and if there are potential outliers, then the G value in this direction will be increased. Thus, the proposed hset can be used to mitigate the increasingly excessive diffusion.
This selection of uset is often seriously affected when there are many extreme values in a dataset, and thus, the midpoint uset that is obtained may not represent the tendency of the dataset although uset can be set as the mean of the samples, and the mean can then be used as the estimation of the data tendency when the number of samples is sufficient. However, setting the mean of uset is unstable in terms of estimating the data tendency in small samples. Thus, when considering the estimation of the tendency of the dataset for a small dataset, the median and mode are considered to be slightly affected by extreme values. k modes can be used to discover the frequency of repeated values in the dataset to predict the tendency of the dataset. Therefore, the proposed setting of uset can be defined as follows:
uset,k={Me, where k=1Mok, where 2⩽k⩽⌊n2⌋ | (20) |
Me={x[n2]+x[n2+1]2, for n=2r, where ∀r∈Nx[n+12], for n=2r+1, where ∀r∈N | (21) |
Mok=xi, for i=1,2,...,n, where 2⩽k⩽⌊n2⌋ | (22) |
Given a dataset X={x1,x2,...,xn}, n represents the number of samples, and k is the number of modes, where the mode is defined as identical data that occurs more than twice in a dataset. If no mode exists in a dataset, then the uset is the median.
Assuming a dataset X={x1,x2,...,xn}, the forms of Eqs (6), (7) and Eqs (16), (17) can be modified with k uset as follows:
ak={uset,k−SkewL,k×√−2×|ˆsx|GL,k×ln(φ)uset,k, where GL,k=0 | (23) |
bk={uset,k+SkewU,k×√−2×|ˆsx|GU,k×ln(φ)uset,k,where GU,k=0 | (24) |
GL,k=n∑i=1|xi−uset,k|=n∑i=1uset,k−xi, for xi<uset,k | (25) |
GU,k=n∑i=1|xi−uset,k|=n∑i=1xi−uset,k, for xi>uset,k | (26) |
Eqs (25) and (26) represent the sum of the distances between xi and uset,k. GL,k and GU,k represent the sum of the distances between the sample points fewer and greater than kth uset, respectively. Therefore, in this case, we use the distance function on two sides of kth uset as an expression of skewness, as shown in Eqs (27) and (28).
SkewL,k=GL,kGL,k+GU,k | (27) |
SkewU,k=GU,kGL,k+GU,k | (28) |
In Eqs (23) and (24), considering the diffusion coefficient presented in Section 3.4, two specific cases are discussed: If the kth mode is one of the extreme values, this will cause GL,k or GU,k to be close to 0 because xi≈uset→|xi−uset|→0; then, the diffusion in the direction of the extreme value cannot be defined, and the lower bound ak or the upper bound bk of the kth mode is set as uset,k. In addition, when the total number of samples is one, both GL,k and GU,k are 0.
Since the previous MTD method set uset as the midpoint, the method is susceptible to extreme values. In addition, the method assumes that the data is fitted into a unimodal distribution, but the data may come from a multi-modal distribution. For these two reasons, the new DB-MTD method proposed in this paper generates virtual samples from a unimodal distribution or a multi-modal distribution, where the median or mode is set as uset to reduce the influence of outliers. The process proposed in the DB-MTD used to generate virtual samples is shown in Figure 8.
When using the proposed method to estimate the data range, some possible scenarios are illustrated as follows:
1) When a dataset does not have a mode, the median is taken as the uset, as shown in Figure 9, Scenario A. Or, when the mode exists in the dataset, then uset is the mode, as shown in Figure 9, Scenario B.
2) When there is only one mode, the uset is the extreme minimum, as shown in Figure 9, Scenario C. If uset is the extreme maximum, the estimation of membership function is as shown in Figure 9, Scenario D.
3) When there is more than one mode, then it is necessary to generate more than one triangular fuzzy membership function, and we thus need to consider whether multiple triangular membership functions will overlap or not.
a. When there is no overlapping area between any two modes, there are two triangular membership functions, as shown in Figure 9, Scenario E.
b. When these two triangular fuzzy membership functions overlap, there is one trapezoidal membership function, as shown in Figure 9, Scenario F.
Since the overlapping of two triangular fuzzy membership functions in Scenarios E and F mean that important data may be located in the overlapping area, the value of the membership function of the sample point is set as 1. In other words, we thus set the possibility of the virtual sample being located in the interval as 1.
After estimating the data range given by Eqs (23) and (24), the virtual sample is randomly generated within the specified range with a uniform distribution.
As mentioned above, we discuss possible scenarios of the membership functions, as shown in Figure 9, where the peak of the membership function is set at 1, and the membership function value of the upper bound and lower bound is set at 0. When k uset exists, and sample x lies in the interval [ ak, bk ], the value of the membership function of x can be calculated when multiple triangular membership functions do not overlap each other, using Eq (29). When multiple triangular membership functions overlap each other, trapezoidal membership functions are used, as shown in Eq (30).
Triangular MF(x)={x−akuset,k−ak, as x<uset,kbk−xbk−uset,k, as x>uset,k1, as x=uset,k0, as x=ak or bk | (29) |
Trapezoidal MF(x)={x−min(ak)min(uset,k)−min(ak), as x∈[min(ak),min(uset,k)]max(bk)−xmax(bk)−max(uset,k), as x∈[max(uset,k),max(bk)]1, as x∈[min(uset,k),max(uset,k)]0, as x=min(ak) or max(bk) | (30) |
According to the virtual sample generation method, the values of the fuzzy membership functions are used to randomly generate virtual samples in order to fill the gaps in the data in the case of small sample learning. In this paper, a procedure was developed for a PAM that could be used to select the virtual samples based on an inferred data distribution so as to increase learning robustness. This mechanism can be used to examine whether the randomly generated virtual values are qualified or not, and it is described as follows in detail: Firstly, a tv (temporary value) is randomly generated from the estimated interval [a, b], after which the corresponding membership function values, MF (tv), are discovered, as shown in Figure 10.
Three possible scenarios are illustrated, as follows: In Figure 10, Scenario A only has one uset; Scenario B is two triangular membership functions, and Scenario C is a trapezoid fuzzy membership function with two uset. When determining whether tv can be kept as a virtual sample or not, we additionally generate a rs (random seed) ranging between [0, 1], as shown in Figure 10. This random seed is defined based on a uniform distribution. In the testing process, if the random seed is less than the membership function value of tv, then this tv will be kept as a suitable virtual sample v; if not, this tv will be abandoned, and a new virtual value will be generated. When a tv is close to the uset, then the membership function value of tv is close to 1, and thus, it will have a higher likelihood of being be kept as a qualified virtual sample, as shown in Eq (31).
v=tv, as rs⩽MF(tv) | (31) |
This section illustrates the experimental procedures designed for the purposes of this study. We compare the proposed method with other VSG methods with two bladder cancer cases. The experiments were implemented with Python 3.8.10. The experimental results are discussed in the following section.
Two bladder cancer cases are examined in our experiments. One case provided by Liao [6] is used to classify whether a patient suffers from bladder cancer, and the other case was obtained in [7] to predict a bladder cancer patient's resistance to radiotherapy.
The BC dataset comprises nine bladder cancer patients and nine healthy persons, as listed in Table 2. The first nine examples are bladder cancer patients defined as class 1, and the last nine samples are healthy persons defined as class 0. Thirteen gene proteins MDR, Topo II, Rb, EGFR, Neu, c-ErbB-3, c-ErbB-4, Cyclin A, Cyclin D1, P16, Cdc 2, Bcl-2, and Bax extracted in bladder cell lines are set as the input attributes.
Bladder cell lines | Input attributes | Output | |||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | class | ||
1 | BFTC-905 | 1.8 | 0.1 | 2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | 0.1 | 2 | 0.1 | 1.5 | 1 | ||
2 | HT 1197 | 1 | 1.3 | 1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | 2.8 | 2.5 | 1 | 1.3 | 1 | ||
3 | T24 | 1.5 | 3.5 | 0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | 2.3 | 2 | 2 | 0.1 | 1 | ||
4 | TCC-SUP | 0.1 | 1 | 1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | 2.3 | 2.8 | 0.1 | 0.2 | 1 | ||
5 | J82 | 0.3 | 1.3 | 2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | 0.1 | 2.5 | 1.3 | 0.2 | 1 | ||
6 | TSGH-8301 | 2 | 0.3 | 0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | 0.1 | 2.5 | 1.3 | 1.5 | 1 | ||
7 | HT 1376 | 1 | 2 | 1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | 1.3 | 2.5 | 0.1 | 0.8 | 1 | ||
8 | Sca-BER | 1 | 1 | 0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | 0.5 | 2.8 | 1.8 | 0.3 | 1 | ||
9 | 5637 | 1.2 | 0.8 | 1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | 0.1 | 2.5 | 0.1 | 1.3 | 1 | ||
10 | A | 0.3 | 0.1 | 6 | 0.5 | 0.1 | 0.2 | 0.5 | 0.1 | 0.5 | 5 | 0.5 | 5 | 8 | 0 | ||
11 | B | 0.1 | 0.3 | 4 | 1 | 0.2 | 0.1 | 2 | 0.3 | 0.1 | 4 | 0.5 | 9 | 5 | 0 | ||
12 | C | 0.2 | 0.1 | 5 | 0.5 | 0.1 | 0.1 | 1.5 | 0.2 | 0.1 | 6 | 1 | 7 | 5 | 0 | ||
13 | D | 0.1 | 0.8 | 9 | 1.5 | 0.1 | 0.5 | 0.5 | 0.1 | 1 | 6 | 0.8 | 6 | 6 | 0 | ||
14 | E | 0.5 | 0.1 | 6 | 2 | 0.5 | 0.1 | 0.5 | 0.3 | 0.1 | 5 | 0.5 | 4 | 4 | 0 | ||
15 | F | 0.5 | 0.5 | 6 | 2 | 0.1 | 0.8 | 0.8 | 0.1 | 0.2 | 6 | 1 | 5 | 5 | 0 | ||
16 | G | 0.1 | 1 | 4 | 1 | 0.3 | 0.2 | 0.5 | 0.5 | 0.2 | 5 | 1.5 | 5 | 7 | 0 | ||
17 | H | 0.5 | 0.1 | 5 | 0.5 | 0.2 | 0.1 | 1 | 0.1 | 0.1 | 8 | 0.5 | 7 | 7 | 0 | ||
18 | I | 0.1 | 0.1 | 7 | 0.8 | 0.1 | 0.1 | 2 | 0.2 | 0.5 | 5 | 0.5 | 6 | 4 | 0 |
The RBC dataset has 36 examples used to predict the resistance to Co-60 radiotherapy for bladder cancer cells. Each example has thirteen gene proteins, as mentioned above, and one additional attribute, the energy of gamma radiation from the Co-60 isotope, as listed in Table 3. The example output represents the patients' resistance to radiotherapy treatment.
Bladder cell lines | Input attributes | Output | ||||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 | Resistance to radiotherapy | ||
1 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 5 | 97 | ||
2 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 10 | 90 | ||
3 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 20 | 84 | ||
4 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 30 | 82 | ||
5 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 5 | 92 | ||
6 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 10 | 78 | ||
7 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 20 | 72 | ||
8 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 30 | 73 | ||
9 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 5 | 94 | ||
10 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 10 | 70 | ||
11 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 20 | 50 | ||
12 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 30 | 41 | ||
13 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 5 | 84 | ||
14 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 10 | 70 | ||
15 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 20 | 48 | ||
16 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 30 | 40 | ||
17 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 5 | 82 | ||
18 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 10 | 57 | ||
19 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 20 | 39 | ||
20 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 30 | 36 | ||
21 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 5 | 90 | ||
22 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 10 | 58 | ||
23 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 20 | 39 | ||
24 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 30 | 34 | ||
25 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 5 | 83 | ||
26 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 10 | 50 | ||
27 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 20 | 32 | ||
28 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 30 | 28 | ||
29 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 5 | 60 | ||
30 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 10 | 30 | ||
31 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 20 | 29 | ||
32 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 30 | 30 | ||
33 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 5 | 63 | ||
34 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 10 | 33 | ||
35 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 20 | 21 | ||
36 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 30 | 18 |
We randomly drew 10 samples in the RBC dataset as an example to explain the different scenarios in the proposed DB-MTD method. The drawn samples were set as a training dataset, as listed in Table 4. The implementation procedure for the DB-MTD method is explained in the following discussion.
Bladder cell lines | Input attributes | |||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 | |
1 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 5 | |
2 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 5 | |
3 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 5 | |
4 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 10 | |
5 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 30 | |
6 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 30 | |
7 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 20 | |
8 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 5 | |
9 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 30 | |
10 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 20 |
We use the Min-Max data normalization process from Eq (13) to transform the training data domain into [0, 1]. The lower bound a and upper bound b of the input attributes can be obtained from Eqs (23) and (24), as shown in Table 5. The Rb, Cyclin D1, Bax attributes are determined to be the trapezoidal MF based on our method, where we could obtain two CLs. The remaining attributes were defined as the triangular MF.
Input attributes | ||||||||||||||
MF type | Tri | Tri | Trap | Tri | Tri | Tri | Tri | Tri | Trap | Tri | Tri | Tri | Trap | Tri |
estimations | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 |
a1 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
uset,1 | 0.643 | 0.315 | 0.5 | 0.176 | 0.333 | 0.219 | 0.062 | 0.105 | 0.3 | 0.185 | 0.625 | 0.474 | 0.0 | 0.4 |
b1 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
a2 | - | - | 0.0 | - | - | - | - | - | 0.0 | - | - | - | 0.0 | - |
uset,2 | - | - | 0.8 | - | - | - | - | - | 0.9 | - | - | - | 1.0 | - |
b2 | - | - | 1.0 | - | - | - | - | - | 1.0 | - | - | - | 1.0 | - |
Note: "Tri" and "Trap" represents that the attribute is fitted by the triangular and trapezoidal MF, respectively. |
To explain the PAM used to select the virtual samples, we use the MDR and Rb attributes as examples. The virtual sample (vs) could be created based on different MFs in the two attributes, as shown in Figure 11. Then, the vs was randomly generated within [a, b] to calculate the possibility MF(vs). When the MF(vs) was greater than rs~Uniform(0, 1), then vs could be regarded as an appropriate virtual sample.
After performing the PAM, the virtual samples could be generated, as shown in Table 6. The virtual sample output was derived using the built BPNN model with small datasets. Then, we added the generated virtual examples into the original training dataset to build a new training dataset.
NO. | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 |
1 | 0.336 | 0.631 | 0.032 | 0.104 | 0.600 | 0.201 | 0.186 | 0.889 | 0.127 | 0.05 | 0.223 | 0.696 | 0.407 | 0.425 |
2 | 0.862 | 0.075 | 0.806 | 0.278 | 0.329 | 0.109 | 0.837 | 0.581 | 0.353 | 0.113 | 0.145 | 0.529 | 0.305 | 0.052 |
3 | 0.966 | 0.141 | 0.231 | 0.381 | 0.534 | 0.338 | 0.169 | 0.377 | 0.57 | 0.171 | 0.774 | 0.171 | 0.444 | 0.165 |
4 | 0.937 | 0.656 | 0.256 | 0.715 | 0.047 | 0.561 | 0.314 | 0.727 | 0.108 | 0.700 | 0.483 | 0.488 | 0.024 | 0.477 |
5 | 0.515 | 0.514 | 0.818 | 0.754 | 0.882 | 0.912 | 0.647 | 0.178 | 0.037 | 0.39 | 0.904 | 0.554 | 0.224 | 0.012 |
6 | 0.129 | 0.147 | 0.139 | 0.685 | 0.486 | 0.576 | 0.485 | 0.646 | 0.148 | 0.782 | 0.04 | 0.871 | 0.491 | 0.724 |
7 | 0.468 | 0.595 | 0.631 | 0.289 | 0.631 | 0.877 | 0.581 | 0.059 | 0.185 | 0.721 | 0.513 | 0.873 | 0.035 | 0.347 |
8 | 0.882 | 0.648 | 0.877 | 0.774 | 0.722 | 0.734 | 0.124 | 0.863 | 0.260 | 0.459 | 0.865 | 0.646 | 0.175 | 0.173 |
9 | 0.795 | 0.467 | 0.528 | 0.189 | 0.775 | 0.433 | 0.295 | 0.246 | 0.778 | 0.442 | 0.031 | 0.076 | 0.615 | 0.466 |
10 | 0.208 | 0.173 | 0.741 | 0.615 | 0.927 | 0.338 | 0.136 | 0.762 | 0.717 | 0.629 | 0.439 | 0.899 | 0.32 | 0.292 |
In our experiment, we used the random sampling method to create different scenarios for small sample sizes. The different training datasets were created by sampling from the original datasets with the BC dataset ranging in size from 3 to 17 and training data sizes {5, 10, 15, 20, 25, 30, 35} for the RBC dataset, and the remaining data was set as a testing dataset. Then, we used Eqs (23) and (24) to calculate the lower and upper bounds of the attributes of the training dataset. The virtual sample could be randomly generated from the evaluated range. In addition, we calculated the MF values of the virtual samples to select virtual samples using the PAM. The virtual data size was set as 100. The f-k-l BPNN network was set as a learning model in the experiment, where f represents f attributes of the training dataset as the number of nodes in the input layer; k is the number of nodes in the hidden layer, and l is the number of nodes in the output layer. The testing dataset was inputted into the learned BPNN model with 100 nodes in the hidden layer and 50 epochs in the training process. In the BC dataset, we used classification accuracy as the evaluation metric for the BPNN model. In the RBC dataset, the prediction performance of the BPNN model was evaluated using the root mean squared error (RMSE), which was defined as:
RMSE=√1testNtestN∑i=1(yi−ˆyi)2 | (32) |
where testN is the number of testing samples; yi is the actual value, and ˆyi is the predicted values of the ith testing sample. A total of 100 iterations were used in this experiment. After performing the experiments, we calculated the average and standard deviation of the classification accuracy for the BC dataset and the average and standard deviation of the RMSE for the RBC dataset.
To verify the effectiveness of the proposed DB-MTD method, we compared it using four methods: RAW (using raw training samples), MTD (generating virtual samples based on the triangular MF), GD (generating virtual samples base on a normal distribution), and MD-MTD (generating virtual samples based on multiple distributions). For the BC dataset, the average (Avg-accuracy) and the standard deviation (SD-accuracy) of the classification accuracies are respectively shown in Figures 12 and 13. The improvements (%) in classification accuracy using the proposed DB-MTD method are listed in Table 7. In Figure 12, it can be seen that the results for most of the training subsets indicated superior classification accuracies as compared to those for the other methods under consideration. The SD-accuracy in these methods had smaller variances that were close to each other with different training data sizes. For example, when the training sample size was 3, the Avg-accuracy using the DB-MTD method was improved from 86.2 to 89.9%, where the improvement was approximately 3.7%, as shown in Table 7. For the RBC dataset, the average (Avg-RMSE) and the standard deviation (SD-RMSE) of RMSE are respectively shown in Figures 14 and 15. In Figure 14, the Avg-RMSE using DB-MTD method was improved from 24.79 to 24.47 at a training sample size of 5, where the improvement was approximately 0.32. The other improvements are shown in Table 8.
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
3 | 3.70 | 4.00 | 0.80 | 3.90 |
4 | 2.60 | 2.70 | 0.50 | 2.50 |
5 | 2.30 | 1.80 | 0.20 | 2.20 |
6 | 2.00 | 1.80 | 0.50 | 1.90 |
7 | 2.40 | 1.80 | 1.20 | 1.80 |
8 | 1.60 | 1.10 | 0.80 | 1.20 |
9 | 0.90 | 0.50 | 0.50 | 0.20 |
10 | 0.70 | 0.00 | 0.20 | 0.30 |
11 | 0.70 | 0.70 | 1.20 | 0.20 |
12 | 0.30 | -0.20 | 0.80 | -0.20 |
13 | 0.60 | 0.60 | 0.90 | 0.20 |
14 | 0.00 | 0.00 | 0.40 | 0.10 |
15 | 0.60 | 0.30 | 0.70 | 0.10 |
16 | 0.30 | 0.30 | 0.50 | 0.30 |
17 | 0.00 | 0.00 | 0.30 | 0.30 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
5 | 0.32 | 0.03 | 0.05 | 0.01 |
10 | 0.37 | 0.07 | 0.16 | 0.06 |
15 | 0.47 | 0.15 | 0.23 | 0.07 |
20 | 0.87 | 0.18 | 0.25 | 0.22 |
25 | 1.45 | 0.29 | 0.40 | 0.37 |
30 | 1.31 | 0.51 | 0.65 | 0.43 |
35 | 2.61 | 0.59 | 0.16 | 0.55 |
A paired t-test was used to verify the significance of the learning improvements using the proposed DB-MTD method in BC and RBC datasets. The p-values of the paired t-test for the four methods and the DB-MTD method are listed in Tables 9 and 10, where the symbol "*" indicates that the prediction results demonstrated a statistically significant difference (p-value < 0.05). In these tables, the results show statistical significance in terms of the difference in the Avg-accuracy for the BC dataset and the Avg-RMSE for the RBC dataset in some training data sizes. For example, when the training data size was set at 10 in the BC dataset, "0.02*" represents that the prediction results based on the DB-MTD demonstrated a significant difference in learning performance for the RAW method. Although the DB-MTD method was only statistically superior to the other methods in some training sizes, the Avg-accuracy of the DB-MTD method was greater than that when using the methods for BC dataset. It is the same results based on the Avg-RMSE were obtained for the RBC dataset, as shown in Table 10.
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
3 | 0.06 | 0.87 | 0.69 | 0.66 |
4 | 0.00* | 0.00* | 0.01* | 0.00* |
5 | 0.91 | 0.73 | 0.31 | 0.79 |
6 | 0.00* | 0.00* | 0.01* | 0.00* |
7 | 0.00* | 0.00* | 0.00* | 0.00* |
8 | 0.00* | 0.00* | 0.00* | 0.00* |
9 | 0.01* | 0.10 | 0.06 | 0.56 |
10 | 0.02* | 1.00 | 0.44 | 0.32 |
11 | 0.00* | 0.00* | 0.00* | 0.23 |
12 | 0.16 | 0.47 | 0.00* | 0.62 |
13 | 0.03* | 0.02* | 0.00* | 0.42 |
14 | 0.66 | 1.00 | 0.10 | 0.42 |
15 | 0.03* | 0.08 | 0.01* | 0.32 |
16 | 0.16 | 0.16 | 0.08 | 0.16 |
17 | Not applicable | Not applicable | 0.32 | 0.32 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
5 | 0.11 | 0.68 | 0.40 | 0.98 |
10 | 0.00* | 0.07 | 0.00* | 0.04* |
15 | 0.00* | 0.00* | 0.00* | 0.11 |
20 | 0.00* | 0.01* | 0.00* | 0.00* |
25 | 0.00* | 0.00* | 0.00* | 0.00* |
30 | 0.00* | 0.00* | 0.00* | 0.00* |
35 | 0.00* | 0.01* | 0.44 | 0.02* |
In many fields, the virtual sample generation (VSG) approach has been regarded as an effective method to improve the learning performance of machine learning models with small sample sizes. The popular MTD method has been widely applied in many VSG studies to generate virtual samples within the estimated data range to extend the amount of the original training data. The problem with the MTD method is that extreme values have serious effects, including making it difficult to estimate the central location and data skewness. To deal with this problem, a new distance-based MTD (DB-MTD) method based on a defined distance function between data was proposed in the present work to improve the degree of data diffusion. The distance became the basis of the coefficient of skewness, which made the inferred distribution close to the pattern in the existing samples. The proposed DB-MTD coefficient could more effectively reduce the excessive diffusion problem that existed in the original MTD method based on a premise where only the triangular membership function was considered.
As to the limitations of this method, the proposed DB-MTD method can be used for small data with continuous variables, but it is not suitable for categorical or discrete variables. In our future research, two research directions can be considered: One is finding discrete density functions to create virtual samples for categorical or discrete attributes. The other is validating the proposed method on practical medical datasets.
This study is supported by the Ministry of Science and Technology and National Taipei University of Nursing and Health Sciences, Taiwan. This research was funded by the Ministry of Science and Technology Taiwan under MOST 110-2222-E-227-001-MY2.
The authors declare that there are no conflicts of interest.
[1] |
P. Gontero, A. Tizzani, G. H. Muir, E. Caldarera, M. Pavone Macaluso, The genetic alterations in the oncogenic pathway of transitional cell carcinoma of the bladder and its prognostic value, Urol. Res., 29 (2001), 377–387. https://doi.org/10.1007/s002400100216 doi: 10.1007/s002400100216
![]() |
[2] |
V. Tut, K. Braithwaite, B. Angus, D. Neal, J. Lunec, J. Mellon, Cyclin D1 expression in transitional cell carcinoma of the bladder: correlation with p53, waf1, pRb and Ki67, Br. J. Cancer, 84 (2001), 270–275. https://doi.org/10.1054/bjoc.2000.1557 doi: 10.1054/bjoc.2000.1557
![]() |
[3] |
A. Colquhoun, S. Sundar, P. Rajjayabun, T. Griffiths, R. Symonds, J. Mellon, Epidermal growth factor receptor status predicts local response to radical radiotherapy in muscle-invasive bladder cancer, Clin. Oncol., 18 (2006), 702–709. https://doi.org/10.1016/j.clon.2006.08.003 doi: 10.1016/j.clon.2006.08.003
![]() |
[4] |
P. Luukka, Similarity classifier in diagnosis of bladder cancer, Comput. Methods Programs Biomed., 89 (2008), 43–49. https://doi.org/10.1016/j.cmpb.2007.10.001 doi: 10.1016/j.cmpb.2007.10.001
![]() |
[5] |
G. Y. Chao, T. I. Tsai, T. J. Lu, H. C. Hsu, B. Y. Bao, W. Y. Wu, et al, A new approach to prediction of radiotherapy of bladder cancer cells in small dataset analysis, Expert Syst. Appl., 38 (2011), 7963–7969. https://doi.org/10.1016/j.eswa.2010.12.035 doi: 10.1016/j.eswa.2010.12.035
![]() |
[6] |
T. W. Liao, Diagnosis of bladder cancers with small sample size via feature selection, Expert Syst. Appl., 38 (2011), 4649–4654. https://doi.org/10.1016/j.eswa.2010.09.135 doi: 10.1016/j.eswa.2010.09.135
![]() |
[7] |
T. I. Tsai, Y. Zhang, Z. Zhang, G. Y. Chao, C. C. Tsai, Considering relationship of proteins for radiotherapy prognosis of bladder cancer cells in small data set, Methods Inf. Med., 57 (2018), 220–229. https://doi.org/10.3414/ME17-02-0003 doi: 10.3414/ME17-02-0003
![]() |
[8] |
M. D. Robinson, G. K. Smyth, Small-sample estimation of negative binomial dispersion, with applications to SAGE data, Biostatistics, 9 (2008), 321–332. https://doi.org/10.1093/biostatistics/kxm030 doi: 10.1093/biostatistics/kxm030
![]() |
[9] |
S. Lee, M. J. Emond, M. J. Bamshad, K. C. Barnes, M. J. Rieder, D. A. Nickerson, et al., Optimal unified approach for rare-variant association testing with application to small-sample case-control whole-exome sequencing studies, Am. J. Hum. Genet., 91 (2012), 224–237. https://doi.org/10.1016/j.ajhg.2012.06.007 doi: 10.1016/j.ajhg.2012.06.007
![]() |
[10] |
Y. Zhao, N. J. Fesharaki, H. Liu, J. Luo, Using data-driven sublanguage pattern mining to induce knowledge models: application in medical image reports knowledge representation, BMC Med. Inf. Decis. Making, 18 (2018), 1–13. https://doi.org/10.1186/s12911-018-0645-3 doi: 10.1186/s12911-017-0580-8
![]() |
[11] |
L. Stainier, A. Leygue, M. Ortiz, Model-free data-driven methods in mechanics: material data identification and solvers, Comput. Mech., 64 (2019), 381–393. https://doi.org/10.1007/s00466-019-01731-1 doi: 10.1007/s00466-019-01731-1
![]() |
[12] |
E. Ntoutsi, P. Fafalios, U. Gadiraju, V. Iosifidis, W. Nejdl, M. E. Vidal, et al., Bias in data‐driven artificial intelligence systems—An introductory survey, Wiley Interdiscip. Rev.: Data Min. Knowl. Discovery, 10 (2020), e1356. https://doi.org/10.1002/widm.1356 doi: 10.1002/widm.1356
![]() |
[13] |
T. Mao, L. Yu, Y. Zhang, L. Zhou, Modified Mahalanobis-Taguchi System based on proper orthogonal decomposition for high-dimensional-small-sample-size data classification, Math. Biosci. Eng., 18 (2020), 426–444. https://doi.org/10.3934/mbe.2021023 doi: 10.3934/mbe.2021023
![]() |
[14] |
I. Izonin, R. Tkachenko, I. Dronyuk, P. Tkachenko, M. Gregus, M. Rashkevych, Predictive modeling based on small data in clinical medicine: RBF-based additive input-doubling method, Math. Biosci. Eng., 18 (2021), 2599–2613. https://doi.org/10.3934/mbe.2020392 doi: 10.3934/mbe.2021132
![]() |
[15] |
Y. Liu, Y. Zhou, X. Liu, F. Dong, C. Wang, Z. Wang, Wasserstein GAN-based small-sample augmentation for new-generation artificial intelligence: a case study of cancer-staging data in biology, Engineering, 5 (2019), 156–163. https://doi.org/10.1016/j.eng.2018.11.018 doi: 10.1016/j.eng.2018.11.018
![]() |
[16] |
H. Han, M. Zhou, Y. Zhang, Can virtual samples solve small sample size problem of KISSME in pedestrian re-identification of smart transportation?, IEEE Trans. Intell. Transp. Syst., 21 (2020), 3766–3776. https://doi.org/10.1109/TITS.2019.2933509 doi: 10.1109/TITS.2019.2933509
![]() |
[17] |
Z. Liu, Y. Li, Small data-driven modeling of forming force in single point incremental forming using neural networks, Eng. Comput., 36 (2020), 1589–1597. https://doi.org/10.1007/s00366-019-00781-6 doi: 10.1007/s00366-019-00781-6
![]() |
[18] |
Q. X. Zhu, Z. S. Chen, X. H. Zhang, A. Rajabifard, Y. Xu, Y. Q. Chen, Dealing with small sample size problems in process industry using virtual sample generation: a Kriging-based approach, Soft Comput., 24 (2020), 6889–6902. https://doi.org/10.1007/s00500-019-04326-3 doi: 10.1007/s00500-019-04326-3
![]() |
[19] |
N. V. Chawla, K. W. Bowyer, L. O. Hall, W. P. Kegelmeyer, SMOTE: synthetic minority over-sampling technique, J. Artif. Intell. Res., 16 (2002), 321–357. https://doi.org/10.1613/jair.953 doi: 10.1613/jair.953
![]() |
[20] | B. Efron, R. LePage, Introduction to Bootstrap, Wiley & Sons, New York, 1992. |
[21] |
S. Lee, A. Ahmad, G. Jeon, Combining bootstrap aggregation with support vector regression for small blood pressure measurement, J. Med. Syst., 42 (2018), 1–7. https://doi.org/10.1007/s10916-018-0913-x doi: 10.1007/s10916-017-0844-y
![]() |
[22] |
M. F. Ijaz, M. Attique, Y. Son, Data-driven cervical cancer prediction model with outlier detection and over-sampling methods, Sensors, 20 (2020), 2809. https://doi.org/10.3390/s20102809 doi: 10.3390/s20102809
![]() |
[23] |
M. La Rocca, C. Perna, Nonlinear autoregressive sieve bootstrap based on extreme learning machines, Math. Biosci. Eng., 17 (2020), 636–653. https://doi.org/10.3934/mbe.202003 doi: 10.3934/mbe.2020033
![]() |
[24] |
S. Cho, M. Jang, S. Chang, Virtual sample generation using a population of networks, Neural Process Lett., 5 (1997), 21–27. https://doi.org/10.1023/A:1009653706403 doi: 10.1023/A:1009653706403
![]() |
[25] |
C. Huang, C. Moraga, A diffusion-neural-network for learning from small samples, Int. J. Approx. Reasoning, 35 (2004), 137–161. https://doi.org/10.1016/j.ijar.2003.06.001 doi: 10.1016/j.ijar.2003.06.001
![]() |
[26] | I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et al., Generative adversarial nets, in Proceedings of the International Conference on Neural Information Processing Systems (NIPS), (2014), 2672–2680. |
[27] |
X. H. Zhang, Y. Xu, Y. L. He, Q. X. Zhu, Novel manifold learning based virtual sample generation for optimizing soft sensor with small data, ISA Trans., 109 (2021), 229–241. https://doi.org/10.1016/j.isatra.2020.10.006 doi: 10.1016/j.isatra.2020.10.006
![]() |
[28] |
D. C. Li, C. S. Wu, T. I. Tsai, Y. S. Lina, Using mega-trend-diffusion and artificial samples in small data set learning for early flexible manufacturing system scheduling knowledge, Comput. Oper. Res., 34 (2007), 966–982. https://doi.org/10.1016/j.cor.2005.05.019 doi: 10.1016/j.cor.2005.05.019
![]() |
[29] |
M. R. Rahimi, H. Karimi, F. Yousefi, Prediction of carbon dioxide diffusivity in biodegradable polymers using diffusion neural network, Heat Mass Transfer, 48 (2012), 1357–1365. https://doi.org/10.1007/s00231-012-0982-1 doi: 10.1007/s00231-012-0982-1
![]() |
[30] |
A. Majid, S. Ali, M. Iqbal, N. Kausar, Prediction of human breast and colon cancers from imbalanced data using nearest neighbor and support vector machines, Comput. Methods Programs Biomed., 113 (2014), 792–808. https://doi.org/10.1016/j.cmpb.2014.01.001 doi: 10.1016/j.cmpb.2014.01.001
![]() |
[31] |
B. Zhu, Z. Chen, L. Yu, A novel mega-trend-diffusion for small sample, CIESC J., 67 (2016), 820–826. https://doi.org/10.11949/j.issn.0438-1157.20151921 doi: 10.11949/j.issn.0438-1157.20151921
![]() |
[32] |
L. Yu, X. Zhang, Can small sample dataset be used for efficient internet loan credit risk assessment? Evidence from online peer to peer lending, Finance Res. Lett., 38 (2021), 101521. https://doi.org/10.1016/j.frl.2020.101521 doi: 10.1016/j.frl.2020.101521
![]() |
[33] |
J. Yang, X. Yu, Z. Q. Xie, J. P. Zhang, A novel virtual sample generation method based on Gaussian distribution, Knowl. Based. Syst., 24 (2011), 740–748. https://doi.org/10.1016/j.knosys.2010.12.010 doi: 10.1016/j.knosys.2010.12.010
![]() |
[34] |
K. Wang, J. Li, F. Tsung, Distribution inference from early-stage stationary data streams by transfer learning, ⅡSE Trans., (2021), 1–25. https://doi.org/10.1080/24725854.2021.1875520 doi: 10.1080/24725854.2021.1875520
![]() |
[35] |
O. Troyanskaya, M. Cantor, G. Sherlock, P. Brown, T. Hastie, R. Tibshirani, et al., Missing value estimation methods for DNA microarrays, Bioinformatics, 17 (2001), 520–525. https://doi.org/10.1093/bioinformatics/17.6.520 doi: 10.1093/bioinformatics/17.6.520
![]() |
[36] |
G. E. Batista, M. C. Monard, An analysis of four missing data treatment methods for supervised learning, Appl. Artif. Intell., 17 (2003), 519–533. https://doi.org/10.1080/713827181 doi: 10.1080/713827181
![]() |
[37] |
D. V. Nguyen, N. Wang, R. J. Carroll, Evaluation of missing value estimation for microarray data, Data Sci. J., 2 (2004), 347–370. https://doi.org/10.6339/JDS.2004.02(4).170 doi: 10.6339/JDS.2004.02(4).170
![]() |
[38] |
A. Jadhav, D. Pramod, K. Ramanathan, Comparison of performance of data imputation methods for numeric dataset, Appl. Artif. Intell., 33 (2019), 913–933. https://doi.org/10.1080/08839514.2019.1637138 doi: 10.1080/08839514.2019.1637138
![]() |
[39] |
T. Cover, P. Hart, Nearest neighbor pattern classification, IEEE Trans. Inf. Theory, 13 (1967), 21–27. https://doi.org/10.1109/TIT.1967.1053964 doi: 10.1109/TIT.1967.1053964
![]() |
[40] | G. H. Cha, Non-metric similarity ranking for image retrieval, in International Conference on Database and Expert Systems Applications: Springer, (2006), 853–862. https://doi.org/10.1007/11827405_83 |
1. | Ivan Izonin, Nataliya Shakhovska, Special issue: informatics & data-driven medicine-2021, 2022, 19, 1551-0018, 9769, 10.3934/mbe.2022454 | |
2. | Claudia C. Tusell-Rey, Oscar Camacho-Nieto, Cornelio Yáñez-Márquez, Yenny Villuendas-Rey, Customized Instance Random Undersampling to Increase Knowledge Management for Multiclass Imbalanced Data Classification, 2022, 14, 2071-1050, 14398, 10.3390/su142114398 | |
3. | Chaudhry Huma, Lee Hawon, Jagasia Sarisha, Tasci Erdal, Camphausen Kevin, Krauze Andra Valentina, Advances in the field of developing biomarkers for re-irradiation: a how-to guide to small, powerful data sets and artificial intelligence, 2024, 9, 2380-8993, 3, 10.1080/23808993.2024.2325936 | |
4. | Liang-Sian Lin, Yao-San Lin, Der-Chiang Li, Yi-Ting Chen, Mega trend diffusion-siamese network oversampling for imbalanced datasets’ SVM classification, 2023, 143, 15684946, 110406, 10.1016/j.asoc.2023.110406 | |
5. | Liang-Sian Lin, Improving small sample prediction performance via novel nonlinear interpolation virtual sample generation with self-supervised learning, 2024, 678, 00200255, 121044, 10.1016/j.ins.2024.121044 |
NO. | Input attributes | Output | ||||
X1 | … | Xj | … | Xm−1 | Xm | |
1 | x1,1 | … | x1,j | … | x1,m−1 | x1,m |
… | … | … | … | … | … | … |
i | xi,1 | … | xi,j | … | xi,m−1 | xi,m |
… | … | … | … | … | … | … |
n | xn,1 | … | xn,j | … | xn,m−1 | xn,m |
Bladder cell lines | Input attributes | Output | |||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | class | ||
1 | BFTC-905 | 1.8 | 0.1 | 2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | 0.1 | 2 | 0.1 | 1.5 | 1 | ||
2 | HT 1197 | 1 | 1.3 | 1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | 2.8 | 2.5 | 1 | 1.3 | 1 | ||
3 | T24 | 1.5 | 3.5 | 0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | 2.3 | 2 | 2 | 0.1 | 1 | ||
4 | TCC-SUP | 0.1 | 1 | 1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | 2.3 | 2.8 | 0.1 | 0.2 | 1 | ||
5 | J82 | 0.3 | 1.3 | 2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | 0.1 | 2.5 | 1.3 | 0.2 | 1 | ||
6 | TSGH-8301 | 2 | 0.3 | 0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | 0.1 | 2.5 | 1.3 | 1.5 | 1 | ||
7 | HT 1376 | 1 | 2 | 1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | 1.3 | 2.5 | 0.1 | 0.8 | 1 | ||
8 | Sca-BER | 1 | 1 | 0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | 0.5 | 2.8 | 1.8 | 0.3 | 1 | ||
9 | 5637 | 1.2 | 0.8 | 1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | 0.1 | 2.5 | 0.1 | 1.3 | 1 | ||
10 | A | 0.3 | 0.1 | 6 | 0.5 | 0.1 | 0.2 | 0.5 | 0.1 | 0.5 | 5 | 0.5 | 5 | 8 | 0 | ||
11 | B | 0.1 | 0.3 | 4 | 1 | 0.2 | 0.1 | 2 | 0.3 | 0.1 | 4 | 0.5 | 9 | 5 | 0 | ||
12 | C | 0.2 | 0.1 | 5 | 0.5 | 0.1 | 0.1 | 1.5 | 0.2 | 0.1 | 6 | 1 | 7 | 5 | 0 | ||
13 | D | 0.1 | 0.8 | 9 | 1.5 | 0.1 | 0.5 | 0.5 | 0.1 | 1 | 6 | 0.8 | 6 | 6 | 0 | ||
14 | E | 0.5 | 0.1 | 6 | 2 | 0.5 | 0.1 | 0.5 | 0.3 | 0.1 | 5 | 0.5 | 4 | 4 | 0 | ||
15 | F | 0.5 | 0.5 | 6 | 2 | 0.1 | 0.8 | 0.8 | 0.1 | 0.2 | 6 | 1 | 5 | 5 | 0 | ||
16 | G | 0.1 | 1 | 4 | 1 | 0.3 | 0.2 | 0.5 | 0.5 | 0.2 | 5 | 1.5 | 5 | 7 | 0 | ||
17 | H | 0.5 | 0.1 | 5 | 0.5 | 0.2 | 0.1 | 1 | 0.1 | 0.1 | 8 | 0.5 | 7 | 7 | 0 | ||
18 | I | 0.1 | 0.1 | 7 | 0.8 | 0.1 | 0.1 | 2 | 0.2 | 0.5 | 5 | 0.5 | 6 | 4 | 0 |
Bladder cell lines | Input attributes | Output | ||||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 | Resistance to radiotherapy | ||
1 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 5 | 97 | ||
2 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 10 | 90 | ||
3 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 20 | 84 | ||
4 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 30 | 82 | ||
5 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 5 | 92 | ||
6 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 10 | 78 | ||
7 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 20 | 72 | ||
8 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 30 | 73 | ||
9 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 5 | 94 | ||
10 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 10 | 70 | ||
11 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 20 | 50 | ||
12 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 30 | 41 | ||
13 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 5 | 84 | ||
14 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 10 | 70 | ||
15 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 20 | 48 | ||
16 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 30 | 40 | ||
17 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 5 | 82 | ||
18 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 10 | 57 | ||
19 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 20 | 39 | ||
20 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 30 | 36 | ||
21 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 5 | 90 | ||
22 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 10 | 58 | ||
23 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 20 | 39 | ||
24 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 30 | 34 | ||
25 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 5 | 83 | ||
26 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 10 | 50 | ||
27 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 20 | 32 | ||
28 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 30 | 28 | ||
29 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 5 | 60 | ||
30 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 10 | 30 | ||
31 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 20 | 29 | ||
32 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 30 | 30 | ||
33 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 5 | 63 | ||
34 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 10 | 33 | ||
35 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 20 | 21 | ||
36 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 30 | 18 |
Bladder cell lines | Input attributes | |||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 | |
1 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 5 | |
2 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 5 | |
3 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 5 | |
4 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 10 | |
5 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 30 | |
6 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 30 | |
7 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 20 | |
8 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 5 | |
9 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 30 | |
10 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 20 |
Input attributes | ||||||||||||||
MF type | Tri | Tri | Trap | Tri | Tri | Tri | Tri | Tri | Trap | Tri | Tri | Tri | Trap | Tri |
estimations | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 |
a1 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
uset,1 | 0.643 | 0.315 | 0.5 | 0.176 | 0.333 | 0.219 | 0.062 | 0.105 | 0.3 | 0.185 | 0.625 | 0.474 | 0.0 | 0.4 |
b1 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
a2 | - | - | 0.0 | - | - | - | - | - | 0.0 | - | - | - | 0.0 | - |
uset,2 | - | - | 0.8 | - | - | - | - | - | 0.9 | - | - | - | 1.0 | - |
b2 | - | - | 1.0 | - | - | - | - | - | 1.0 | - | - | - | 1.0 | - |
Note: "Tri" and "Trap" represents that the attribute is fitted by the triangular and trapezoidal MF, respectively. |
NO. | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 |
1 | 0.336 | 0.631 | 0.032 | 0.104 | 0.600 | 0.201 | 0.186 | 0.889 | 0.127 | 0.05 | 0.223 | 0.696 | 0.407 | 0.425 |
2 | 0.862 | 0.075 | 0.806 | 0.278 | 0.329 | 0.109 | 0.837 | 0.581 | 0.353 | 0.113 | 0.145 | 0.529 | 0.305 | 0.052 |
3 | 0.966 | 0.141 | 0.231 | 0.381 | 0.534 | 0.338 | 0.169 | 0.377 | 0.57 | 0.171 | 0.774 | 0.171 | 0.444 | 0.165 |
4 | 0.937 | 0.656 | 0.256 | 0.715 | 0.047 | 0.561 | 0.314 | 0.727 | 0.108 | 0.700 | 0.483 | 0.488 | 0.024 | 0.477 |
5 | 0.515 | 0.514 | 0.818 | 0.754 | 0.882 | 0.912 | 0.647 | 0.178 | 0.037 | 0.39 | 0.904 | 0.554 | 0.224 | 0.012 |
6 | 0.129 | 0.147 | 0.139 | 0.685 | 0.486 | 0.576 | 0.485 | 0.646 | 0.148 | 0.782 | 0.04 | 0.871 | 0.491 | 0.724 |
7 | 0.468 | 0.595 | 0.631 | 0.289 | 0.631 | 0.877 | 0.581 | 0.059 | 0.185 | 0.721 | 0.513 | 0.873 | 0.035 | 0.347 |
8 | 0.882 | 0.648 | 0.877 | 0.774 | 0.722 | 0.734 | 0.124 | 0.863 | 0.260 | 0.459 | 0.865 | 0.646 | 0.175 | 0.173 |
9 | 0.795 | 0.467 | 0.528 | 0.189 | 0.775 | 0.433 | 0.295 | 0.246 | 0.778 | 0.442 | 0.031 | 0.076 | 0.615 | 0.466 |
10 | 0.208 | 0.173 | 0.741 | 0.615 | 0.927 | 0.338 | 0.136 | 0.762 | 0.717 | 0.629 | 0.439 | 0.899 | 0.32 | 0.292 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
3 | 3.70 | 4.00 | 0.80 | 3.90 |
4 | 2.60 | 2.70 | 0.50 | 2.50 |
5 | 2.30 | 1.80 | 0.20 | 2.20 |
6 | 2.00 | 1.80 | 0.50 | 1.90 |
7 | 2.40 | 1.80 | 1.20 | 1.80 |
8 | 1.60 | 1.10 | 0.80 | 1.20 |
9 | 0.90 | 0.50 | 0.50 | 0.20 |
10 | 0.70 | 0.00 | 0.20 | 0.30 |
11 | 0.70 | 0.70 | 1.20 | 0.20 |
12 | 0.30 | -0.20 | 0.80 | -0.20 |
13 | 0.60 | 0.60 | 0.90 | 0.20 |
14 | 0.00 | 0.00 | 0.40 | 0.10 |
15 | 0.60 | 0.30 | 0.70 | 0.10 |
16 | 0.30 | 0.30 | 0.50 | 0.30 |
17 | 0.00 | 0.00 | 0.30 | 0.30 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
5 | 0.32 | 0.03 | 0.05 | 0.01 |
10 | 0.37 | 0.07 | 0.16 | 0.06 |
15 | 0.47 | 0.15 | 0.23 | 0.07 |
20 | 0.87 | 0.18 | 0.25 | 0.22 |
25 | 1.45 | 0.29 | 0.40 | 0.37 |
30 | 1.31 | 0.51 | 0.65 | 0.43 |
35 | 2.61 | 0.59 | 0.16 | 0.55 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
3 | 0.06 | 0.87 | 0.69 | 0.66 |
4 | 0.00* | 0.00* | 0.01* | 0.00* |
5 | 0.91 | 0.73 | 0.31 | 0.79 |
6 | 0.00* | 0.00* | 0.01* | 0.00* |
7 | 0.00* | 0.00* | 0.00* | 0.00* |
8 | 0.00* | 0.00* | 0.00* | 0.00* |
9 | 0.01* | 0.10 | 0.06 | 0.56 |
10 | 0.02* | 1.00 | 0.44 | 0.32 |
11 | 0.00* | 0.00* | 0.00* | 0.23 |
12 | 0.16 | 0.47 | 0.00* | 0.62 |
13 | 0.03* | 0.02* | 0.00* | 0.42 |
14 | 0.66 | 1.00 | 0.10 | 0.42 |
15 | 0.03* | 0.08 | 0.01* | 0.32 |
16 | 0.16 | 0.16 | 0.08 | 0.16 |
17 | Not applicable | Not applicable | 0.32 | 0.32 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
5 | 0.11 | 0.68 | 0.40 | 0.98 |
10 | 0.00* | 0.07 | 0.00* | 0.04* |
15 | 0.00* | 0.00* | 0.00* | 0.11 |
20 | 0.00* | 0.01* | 0.00* | 0.00* |
25 | 0.00* | 0.00* | 0.00* | 0.00* |
30 | 0.00* | 0.00* | 0.00* | 0.00* |
35 | 0.00* | 0.01* | 0.44 | 0.02* |
NO. | Input attributes | Output | ||||
X1 | … | Xj | … | Xm−1 | Xm | |
1 | x1,1 | … | x1,j | … | x1,m−1 | x1,m |
… | … | … | … | … | … | … |
i | xi,1 | … | xi,j | … | xi,m−1 | xi,m |
… | … | … | … | … | … | … |
n | xn,1 | … | xn,j | … | xn,m−1 | xn,m |
Bladder cell lines | Input attributes | Output | |||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | class | ||
1 | BFTC-905 | 1.8 | 0.1 | 2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | 0.1 | 2 | 0.1 | 1.5 | 1 | ||
2 | HT 1197 | 1 | 1.3 | 1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | 2.8 | 2.5 | 1 | 1.3 | 1 | ||
3 | T24 | 1.5 | 3.5 | 0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | 2.3 | 2 | 2 | 0.1 | 1 | ||
4 | TCC-SUP | 0.1 | 1 | 1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | 2.3 | 2.8 | 0.1 | 0.2 | 1 | ||
5 | J82 | 0.3 | 1.3 | 2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | 0.1 | 2.5 | 1.3 | 0.2 | 1 | ||
6 | TSGH-8301 | 2 | 0.3 | 0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | 0.1 | 2.5 | 1.3 | 1.5 | 1 | ||
7 | HT 1376 | 1 | 2 | 1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | 1.3 | 2.5 | 0.1 | 0.8 | 1 | ||
8 | Sca-BER | 1 | 1 | 0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | 0.5 | 2.8 | 1.8 | 0.3 | 1 | ||
9 | 5637 | 1.2 | 0.8 | 1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | 0.1 | 2.5 | 0.1 | 1.3 | 1 | ||
10 | A | 0.3 | 0.1 | 6 | 0.5 | 0.1 | 0.2 | 0.5 | 0.1 | 0.5 | 5 | 0.5 | 5 | 8 | 0 | ||
11 | B | 0.1 | 0.3 | 4 | 1 | 0.2 | 0.1 | 2 | 0.3 | 0.1 | 4 | 0.5 | 9 | 5 | 0 | ||
12 | C | 0.2 | 0.1 | 5 | 0.5 | 0.1 | 0.1 | 1.5 | 0.2 | 0.1 | 6 | 1 | 7 | 5 | 0 | ||
13 | D | 0.1 | 0.8 | 9 | 1.5 | 0.1 | 0.5 | 0.5 | 0.1 | 1 | 6 | 0.8 | 6 | 6 | 0 | ||
14 | E | 0.5 | 0.1 | 6 | 2 | 0.5 | 0.1 | 0.5 | 0.3 | 0.1 | 5 | 0.5 | 4 | 4 | 0 | ||
15 | F | 0.5 | 0.5 | 6 | 2 | 0.1 | 0.8 | 0.8 | 0.1 | 0.2 | 6 | 1 | 5 | 5 | 0 | ||
16 | G | 0.1 | 1 | 4 | 1 | 0.3 | 0.2 | 0.5 | 0.5 | 0.2 | 5 | 1.5 | 5 | 7 | 0 | ||
17 | H | 0.5 | 0.1 | 5 | 0.5 | 0.2 | 0.1 | 1 | 0.1 | 0.1 | 8 | 0.5 | 7 | 7 | 0 | ||
18 | I | 0.1 | 0.1 | 7 | 0.8 | 0.1 | 0.1 | 2 | 0.2 | 0.5 | 5 | 0.5 | 6 | 4 | 0 |
Bladder cell lines | Input attributes | Output | ||||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 | Resistance to radiotherapy | ||
1 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 5 | 97 | ||
2 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 10 | 90 | ||
3 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 20 | 84 | ||
4 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 30 | 82 | ||
5 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 5 | 92 | ||
6 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 10 | 78 | ||
7 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 20 | 72 | ||
8 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 30 | 73 | ||
9 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 5 | 94 | ||
10 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 10 | 70 | ||
11 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 20 | 50 | ||
12 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 30 | 41 | ||
13 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 5 | 84 | ||
14 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 10 | 70 | ||
15 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 20 | 48 | ||
16 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 30 | 40 | ||
17 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 5 | 82 | ||
18 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 10 | 57 | ||
19 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 20 | 39 | ||
20 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 30 | 36 | ||
21 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 5 | 90 | ||
22 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 10 | 58 | ||
23 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 20 | 39 | ||
24 | T24 | 0.3 | 1.3 | -2.7 | 2.2 | 1 | 0.1 | 2.5 | 0.5 | 2 | -0.1 | 2.5 | 1.3 | -0.2 | 30 | 34 | ||
25 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 5 | 83 | ||
26 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 10 | 50 | ||
27 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 20 | 32 | ||
28 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 30 | 28 | ||
29 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 5 | 60 | ||
30 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 10 | 30 | ||
31 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 20 | 29 | ||
32 | TSGH-8301 | 2 | 0.3 | -0.3 | 9 | 3.5 | 4.5 | 6.5 | 0.1 | 0.1 | -0.1 | 2.5 | 1.3 | -1.5 | 30 | 30 | ||
33 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 5 | 63 | ||
34 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 10 | 33 | ||
35 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 20 | 21 | ||
36 | BFTC-905 | 1.8 | 0.1 | -2.5 | 7 | 0.8 | 1 | 1 | 0.3 | 2.5 | -0.1 | 2 | 0.1 | -1.5 | 30 | 18 |
Bladder cell lines | Input attributes | |||||||||||||||
NO. | ID | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 | |
1 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 5 | |
2 | HT 1376 | 1 | 1 | -0.5 | 5 | 0.5 | 6.5 | 6.5 | 0.7 | 0.1 | -0.5 | 2.8 | 1.8 | -0.3 | 5 | |
3 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 5 | |
4 | TCC-SUP | 0.1 | 1 | -1 | 1.5 | 0.1 | 0.1 | 3 | 0.1 | 0.8 | -2.3 | 2.8 | 0.1 | -0.2 | 10 | |
5 | 5637 | 1 | 1.3 | -1 | 8.5 | 0.3 | 0.1 | 3.5 | 0.3 | 0.8 | -2.8 | 2.5 | 1 | -1.3 | 30 | |
6 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 30 | |
7 | Sca-BER | 1.2 | 0.8 | -1.5 | 10 | 0.5 | 0.1 | 5 | 0.1 | 2.7 | -0.1 | 2.5 | 0.1 | -1.3 | 20 | |
8 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 5 | |
9 | HT 1197 | 1 | 2 | -1.3 | 3 | 0.3 | 3 | 2.5 | 0.1 | 1.3 | -1.3 | 2.5 | 0.1 | -0.8 | 30 | |
10 | J82 | 1.5 | 3.5 | -0.7 | 2 | 1 | 1.5 | 2.5 | 2 | 2.5 | -2.3 | 2 | 2 | -0.1 | 20 |
Input attributes | ||||||||||||||
MF type | Tri | Tri | Trap | Tri | Tri | Tri | Tri | Tri | Trap | Tri | Tri | Tri | Trap | Tri |
estimations | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 |
a1 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
uset,1 | 0.643 | 0.315 | 0.5 | 0.176 | 0.333 | 0.219 | 0.062 | 0.105 | 0.3 | 0.185 | 0.625 | 0.474 | 0.0 | 0.4 |
b1 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
a2 | - | - | 0.0 | - | - | - | - | - | 0.0 | - | - | - | 0.0 | - |
uset,2 | - | - | 0.8 | - | - | - | - | - | 0.9 | - | - | - | 1.0 | - |
b2 | - | - | 1.0 | - | - | - | - | - | 1.0 | - | - | - | 1.0 | - |
Note: "Tri" and "Trap" represents that the attribute is fitted by the triangular and trapezoidal MF, respectively. |
NO. | MDR | Topo II | Rb | EGFR | Neu | c-ErbB-3 | c-ErbB-4 | Cyclin A | Cyclin D1 | P16 | Cdc 2 | Bcl-2 | Bax | Co-60 |
1 | 0.336 | 0.631 | 0.032 | 0.104 | 0.600 | 0.201 | 0.186 | 0.889 | 0.127 | 0.05 | 0.223 | 0.696 | 0.407 | 0.425 |
2 | 0.862 | 0.075 | 0.806 | 0.278 | 0.329 | 0.109 | 0.837 | 0.581 | 0.353 | 0.113 | 0.145 | 0.529 | 0.305 | 0.052 |
3 | 0.966 | 0.141 | 0.231 | 0.381 | 0.534 | 0.338 | 0.169 | 0.377 | 0.57 | 0.171 | 0.774 | 0.171 | 0.444 | 0.165 |
4 | 0.937 | 0.656 | 0.256 | 0.715 | 0.047 | 0.561 | 0.314 | 0.727 | 0.108 | 0.700 | 0.483 | 0.488 | 0.024 | 0.477 |
5 | 0.515 | 0.514 | 0.818 | 0.754 | 0.882 | 0.912 | 0.647 | 0.178 | 0.037 | 0.39 | 0.904 | 0.554 | 0.224 | 0.012 |
6 | 0.129 | 0.147 | 0.139 | 0.685 | 0.486 | 0.576 | 0.485 | 0.646 | 0.148 | 0.782 | 0.04 | 0.871 | 0.491 | 0.724 |
7 | 0.468 | 0.595 | 0.631 | 0.289 | 0.631 | 0.877 | 0.581 | 0.059 | 0.185 | 0.721 | 0.513 | 0.873 | 0.035 | 0.347 |
8 | 0.882 | 0.648 | 0.877 | 0.774 | 0.722 | 0.734 | 0.124 | 0.863 | 0.260 | 0.459 | 0.865 | 0.646 | 0.175 | 0.173 |
9 | 0.795 | 0.467 | 0.528 | 0.189 | 0.775 | 0.433 | 0.295 | 0.246 | 0.778 | 0.442 | 0.031 | 0.076 | 0.615 | 0.466 |
10 | 0.208 | 0.173 | 0.741 | 0.615 | 0.927 | 0.338 | 0.136 | 0.762 | 0.717 | 0.629 | 0.439 | 0.899 | 0.32 | 0.292 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
3 | 3.70 | 4.00 | 0.80 | 3.90 |
4 | 2.60 | 2.70 | 0.50 | 2.50 |
5 | 2.30 | 1.80 | 0.20 | 2.20 |
6 | 2.00 | 1.80 | 0.50 | 1.90 |
7 | 2.40 | 1.80 | 1.20 | 1.80 |
8 | 1.60 | 1.10 | 0.80 | 1.20 |
9 | 0.90 | 0.50 | 0.50 | 0.20 |
10 | 0.70 | 0.00 | 0.20 | 0.30 |
11 | 0.70 | 0.70 | 1.20 | 0.20 |
12 | 0.30 | -0.20 | 0.80 | -0.20 |
13 | 0.60 | 0.60 | 0.90 | 0.20 |
14 | 0.00 | 0.00 | 0.40 | 0.10 |
15 | 0.60 | 0.30 | 0.70 | 0.10 |
16 | 0.30 | 0.30 | 0.50 | 0.30 |
17 | 0.00 | 0.00 | 0.30 | 0.30 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
5 | 0.32 | 0.03 | 0.05 | 0.01 |
10 | 0.37 | 0.07 | 0.16 | 0.06 |
15 | 0.47 | 0.15 | 0.23 | 0.07 |
20 | 0.87 | 0.18 | 0.25 | 0.22 |
25 | 1.45 | 0.29 | 0.40 | 0.37 |
30 | 1.31 | 0.51 | 0.65 | 0.43 |
35 | 2.61 | 0.59 | 0.16 | 0.55 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
3 | 0.06 | 0.87 | 0.69 | 0.66 |
4 | 0.00* | 0.00* | 0.01* | 0.00* |
5 | 0.91 | 0.73 | 0.31 | 0.79 |
6 | 0.00* | 0.00* | 0.01* | 0.00* |
7 | 0.00* | 0.00* | 0.00* | 0.00* |
8 | 0.00* | 0.00* | 0.00* | 0.00* |
9 | 0.01* | 0.10 | 0.06 | 0.56 |
10 | 0.02* | 1.00 | 0.44 | 0.32 |
11 | 0.00* | 0.00* | 0.00* | 0.23 |
12 | 0.16 | 0.47 | 0.00* | 0.62 |
13 | 0.03* | 0.02* | 0.00* | 0.42 |
14 | 0.66 | 1.00 | 0.10 | 0.42 |
15 | 0.03* | 0.08 | 0.01* | 0.32 |
16 | 0.16 | 0.16 | 0.08 | 0.16 |
17 | Not applicable | Not applicable | 0.32 | 0.32 |
Training data size | RAW vs DB-MTD | MTD vs DB-MTD | GD vs DB-MTD | MD-MTD vs DB-MTD |
5 | 0.11 | 0.68 | 0.40 | 0.98 |
10 | 0.00* | 0.07 | 0.00* | 0.04* |
15 | 0.00* | 0.00* | 0.00* | 0.11 |
20 | 0.00* | 0.01* | 0.00* | 0.00* |
25 | 0.00* | 0.00* | 0.00* | 0.00* |
30 | 0.00* | 0.00* | 0.00* | 0.00* |
35 | 0.00* | 0.01* | 0.44 | 0.02* |