Processing math: 100%
Research article Special Issues

Diagnosis of each main coronary artery stenosis based on whale optimization algorithm and stacking model


  • Cardiovascular disease is currently one of the diseases with high morbidity and mortality worldwide. One of the main types is coronary artery disease (CAD), which occurs when one or more of the three main arteries, the left anterior descending (LAD) artery, the left circumflex (LCX) artery, and the right coronary artery (RCA), are narrowed. In this paper, we introduce a computer-aided diagnosis model, which uses the k-nearest neighbor (KNN)-based whale optimization algorithm (WOA) for feature selection and combines stacking model for CAD diagnosis and prediction. In WOA, the values in the solution vectors are all continuous, and a threshold is set for binary-conversion to obtain the optimal feature subsets of each main coronary artery. Then we develop a two-layer stacking model based on the selected feature subsets to diagnosis LAD, LCX and RCA. By the proposed method, we select 17 features for each main artery diagnosis, and the classification accuracy on LAD, LCX, and RCA test sets is 89.68, 88.71 and 85.81%, respectively. On the Z-Alizadeh Sani dataset, we compare the proposed feature selection method with other metaheuristics and compare the performance of WOA based on different wrappers. The experimental results show that, the KNN-based WOA method selects the optimal feature subsets, and the classification performance of the stacking model is better than other machine learning algorithms.

    Citation: Ziyu Jin, Ning Li. Diagnosis of each main coronary artery stenosis based on whale optimization algorithm and stacking model[J]. Mathematical Biosciences and Engineering, 2022, 19(5): 4568-4591. doi: 10.3934/mbe.2022211

    Related Papers:

    [1] Li Cai, Qian Zhong, Juan Xu, Yuan Huang, Hao Gao . A lumped parameter model for evaluating coronary artery blood supply capacity. Mathematical Biosciences and Engineering, 2024, 21(4): 5838-5862. doi: 10.3934/mbe.2024258
    [2] Xintong Wu, Yingyi Geng, Xinhong Wang, Jucheng Zhang, Ling Xia . Continuous extraction of coronary artery centerline from cardiac CTA images using a regression-based method. Mathematical Biosciences and Engineering, 2023, 20(3): 4988-5003. doi: 10.3934/mbe.2023231
    [3] Bin Zhang, Kuan Zeng, Rongzhen Li, Huiqi Jiang, Minnan Gao, Lu Zhang, Jianfen Li, Ruicong Guan, Yuqiang Liu, Yongjia Qiang, Yanqi Yang . Construction of the gene expression subgroups of patients with coronary artery disease through bioinformatics approach. Mathematical Biosciences and Engineering, 2021, 18(6): 8622-8640. doi: 10.3934/mbe.2021427
    [4] Nattawan Chuchalerm, Wannika Sawangtong, Benchawan Wiwatanapataphee, Thanongchai Siriapisith . Study of Non-Newtonian blood flow - heat transfer characteristics in the human coronary system with an external magnetic field. Mathematical Biosciences and Engineering, 2022, 19(9): 9550-9570. doi: 10.3934/mbe.2022444
    [5] Yongquan Zhou, Yanbiao Niu, Qifang Luo, Ming Jiang . Teaching learning-based whale optimization algorithm for multi-layer perceptron neural network training. Mathematical Biosciences and Engineering, 2020, 17(5): 5987-6025. doi: 10.3934/mbe.2020319
    [6] Natalya Shakhovska, Vitaliy Yakovyna, Valentyna Chopyak . A new hybrid ensemble machine-learning model for severity risk assessment and post-COVID prediction system. Mathematical Biosciences and Engineering, 2022, 19(6): 6102-6123. doi: 10.3934/mbe.2022285
    [7] Juan Du, Jie Hou, Heyang Wang, Zhi Chen . Application of an improved whale optimization algorithm in time-optimal trajectory planning for manipulators. Mathematical Biosciences and Engineering, 2023, 20(9): 16304-16329. doi: 10.3934/mbe.2023728
    [8] B. Wiwatanapataphee, D. Poltem, Yong Hong Wu, Y. Lenbury . Simulation of Pulsatile Flow of Blood in Stenosed Coronary Artery Bypass with Graft. Mathematical Biosciences and Engineering, 2006, 3(2): 371-383. doi: 10.3934/mbe.2006.3.371
    [9] Benchawan Wiwatanapataphee, Yong Hong Wu, Thanongchai Siriapisith, Buraskorn Nuntadilok . Effect of branchings on blood flow in the system of human coronary arteries. Mathematical Biosciences and Engineering, 2012, 9(1): 199-214. doi: 10.3934/mbe.2012.9.199
    [10] Xiaoye Zhao, Yinlan Gong, Lihua Xu, Ling Xia, Jucheng Zhang, Dingchang Zheng, Zongbi Yao, Xinjie Zhang, Haicheng Wei, Jun Jiang, Haipeng Liu, Jiandong Mao . Entropy-based reliable non-invasive detection of coronary microvascular dysfunction using machine learning algorithm. Mathematical Biosciences and Engineering, 2023, 20(7): 13061-13085. doi: 10.3934/mbe.2023582
  • Cardiovascular disease is currently one of the diseases with high morbidity and mortality worldwide. One of the main types is coronary artery disease (CAD), which occurs when one or more of the three main arteries, the left anterior descending (LAD) artery, the left circumflex (LCX) artery, and the right coronary artery (RCA), are narrowed. In this paper, we introduce a computer-aided diagnosis model, which uses the k-nearest neighbor (KNN)-based whale optimization algorithm (WOA) for feature selection and combines stacking model for CAD diagnosis and prediction. In WOA, the values in the solution vectors are all continuous, and a threshold is set for binary-conversion to obtain the optimal feature subsets of each main coronary artery. Then we develop a two-layer stacking model based on the selected feature subsets to diagnosis LAD, LCX and RCA. By the proposed method, we select 17 features for each main artery diagnosis, and the classification accuracy on LAD, LCX, and RCA test sets is 89.68, 88.71 and 85.81%, respectively. On the Z-Alizadeh Sani dataset, we compare the proposed feature selection method with other metaheuristics and compare the performance of WOA based on different wrappers. The experimental results show that, the KNN-based WOA method selects the optimal feature subsets, and the classification performance of the stacking model is better than other machine learning algorithms.



    Cardiovascular disease has always been regarded as the most serious and fatal disease for humans in the world. The increase in morbidity and mortality of cardiovascular disease has brought enormous risks and burdens to healthcare systems around the world [1]. Despite the efforts of medical staff to prevent, diagnose, and treat different types of cardiovascular diseases, the number of deaths from cardiovascular disease in the world continues to increase every year, and by 2019 the number has increased to 18.6 million [2]. World Health Organization estimates that the deaths from cardiovascular disease in 2020 accounts for approximately 32% of the total deaths worldwide [3]. According to NHANES report, the prevalence of cardiovascular disease among adults over 20 years old from 2013 to 2016 was 48%, and the prevalence increased with age [4].

    CAD is one of the common clinical cardiovascular diseases. Clinically, Coronary Angiography (CAG) is mainly used to determine the location and extent of arterial stenosis. The CAG technique is to obtain coronary artery images by X-ray after direct injection of contrast medium into the femoral artery. The inspection is time-consuming, expensive, traumatic, and it has high technical threshold and equipment requirements [5]. Coronary Computed Tomography Angiography (CTA) is an emerging examination technique. The CTA technology obtains accurate and clear images of cardiac coronary artery by intravenous injection of contrast medium and computer reconstruction spiral CT scanning, which is non-invasive. However, the patient's respiration, heart rate, cardiac function, and other factors could affect the imaging quality, which causes the inspection effect of this method to be inferior to CAG [6]. To avoid the harm of CAG technology to patients and the limitations of CTA technology on patients' factors, researchers have widely applied machine learning and data mining techniques to diagnose CAD [7].

    This research will build models for diagnosing LAD, LCX, and RCA based on swarm intelligence optimization algorithm and machine learning techniques. All the research in this paper is done on the extension of Z-Alizadeh Sani dataset, which is derived from the Mendeley Data [8]. We could know the severity of the patient's condition by diagnosing the stenosis of individual artery, which can assist physicians to take corresponding treatments according to different degrees of disease. Feature selection optimization based on meta-heuristic optimizer is a new feature selection algorithm proposed in recent years, which can effectively solve global optimization problems and avoid falling into local optimal solutions [9]. For feature selection, we first apply the filtering method to delete the features with the variance of 0. Then we use the KNN-based WOA to select the feature subsets and use the classification accuracy of KNN and the number of features to guarantee the quality of the selected feature subsets [10]. In this study, a two-layer stacking model is established to blend the results of individual and ensemble classifiers. Four classifiers with best performance are selected from KNN, Support Vector Machine (SVM), Decision Tree (DT), Random Forest (RF), Gradient Boosting Decision Tree (GBDT), Extreme Gradient Boosting (XGBoost), and Adaptive Boosting (AdaBoost) algorithms as primary learners, and Logistic Regression (LR) is applied as the secondary learner to reduce the complexity of the model [11]. We adopt accuracy, recall, precision, F1-score, and AUC value as the evaluation metrics of the model.

    The rest of the paper consists of the following contents: Section 2 provides an overview of relevant studies in previous literature; Section 3 introduces the feature selection methods and machine learning algorithms. Section 4 discusses the results of feature selection and the classification performance of the proposed method. Section 5 summarizes this paper and looks forward to future study.

    In this section, we briefly discuss some relevant studies that use machine learning algorithms, data mining techniques, and improved methods to diagnose and predict diseases.

    The methods and results of feature selection determine the classification performance of the model to some extent. In [12,13,17,18,19,20], Roohallah Alizadehsani and Zeinab Arabasadi et al. adopted the weight by SVM method to complete feature selection. This method uses the normal vector coefficients of linear support vector machine as feature weights. In [14,15], Roohallah Alizadehsani et al. used information gain to measure the importance of features, and selected features with information gain higher than a certain value as feature subset. In [16], Roohallah Alizadehsani et al. measured feature importance by calculating the Gini index of each feature. In [21], Roohallah Alizadehsani et al. proposed the assurance feature selection method. This method measured the importance of a feature by calculating the ratio of the number of patients associated with a feature to the total number of patients. In [22], Roohallah Alizadehsani et al. applied Gini index and principal component analysis (PCA) to calculate the weights of features, and determined the threshold of the weights for feature selection through experiments. In the above studies, researchers implemented feature selection by calculating and evaluating the importance of single features. These methods are computationally fast and easy to implement, focusing on the ability to select features that have a great impact on disease classification.

    Metaheuristic optimization algorithm (MOA) mainly simulates natural and human intelligence to solve the optimal solution [23]. MOA can be divided into four main categories: evolutionary, swarm intelligence, human, and physical and chemical based algorithms. Among them, Genetic Algorithm (GA) [24], Particle Swarm Optimization (PSO) [25], Sine Cosine Algorithm (SCA) [26], Moth-flame Optimization (MFO) [27], WOA [28], Grey Wolf Optimizer (GWO) [29] and other algorithms are widely used for feature selection.

    Moloud Abdar et al. [30] used three types SVMs to establish models for CAD diagnosis, and compared the performance of GA and PSO for feature selection and model parameter optimization in parallel. This method could simultaneously select the optimal feature subset and parameter combination of the model. Bayu Adhi Tama et al. [5] combined the Correlation-based Feature Selection (CFS) and Credal Decision Tree (CDT)-based BPSO to identify important features. The CFS could identify the unimportant and unnecessary features for classification. Shafaq Abbas et al. [31] applied Extremely Randomized Tree (ERT)-based WOA to conduct feature selection and classification on breast cancer dataset. Hoda Zamani et al. [32] proposed a FSWOA algorithm for feature selection, which achieved effective dimensionality reduction in medical datasets. In the above studies, researchers used different machine learning algorithms as wrappers to combine MOA to achieve feature selection for disease diagnosis problems. By preliminarily filtering redundant features in the dataset, the initial population of the MOA can be optimized and the efficiency of feature selection can be improved.

    Since most metaheuristic algorithms are proposed to solve continuous problems, researchers have used transfer functions to convert each dimension of the solution vectors into binary form for feature selection, especially for medical data [33]. E. Emary et al. [34] proposed two binary-conversion methods for the GWO algorithm. The second method is to use the S-shaped transfer function to convert the updated gray wolf position vector into binary form. Shokooh Taghian et al. [35] proposed SBSCA and VBSCA method for feature selection, researchers used S-shaped and V-shaped transfer function to achieve binary-conversion of SCA. Mohammad H. Nadimi-Shahraki et al. [23] improved the MFO algorithm using S-shaped, V-shaped and U-shaped transfer functions for feature selection on medical datasets.

    In previous studies, researchers have applied a variety of basic classification algorithms and ensemble classification algorithms to establish CAD diagnostic models. In [12,13], Roohallah Alizadehsani et al. introduced the cost sensitive algorithm into the model construction, and combined the 10-fold cross-validation (cv) method to evaluate the classification performance of the Sequence Minimum Optimization (SMO) algorithm for CAD. In [15], Roohallah Alizadehsani et al. used ensemble learning method to combine classification results of SMO algorithm and Naïve Bayes (NB) to diagnose CAD. In [16], the researchers used bagging algorithm to obtain high accuracy in the diagnose of LAD. In [18,20,21], Roohallah Alizadehsani et al. used SVM to establish diagnostic models for CAD and main coronary artery stenosis. Zeinab Arabasadi et al. [19] applied Neural Network to establish CAD diagnostic model, and adjusted the weight of Neural Network through GA to obtain ideal classification performance. In [22], the improved SVM algorithm was used to diagnose LAD, LCX and RCA stenosis. The researchers combined the distance between the sample and the separating hyperplane with the accuracy of the classifier to improve the model's performance. Md Mamun Ali et al. [36] applied KNN, DT, and RF to establish disease diagnosis models. Bayu Adhi Tama et al. [5] built a two-layer stacking model to diagnose CAD, RF, GBDT, and XGBoost were used to obtain classification results in the first layer, and the Generalized Linear Model (GLM) in the second layer to generate the final predictions.

    According to the above summary, we can find that there are few studies on the diagnosis of each main coronary artery stenosis. In this study, we will apply machine learning and data mining algorithms to diagnose stenosis of each main coronary artery. We will divide the training set and test set at a ratio of 9: 1, combine randomized search and 10-fold cv to train the model on the training set, and then use the trained model to make predictions on the test set. We apply filtering and KNN-based WOA to select the optimal feature subset for each main coronary artery and then build a two-layer stacking model based on the selected feature subset. At last, we compare the performance achieved by the proposed method with the classification performance obtained in the existing literature.

    The extension of Z-Alizadeh Sani dataset contains the clinical record information of 303 patients, and each patient has 55 features [8]. The patients with LAD, LCX, and RCA stenosis were 160,109, and 101. A total of 216 people were diagnosed with CAD. The features can be divided into four fields: demographic features, symptoms and physical examination, electrocardiography (ECG), laboratory and echocardiography [13]. Table 1 shows the information of features in the extension of Z-Alizadeh Sani dataset.

    Table 1.  Features and types of the extension of Z-Alizadeh Sani dataset.
    Feature type Feature name Count
    Continuous Age, Weight, Height, Body Mass Index (BMI), Blood Pressure (BP), Pulse Rate (PR), Fasting Blood Sugar (FBS), Creatine (Cr), Triglyceride (TG), Low Density Lipoprotein (LDL), High Density Lipoprotein (HDL), Blood Urea Nitrogen (BUN), Erythrocyte Sedimentation Rate (ESR), Hemoglobin (HB), Potassium (K), Sodium (Na), White Blood Cell (WBC), Lymphocyte (Lymph), Neutrophil (Neut), Platelet (PLT), Ejection Fraction (EF-TTE) 21
    Binary Sex, Diabetes Mellitus (DM), Hyper Tension (HTN), Current Smoker, EX-Smoker, Family History (FH), Obesity, Chronic Renal Failure (CRF), Cerebrovascular Accident (CVA), Airway disease, Thyroid Disease, Congestive Heart Failure (CHF), Dyslipidemia (DLP), Edema, Weak Peripheral Pulse, Lung Rales, Systolic Murmur, Diastolic Murmur, Typical Chest Pain (Typical CP), Dyspnea, Atypical, Nonanginal Chest Pain (Nonanginal CP), Exertional Chest Pain (Exertional CP), Low Threshold Angina (LowTH Ang), Q Wave, St Elevation, St Depression, T inversion, Left Ventricular Hypertrophy (LVH), Poor R Progression 30
    Categorical Function Class, Bundle Branch Block (BBB), Region with Regional wall motion abnormality (Region RWMA), Valvular Heart Disease (VHD) 4

     | Show Table
    DownLoad: CSV

    The data mining process starts from the preprocessing stage, followed by feature engineering, and finally uses machine learning algorithms to establish models. The algorithm flow of this research is shown in Figure 1.

    Figure 1.  Algorithm flow chart.

    In one-hot encoding, each category of features is represented by a vector whose length is equal to the number of categories. The i-th vector only takes the value of 1 at the i-th component, and the rest are all 0. One-hot encoding of features can extend the value of discrete features to Euclidean space, making the distance calculation between features more reasonable.

    Standardization is to scale the values of a column of numerical features to a state of mean 0 and variance 1. The standardized formula is shown in Eq (1).

    S(x)=xxσ (1)

    Among them, x is the instance in an n-dimensional space, n is the number of features, and x and σ represent the mean and standard deviation of each feature [37].

    In this research, we use accuracy, precision, recall, F1-score, and AUC value to evaluate the classification performance of proposed model. The calculation formulas are shown in Eqs (2)–(5) [38].

    Accuracy=TP+TNTP+FN+FP+TN (2)
    Precision=TPTP+FP (3)
    Recall=TPTP+FN (4)
    F1score=2×Precision×RecallPrecision+Recall (5)

    AUC (Area under curve) is defined as the area under the ROC curve, which is usually greater than 0.5 and less than 1. The larger the AUC value, the better the classification performance of the classifier.

    WOA is a meta-heuristic global optimization algorithm related to swarm intelligence proposed by Mirjalili and Lewis [28]. The algorithm is inspired by the bubble net foraging behavior of humpback whales and finds the optimal solution by simulating this unique behavior. They hunt for food by continuously shrinking the enclosure, spirally updating their positions, and hunting randomly [39].

    The algorithm mainly includes two stages: first, achieve the encirclement of the prey, and update the spiral position (also known as hunting behavior); second, search for the prey randomly [39]. Next, we will introduce each stage in detail:

    1) Surround the prey. Humpback whales can identify the location of their prey and circle around them. In the initial stage of the algorithm, since we don't know the location of the optimal solution in the search space, the WOA will assume that the best candidate solution currently obtained is the target solution or is close to the optimal solution. After defining the best candidate solution, the whales will attempt to move from other candidate positions to the best position and update their positions. This process is represented by Eq (6):

    P(t+1)=P(t)A|CP(t)P(t)| (6)

    Among them, t is the number of iterations in the current state, P(t) is the position vector of the optimal solution currently, P(t) is the current position vector, and A and C are the constant vectors. The vectors A and C are calculated by Eqs (7) and (8):

    A=2ara (7)
    C=2r (8)

    In the above equations, a is decreased linearly from 2 to 0 in the iterative process, and r is a random vector in the range of [0, 1].

    2) Hunting behavior. Humpback whales hunt by swimming towards their prey in a spiral motion. The mathematical model of hunting behavior is as follows:

    P(t+1)=|P(t)P(t)|eblcos(2πl)+P(t) (9)

    Among them, b is a constant used to define the shape of the logarithmic spiral, and l is a random number in the range of [-1, 1]. During the hunting process, each humpback whale randomly chooses to shrink to surround the prey or spiral upward to chase the prey. The probability of each whale choosing these two behaviors is 50%. The researchers simulated this behavior through the following mathematical model:

    P(t+1)={P(t)A|CP(t)P(t)|,if p<0.5|P(t)P(t)|eblcos(2πl)+P(t),if p>0.5 (10)

    3) Search for prey. The algorithm searches for prey according to the change of the value of A. When |A|>1, the algorithm randomly selects a search individual, and then updates the position of other individuals according to the location of the randomly selected individual, forcing the whale to deviate from the prey, thereby finding a more suitable prey and make the WOA realize global search. When |A|<1, the whales attack the preys. The mathematical model is shown in Eq (11):

    P(t+1)=PrandA|CPrandP(t)| (11)

    Among them, Prand is the position vector of the whale randomly selected from the population.

    In this study, we apply the KNN-based WOA for feature selection. WOA is used to adaptively search for the optimal feature subset to maximize classification accuracy, and KNN is used to ensure the quality of the selected feature subset. In WOA, the whale takes any point in space as a starting point, and continuously adjusts its position to the best candidate solution. Each solution obtained by this algorithm is a continuous vector of the same dimension, bounded in [0, 1] [39]. The function of feature selection can be achieved by setting a threshold to perform binary-conversion on the solution vector. In this study, we set the threshold to 0.5, and the value in the solution is 1 when it is greater than 0.5, and 0 when it is less than 0.5. The length of each solution is M and consists of 0 and 1, where M is the total number of features, 1 means that the feature at the corresponding position is selected, and 0 means that the feature is abandoned. Multiple solution vectors are obtained by changing the initial population size and the number of iterations. The quality of the solution is evaluated by the fitness function. Eq (12) shows the fitness function used in this paper.

    f=αE+(1α)mM (12)

    where f is the fitness of a given solution vector of size M, m is the number of selected features, E is the classification error rate of the classifier, and α is a constant that balances the error rate of the classifier with the number of selected features [10]. The smaller the fitness value, the better the performance of the feature, and the closer to the optimal solution. In this research, E is the classification error rate of KNN, and α is set to 0.99. The pseudocode of WOA feature selection algorithm is shown in Algorithm 1.

    Table Algorithm1.  Pseudocode of WOA feature selection algorithm.
    Input: Number of iterations (T), size of initial population (N).
    Output: The global optimal position vector P*.
    1 Initialize the population Pi (i = 1, 2, 3, …, N).
    2 Binary-conversion.
    3 Compute the fitness of each solution.
    4 Set P* as the best solution.
    5 while (t < T) do
    6  for (each solution) do
    7    Update a, A, C, l and p.
    8    if (p < 0.5) then
    9      if (|A| < 1) then
    10        The whale position is updating by the Eq (6).
    11      else
    12        if (|A| ≥ 1) then
    13          Select the random whale Prand.
    14          The whale position is updating by the Eq (11).
    15        end
    16      end
    17    else
    18       if (p ≥ 0.5) then
    19        Modify the whale position by the Eq (9).
    20      end
    21    end
    22   end
    23  Check if any search agent goes beyond the search space and amend it.
    24  Binary-conversion.
    25  Compute the fitness of each search agent.
    26  Update P* if there is a better solution.
    27  t = t + 1
    28  end
    29 return P*

     | Show Table
    DownLoad: CSV

    In this study, we construct a two-layer stacking model for diagnosing each main coronary artery stenosis. The primary learners in this model are selected from KNN, SVM, DT, RF, GBDT, XGBoost, and AdaBoost, and LR is used as a secondary learner to blend the classification results of multiple primary learners to obtain the final prediction results.

    LR is a generalized linear model used to solve binary classification problems. The output value of the linear model is processed by the sigmoid function and positioned between (0, 1) for the task of binary classification [36]. KNN is a supervised algorithm, and the principle of KNN is that when predicting a new sample, it can judge which category the sample belongs to according to the category of the k points closest to it [40]. SVM is a two-class model with superior performance and flexibility, which can minimize both empirical and structural risks. For a sample set in a finite-dimensional space, SVM performs classification by mapping the sample set from the original feature space to a high-dimensional space [41]. DT is a non-parametric supervised learning method, and the generation of DT is to continuously select the optimal features to divide the training set. ID3, C4.5, and CART are the three main DT algorithms [42]. RF is a special bagging method [43]. For each training set, a decision tree is constructed. When nodes are selected for feature splitting, some features are randomly selected from all the features, and the optimal solution is found from the selected features and applied to node splitting. GBDT uses the addition model and forward stepwise algorithm for greedy learning, and learns a CART tree in each iteration to fit the residuals between the predicted results of the previous (t-1) trees and the real values of the training samples [44]. XGBoost is an optimized distributed gradient boosting library [45]. XGBoost supports multiple types of base classifiers. When using CART as a base classifier, XGBoost improves its generalization ability by adding a regular term to control the complexity of the model. AdaBoost is an iterative algorithm implemented by changing the distribution of dataset. The weight of the sample incorrectly classified by the previous classifier in the training set will increase, and the weight of the sample correctly classified will decrease. Then the new dataset with modified weight is passed to the next classifier for training. At last, the algorithm combines the classifiers obtained each time as the final decision classifier [46].

    Stacking is an ensemble learning algorithm that learns how to best combine the prediction results from multiple well-performing machine learning models. In the stacking model, we call the base learners the primary learners and the learner used for blending is called the secondary learner or meta-learner. Specifically, the original dataset is divided into several subsets, which are input into each primary learner of the first layer and predicted by k-fold cv. Then, the output of each primary learner in the first layer is taken as the input value of the secondary learner in the second layer. The final prediction result is obtained by fitting the trained model to the test set [47]. The algorithm flow of the stacking model is shown in Figure 2.

    Figure 2.  Flow chart of stacking model.

    In this section, we systematically describe the implementation of the proposed method on the extension of Z-Alizadeh Sani dataset and report the results. All experiments in this research are carried out on a Windows machine with 8GB memory and Intel (R) Core (TM) i5-7200U CPU @ 2.50GHZ. Python 3.8 is used in the Jupyter Notebook IDE to implement the entire experiment.

    First, for the processing of categorical features, one-hot encoding is performed on the feature Bundle Branch Block (BBB) to obtain three binary features BBB_LBBB, BBB_RBBB and BBB_N. For feature Valvular Heart Disease (VHD), its values "Normal", "Mild", "Moderate", and "Severe" are denoted as 0, 1, 2 and 3 respectively. Second, for the features Function Class and Region with Regional wall motion abnormality (Region RWMA), they are processed according to the discretization range provided in the Braunwald heart book, when the value of the feature is zero, it is recorded as "Normal", and non-zero is recorded as "High" [48]. Then, all the categorical features with two values are transformed into numerical values. Next, the dataset is divided in a ratio of 9: 1 to obtain training data and test data for LAD, LCX and RCA diagnosis respectively. The training data is used to develop the model, while the test data is used to evaluate the classification performance of the model. At last, the training set is standardized, and the mean and variance of the training set are used for standardizing the test set. Mark "Stenotic" as 1, and "Normal" as 0 in the three labels of LAD, LCX and RCA.

    In this study, we first use the filtering method to delete the feature Exertional CP with the variance of 0 and then use KNN-based WOA to select the feature subsets for diagnosing each main coronary artery. The parameters of feature selection algorithm are shown in Table 2.

    Table 2.  Parameters of WOA feature selection algorithm.
    Parameter Value
    Number of iterations (T) 200,300,400
    Size of initial population (N) 20, 30
    Search domain [0, 1]
    K value 5

     | Show Table
    DownLoad: CSV

    We first compare the results of feature selection by KNN-based WOA, BPSO, GA and BGWO, and the number of iterations and the size of initial population of each algorithm are the same as WOA, as shown in Table 2. Each method is run for 60 times. Table 3 shows the average fitness of the feature subsets obtained by these four algorithms, the average classification accuracy, and the average AUC value on validation set. It can be seen from Table 3 that the KNN-based WOA method has the best performance, and Friedman test results show that there are significant differences in the performance of different feature selection methods. The bold font indicates the best result in the following tables.

    Table 3.  Comparison of KNN-based WOA, BPSO, GA, and BGWO feature selection methods.
    Artery Metrics WOA BPSO GA BGWO
    LAD Fitness 0.0685 0.0999 0.1306 0.0783
    Accuracy 0.9801 0.9024 0.8714 0.9238
    AUC 0.9627 0.8581 0.8251 0.9080
    Friedman rank 1 3 4 2
    LCX Fitness 0.0764 0.0771 0.1341 0.0737
    Accuracy 0.9464 0.9262 0.8679 0.9262
    AUC 0.9382 0.9039 0.8325 0.9077
    Friedman rank 1.33 2.83 4 1.84
    RCA Fitness 0.0761 0.0899 0.1006 0.0519
    Accuracy 0.9559 0.9428 0.9269 0.9452
    AUC 0.9287 0.9089 0.8876 0.9052
    Friedman rank 1.33 2.67 4 2

     | Show Table
    DownLoad: CSV

    Next, we compare WOA feature selection methods based on different wrappers. We choose SVM, DT, and RF to compare with KNN, and run each method 60 times to generate multiple feature subsets. The average performance of these four algorithms on validation set and Friedman test results are given in Table 4. We can conclude from Table 4 that the KNN-based WOA is more suitable for feature selection of the problems in this study. Because the WOA has low complexity, fast convergence, and good optimization performance, and the KNN has low computational complexity and is easy to repeat, the KNN-based WOA method is suitable for the feature selection problem of this study.

    Table 4.  Comparison of WOA feature selection methods based on different wrappers.
    Artery Metrics KNN-WOA SVM-WOA DT-WOA RF-WOA
    LAD Fitness 0.0685 0.1702 0.1171 0.2506
    Accuracy 0.9801 0.8309 0.8857 0.75
    AUC 0.9627 0.7652 0.8823 0.7599
    Friedman rank 1 3 2 4
    LCX Fitness 0.0764 0.1615 0.1401 0.1729
    Accuracy 0.9464 0.8405 0.8619 0.8286
    AUC 0.9382 0.7811 0.8606 0.8570
    Friedman rank 1 3.33 2 3.67
    RCA Fitness 0.0761 0.1217 0.1463 0.1727
    Accuracy 0.9559 0.8809 0.8549 0.8286
    AUC 0.9287 0.8307 0.8269 0.8043
    Friedman rank 1 2 3 4

     | Show Table
    DownLoad: CSV

    In the feature selection process, we obtain multiple feature subsets by setting different iterations and the sizes of the initial population for WOA. We run 20 times for each parameter combination to find the corresponding optimal feature subset. So, for all parameter combinations, we run a total of 120 times. The number of features, classification accuracy and AUC values of the optimal feature subsets on test sets obtained by multiple operations and comparison under different parameter combinations are given in Table 5.

    Table 5.  Classification performance of feature subsets obtained under different parameter combinations.
    Artery Metrics N = 20
    T = 200
    N = 20
    T = 300
    N = 20
    T = 400
    N = 30
    T = 200
    N = 30
    T = 300
    N = 30
    T = 400
    LAD Number 17 8 17 17 19 18
    Accuracy 0.8129 0.7290 0.7742 0.8968 0.6548 0.7742
    Recall 0.9118 0.8235 0.8824 1 0.7235 0.9412
    AUC 0.9005 0.7546 0.7761 0.9067 0.7386 0.8739
    Friedman rank 2.5 4 3.625 1.5 6 13.5
    LCX Number 10 11 17 21 17 12
    Accuracy 0.6451 0.6871 0.8871 0.5903 0.7419 0.7774
    Recall 0.77 0.63 0.82 0.52 0.5 0.6
    AUC 0.7057 0.6381 0.9019 0.5519 0.6333 0.7048
    Friedman rank 2.5 3.25 1.875 5.75 4.625 3
    RCA Number 18 17 22 18 13 10
    Accuracy 0.5613 0.8581 0.6129 0.7452 0.6829 0.7097
    Recall 0.4923 0.7923 0.5 0.5371 0.4219 0.3846
    AUC 0.5184 0.9252 0.6731 0.7603 0.7134 0.7692
    Friedman rank 5.125 1.5 4.75 2.875 3.25 3

     | Show Table
    DownLoad: CSV

    For the diagnosis of three main coronary artery stenosis, the optimal features with better performance are obtained under the parameter combinations of (N = 30, T = 200), (N = 20, T = 400), and (N = 20, T = 300) respectively. Through the KNN-based WOA feature selection method, 17 features are randomly obtained for the diagnosis of LAD, LCX and RCA. The results of feature selection and the Pearson correlation coefficient between features and labels are shown in Table 6.

    Table 6.  Feature selection results and Pearson correlation coefficient of LAD, LCX and RCA.
    LAD (N = 30, T = 200) LCX (N = 20, T = 400) RCA (N = 20, T = 300)
    Feature Coefficient Feature Coefficient Feature Coefficient
    Typical CP 0.47 Typical CP 0.31 Lymph -0.23
    Atypical -0.36 Age 0.29 Age 0.22
    Region RWMA 0.36 HTN 0.18 Atypical -0.22
    Age 0.27 PLT -0.16 Poor R Progression 0.2
    Nonanginal CP -0.22 FBS 0.13 ESR 0.17
    T inversion 0.2 K 0.11 HTN 0.16
    Lymph -0.19 EF-TTE -0.11 FBS 0.15
    Neut 0.18 Nonanginal CP -0.099 WBC 0.14
    PR 0.18 Sex 0.086 Na -0.12
    HTN 0.16 Thyroid Disease -0.086 Q wave 0.11
    BP 0.16 ESR 0.08 K 0.089
    Na -0.15 T inversion 0.043 PR 0.088
    FBS 0.14 Dyspnea -0.038 Dyspnea -0.075
    Poor R Progression 0.11 LDL 0.035 Airway disease 0.052
    BMI -0.087 BMI -0.033 Current Smoker 0.023
    DLP -0.086 Lung rales 0.023 BMI -0.021
    LDL 0.011 LVH 0.019 BBB -0.017

     | Show Table
    DownLoad: CSV

    In this section, we first train multiple basic classifiers, then evaluate and compare the performance of these classifiers. The parameters of each classifier are adjusted on the training set by randomized search combined with 10-fold cv, and the average AUC value is taken as the parameter tuning metric to obtain the optimal hyperparameter combination. We compare the KNN, SVM, DT, RF, GBDT, XGBoost, and AdaBoost algorithms to select four best classifiers as the primary learners to build stacking models. The performance of each algorithm on each main coronary artery test set is shown in Table 7.

    Table 7.  Performance of each algorithm in LAD, LCX, and RCA diagnosis.
    Artery Metrics Algorithm
    KNN SVM DT RF GBDT XGBoost AdaBoost
    LAD Accuracy (%) 80.65 83.87 77.42 83.87 87.1 87.1 87.1
    Recall (%) 94.12 100 88.24 100 94.12 94.12 88.24
    Precision (%) 76.19 77.27 75 77.28 84.21 84.21 88.24
    F1-score (%) 84.21 87.18 81.08 87.18 88.89 88.89 88.24
    AUC 0.7374 0.916 0.8361 0.8782 0.9118 0.9076 0.916
    LCX Accuracy (%) 83.87 83.87 80.65 74.19 87.1 83.87 87.1
    Recall (%) 70 80 100 50 80 90 90
    Precision (%) 77.78 72.73 62.5 62.5 80 69.23 75
    F1-score (%) 73.68 76.19 76.92 55.56 80 78.26 81.82
    AUC 0.9 0.8667 0.8 0.819 0.9 0.8333 0.8952
    RCA Accuracy (%) 83.87 80.65 64.52 67.74 83.87 80.65 83.87
    Recall (%) 84.62 84.62 30.77 30.77 69.23 69.23 76.92
    Precision (%) 78.57 73.33 66.67 80 90 81.82 83.33
    F1-score (%) 81.48 78.57 42.11 44.44 78.26 75 80
    AUC 0.9316 0.8205 0.7607 0.7009 0.8248 0.7991 0.8419

     | Show Table
    DownLoad: CSV

    According to the "strong-strong combination" strategy of the stacking model, we should try to choose classifiers with good performance and maintain the diversity of primary learners. According to the performance indicators of each machine learning algorithm in the above table, we choose SVM, GBDT, XGBoost, and AdaBoost as primary learners for LAD diagnosis, choose KNN, SVM, GBDT, and AdaBoost for LCX diagnosis, choose KNN, SVM, GBDT, and AdaBoost for RCA diagnosis, and use LR as the secondary learner of these three models. For these three stacking models, we compare the performance of the stacking algorithm in the case of 3-fold, 5-fold, and 8-fold cv to select the best cross-validation fold. Moreover, we generate three sets of 10 random numbers between 0 and 1000 to control the random state of the stacking algorithm in each model. Then, we calculate the average value of the model evaluation indicators under 10 different random states to test the stability of the models.

    Table 8 shows the performance of the model in the diagnosis of main coronary arteries when different cross-validation folds are set for the stacking algorithm. From the data in the table, we can conclude that the 5-fold cv of the stacking algorithm can make our model obtain the best classification performance.

    Table 8.  The performance of the stacking model under different cross-validation folds.
    Artery CV Metrics
    Accuracy (%) Recall (%) Precision (%) F1-score (%) AUC
    LAD cv = 3 87.42 ± 2.82 98.24 ± 2.84 82.44 ± 3.59 89.58 ± 2.11 0.9029 ± 0.0169
    cv = 5 89.68 ± 2.04 100 ± 0.00 84.23 ± 2.44 91.42 ± 1.49 0.9067 ± 0.0018
    cv = 8 87.42 ± 3.21 99.41 ± 1.86 81.83 ± 3.93 89.71 ± 2.36 0.9029 ± 0.0169
    LCX cv = 3 87.1 ± 2.15 73 ± 11.6 87.62 ± 11.59 78.23 ± 4.06 0.9038 ± 0.0406
    cv = 5 88.71 ± 1.7 82 ± 4.22 83.03 ± 4.11 82.41 ± 2.59 0.9019 ± 0.0356
    cv = 8 86.45 ± 2.96 76 ± 8.43 81.63 ± 6.95 78.24 ± 5.19 0.9062 ± 0.0127
    RCA cv = 3 82.9 ± 4.57 73.85 ± 7.43 85.33 ± 11.61 78.47 ± 5.03 0.9154 ± 0.0112
    cv = 5 85.81 ± 2.26 79.23 ± 5.2 86.81 ± 7.55 82.44 ± 1.97 0.9252 ± 0.0088
    cv = 8 84.19 ± 3.21 76.92 ± 5.13 84.98 ± 8.25 80.39 ± 3.33 0.9218 ± 0.0962

     | Show Table
    DownLoad: CSV

    In Tables 911, we compare the classification performance of the proposed stacking models with each primary learner on the test set. It can be seen that the stacking models have the highest accuracy and F1-score in all prediction results, and the stacking models can also stabilize the recall, precision, and AUC value in a good range.

    Table 9.  Performance of stacking model and primary learners in LAD diagnosis.
    Metrics Algorithm
    SVM GBDT XGBoost AdaBoost Stacking
    Accuracy (%) 83.87 87.1 87.1 87.1 89.68 ± 2.04
    Recall (%) 100 94.12 94.12 88.24 100 ± 0.00
    Precision (%) 77.27 84.21 84.21 88.24 84.23 ± 2.44
    F1-score (%) 87.18 88.89 88.89 88.24 91.42 ± 1.49
    AUC 0.916 0.9118 0.9076 0.916 0.9067 ± 0.0018
    Friedman rank 3.6 2.5 3.3 2.9 2.1

     | Show Table
    DownLoad: CSV
    Table 10.  Performance of stacking model and primary learners in LCX diagnosis.
    Metrics Algorithm
    KNN SVM GBDT AdaBoost Stacking
    Accuracy (%) 83.87 83.87 87.1 87.1 88.71 ± 1.7
    Recall (%) 70 80 80 90 82 ± 4.22
    Precision (%) 77.78 72.73 80 75 83.03 ± 4.11
    F1-score (%) 73.68 76.19 80 81.82 82.41 ± 2.59
    AUC 0.9 0.8667 0.9 0.8952 0.9019 ± 0.0356
    Friedman rank 4 4.4 2.7 2.7 1.2

     | Show Table
    DownLoad: CSV
    Table 11.  Performance of stacking model and primary learners in RCA diagnosis.
    Metrics Algorithm
    KNN SVM GBDT AdaBoost Stacking
    Accuracy (%) 83.87 80.65 83.87 83.87 85.81 ± 2.26
    Recall (%) 84.62 84.62 69.23 76.92 79.23 ± 5.2
    Precision (%) 78.57 73.33 90 83.33 86.81 ± 7.55
    F1-score (%) 81.48 78.57 78.26 80 82.44 ± 1.97
    AUC 0.9316 0.8205 0.8248 0.8419 0.9252 ± 0.0088
    Friedman rank 2.3 4.1 3.6 3 2

     | Show Table
    DownLoad: CSV

    From Table 9 we can see that compared with other machine learning algorithms, stacking model has the highest accuracy (89.68%), recall (100%), and F1-score (91.42%) in the diagnosis of LAD stenosis, especially the recall of 100%. This shows that the model can identify all patients with artery stenosis in the LAD test set, and minimize the possibility of missed diagnosis in patients with CAD. The high accuracy and F1-score indicate that the model also has good ability to identify patients without CAD. In Table 10, compared with other algorithms, the stacking model achieves the highest accuracy (88.71%), precision (83.03%), F1-score (82.41%), and AUC value (0.9019) in the diagnosis of LCX, indicating that the model can accurately distinguish patients with LCX stenosis from those without stenosis. In Table 11, compared with other machine learning algorithms, stacking model in diagnosis of RCA obtains the highest accuracy (85.81%) and F1-score (82.44%). The AUC value (0.9252) of the model remains at a relatively good level, which also shows that the model can accurately distinguish the patients with RCA stenosis and non-stenosis, and effectively reduce the possibility of missed diagnosis in patients with CAD. From the above analysis, we can know that the classification performance of the stacking model is better than that of individual classifiers and ensemble classifiers based on bagging and boosting. The stacking model can combine the advantages of each primary learner to improve the prediction performance to the highest level.

    We compare the classification performance of the model proposed in this paper with the performance of the model established by using the recursive feature elimination cross-validation (RFE-CV) method based on SVM for feature selection [49]. The results are shown in Table 12. It can be seen that the proposed method has better classification performance.

    Table 12.  Performance of stacking model and primary learners in RCA diagnosis.
    Artery Method Metrics
    Accuracy (%) Recall (%) Precision (%) F1-score (%) AUC
    LAD WOA 89.68 100 84.23 91.42 0.9067
    RFE-CV 77.41 91.18 73.81 81.5 0.7908
    LCX WOA 88.71 82 83.03 82.41 0.9019
    RFE-CV 78.06 51 73.68 59.53 0.7881
    RCA WOA 85.81 79.23 86.81 82.44 0.9252
    RFE-CV 71.61 46.15 76.48 55.53 0.7299

     | Show Table
    DownLoad: CSV

    Figure 3 shows the comparison diagrams of the classification performance between the stacking model and each primary learner in diagnosing LAD, LCX and RCA stenosis.

    Figure 3.  Classification performance of LAD, LCX, and RCA diagnostics using stacking model versus primary learners.

    Figure 4 shows the ROC curves of the stacking model and other machine learning algorithms in diagnosing LAD, LCX and RCA stenosis, respectively.

    Figure 4.  ROC curves of LAD, LCX and RCA diagnostics.

    We compare the number of features and classification accuracy of the proposed model in this study with previous studies and show the differences between the methods and results in Table 13. It can be seen that the feature selection method used in this paper selects fewer features, and the classification accuracy of the proposed model is significantly better than other methods in diagnosing individual LAD, LCX and RCA stenosis.

    Table 13.  Comparison of proposed method with previous studies for detecting main coronary arteries stenosis.
    Study Feature selection method Model Number of features Accuracy (%)
    LAD LCX RCA LAD LCX RCA
    [50] All features C4.5 30 30 30 74.2 63.76 68.33
    [18] Weight by SVM SVM 24 24 24 86.14 83.17 83.5
    [21] Assurance SVM 30 40 40 86.64 83.47 82.85
    [22] Gini index and PCA SVM 23 16 26 86.43 83.67 82.67
    Proposed method KNN-WOA Stacking 17 17 17 89.68 88.71 85.81

     | Show Table
    DownLoad: CSV

    We apply the proposed model to the well-known Cleveland dataset to diagnose heart disease [51]. The dataset contains 303 patient records, each of which has 13 features for diagnosing heart disease [52]. In this study, we select 297 records with no missing values to build the model, of which 137 patients were diagnosed with heart disease. We use the same processing method as the Z-Alizadeh Sani dataset to complete the preprocessing of the Cleveland dataset, and apply the KNN-based WOA method for feature selection. We run 20 times for each parameter combination and obtain the optimal feature subset under parameter combination of (T = 20, N = 300). The feature subset containing 6 features including resting electrocardiographic results (restecg), maximum heart rate achieved (thalach), exercise induced angina (exang), the slope of the peak exercise ST segment (slope), number of major vessels colored by flourosopy (ca), and thallium scan (thal). Then, we build a stacking model based on the selected features and choose KNN, SVM, GBDT, and AdaBoost as primary learners. The proposed method achieves a classification accuracy of 89.67% and an AUC value of 0.9129 on the Cleveland dataset. We compare the classification performance of the stacking model with each primary learner and show the results in Table 14.

    Table 14.  Performance of stacking model and primary learners on the Cleveland dataset.
    Metrics Algorithm
    KNN SVM GBDT AdaBoost Stacking
    Accuracy (%) 86.67 86.67 86.67 86.67 89.67 ± 1.05
    Recall (%) 78.57 71.42 78.57 78.57 77.86 ± 2.26
    Precision (%) 91.67 100 91.67 91.67 100 ± 0.00
    F1-score (%) 84.62 83.33 84.62 84.62 87.53 ± 1.48
    AUC 0.9196 0.9018 0.9063 0.8973 0.9129 ± 0.0212
    Friedman rank 2.7 3.8 3.1 3.5 1.9

     | Show Table
    DownLoad: CSV

    Table 15 shows the comparison of the performance of the proposed model and the models in previous studies on the Cleveland dataset. It can be seen from the comparison results that our proposed method selects fewer features on the Cleveland dataset, and the model can achieve high precision and AUC value in the diagnosis of heart disease.

    Table 15.  Comparison of proposed method with previous studies on the Cleveland dataset.
    Study Feature selection method Model Number of features Accuracy (%) Precision (%) F1-score (%) AUC
    [5] PSO-based Two-tier ensemble 7 85.71 -- 86.49 0.8586
    [53] All features HRFLM 13 88.47 87.5 90 --
    [54] MAFW CNN 8 90.1 88.9 87.8 --
    [55] Chi-Squared SMO 11 86.47 86.5 86.4 0.861
    Proposed method KNN-WOA Stacking 6 89.67 100 87.53 0.9129
    Note:(The "--" in Table 15 indicates that this performance metric was not reported in the study.)

     | Show Table
    DownLoad: CSV

    This paper proposes new models for main coronary arteries diagnosis. We use the KNN-based WOA for feature selection on the extension of Z-Alizadeh Sani dataset and apply stacking models to diagnose LAD, LCX and RCA stenosis.

    In the feature selection process, first, we delete the feature with zero variance in the dataset. Then, by comparing the feature selection results of multiple meta-heuristic optimization algorithms and different wrappers, the KNN-based WOA method is used to select the optimal feature subsets. By using this method, we obtain three optimal feature subsets for diagnosing each main coronary artery, each of which contains 17 features. According to the results of feature selection, it can be seen that the features Age, BMI, FBS, and HTN appear in the feature subsets of each main coronary artery, which indicates that these features are important indicators affecting CAD.

    In the stacking model, we choose different primary learners for the three coronary arteries, and use LR as the secondary learner. In this paper, the average classification performance of the stacking algorithm in multiple random states is calculated to obtain stable results. The diagnostic accuracy of the proposed method in this research for LAD, LCX and RCA stenosis is 89.68, 88.71 and 85.81%, respectively. The classification performance of the stacking model is more stable than that of other machine learning algorithms. Compared with previous studies, we select relatively fewer features in this study, and the diagnostic accuracy of the proposed model is also significantly improved. Our results show that the proposed method can be well applied to CAD datasets and provide a reliable and robust model for clinical diagnosis.

    In the future study, we intend to use improved WOA to select fewer features and make accurate predictions of each main coronary artery stenosis on larger CAD datasets.

    Thanks to our families and colleagues who supported us morally.

    All authors declare no conflicts of interest in this paper.



    [1] E. J. Benjamin, P. Muntner, A. Alonso, M. S. Bittencourt, C. W. Callaway, A. Carson, et al., Heart disease and stroke statistics-2019 update: a report from the American heart association, Circulation, 139 (2019), 56-528. https://doi.org/10.1161/CIR.0000000000000659 doi: 10.1161/CIR.0000000000000659
    [2] G. A. Roth, G. A. Mensah, C. O. Johnson, G. Addolorato, E. Ammirati, L. M. Baddour, et al., Global burden of cardiovascular diseases and risk factors, 1990-2019: Update from the GBD 2019 study, J. Am. Coll. Cardiol., 76 (2020), 2982-3021. https://doi.org/10.1016/j.jacc.2020.11.010 doi: 10.1016/j.jacc.2020.11.010
    [3] Cardiovascular diseases, 2021. Available from: https://www.who.int/health-topics/cardiovascular-diseases.
    [4] S. S. Virani, A. Alonso, E. J. Benjamin, M. S. Bittencourt, C. W. Callaway, A. Carson, et al., Heart disease and stroke statistics-2020 update: a report from the American heart association, Circulation, 141 (2020), 139-596. https://doi.org/10.1161/CIR.0000000000000757 doi: 10.1161/CIR.0000000000000757
    [5] B. A. Tama, S. Im, S. Lee, Improving an intelligent detection system for coronary heart disease using a two-tier classifier ensemble, BioMed. Res. Int., 2020 (2020), 9816142. https://doi.org/10.1155/2020/9816142 doi: 10.1155/2020/9816142
    [6] Y. Yang, Comparison of the diagnostic value of coronary CTA imaging technology and coronary angiography for coronary heart disease, Heilongjiang Med. Pharm., 44 (2021), 113-114. https://doi.org/10.3969/j.issn.1008-0104.2021.02.049 doi: 10.3969/j.issn.1008-0104.2021.02.049
    [7] Y. Khan, U. Qamar, N. Yousaf, Machine learning techniques for heart disease datasets: a survey, in ICMLC '19: Proceedings of the 2019 11th International Conference on Machine Learning and Computing, 2019. https://doi.org/10.1145/3318299.3318343
    [8] R. Alizadehsani, Extention of Z-Alizadeh sani dataset, Mendeley Data, V1 (2017). https://doi.org/10.17632/bgf5czvpg2.1
    [9] Y. Zheng, Y. Li, G. Wang, Y. Chen, Q. Xu, J. Fan, et al., A novel hybrid algorithm for feature selection based on whale optimization algorithm, IEEE Access, 7 (2019), 14908-14923. https://doi.org/10.1109/ACCESS.2018.2879848 doi: 10.1109/ACCESS.2018.2879848
    [10] M. Sharawi, H. M. Zawbaa, E. Emary, H. M. Zawbaa, E. Emary, Feature selection approach based on whale optimization algorithm, in 2017 Ninth International Conference on Advanced Computational Intelligence (ICACI), (2017), 163-168. https://doi.org/10.1109/ICACI.2017.7974502
    [11] J. Wang, C. Liu, L. Li, W. Li, L. Yao, H. Li, et al., A stacking-based model for non-invasive detection of coronary heart disease, IEEE Access, 8 (2020), 37124-37133. https://doi.org/10.1109/ACCESS.2020.2975377 doi: 10.1109/ACCESS.2020.2975377
    [12] R. Alizadehsani, M. J. Hosseini, R. Boghrati, A. Ghandeharioun, F. Khozeimeh, Z. A. Sani, Exerting cost-sensitive and feature creation algorithms for coronary artery disease diagnosis, Int. J. Knowl. Disc. Bioinfo., 3 (2012), 59-79. https://doi.org/10.4018/jkdb.2012010104 doi: 10.4018/jkdb.2012010104
    [13] R. Alizadehsani, M. J. Hosseini, Z. A. Sani, A. Ghandeharioun, R. Boghrati, Diagnosis of coronary artery disease using cost-sensitive algorithms, in 2012 IEEE 12th International Conference on Data Mining Workshops, (2012), 9-16. https://doi.org/10.1109/ICDMW.2012.29
    [14] R. Alizadehsani, J. Habibi, Z. A. Sani, H. Mashayekhi, R. Boghrati, A. Ghandeharioun, et al., Diagnosis of coronary artery disease using data mining based on lab data and echo features, J. Med. Bioeng., 1 (2012), 26-29. https://doi.org/10.12720/jomb.1.1.26-29 doi: 10.12720/jomb.1.1.26-29
    [15] R. Alizadehsani, J. Habibi, M. J. Hosseini, R. Boghrati, A. Ghandeharioun, B. Bahadorian, et al., Diagnosis of coronary artery disease using data mining techniques based on symptoms and ECG features, Eur. J. Sci. Res., 82 (2012), 542-553.
    [16] R. Alizadehsani, J. Habibi, Z. A. Sani, H. Mashayekhi, R. Boghrati, A. Ghandeharioun, et al., Diagnosing coronary artery disease via data mining algorithms by considering lab-oratory and Echocardiography Features, Res. Cardiovasc. Med., 2 (2013), 133-139. https://doi.org/10.5812/cardiovascmed.10888 doi: 10.5812/cardiovascmed.10888
    [17] R. Alizadehsani, J. Habibi, M. J. Hosseini, H. Mashayekhi, R. Boghrati, A. Ghandeharioun, et al., A data mining approach for diagnosis of coronary artery disease, Comput. Methods Programs Biomed., 111 (2013), 52-61, https://doi.org/10.1016/j.cmpb.2013.03.004 doi: 10.1016/j.cmpb.2013.03.004
    [18] R. Alizadehsani, M. H. Zangooei, M. J. Hosseini, J. Habibi, A. Khosravi, M. Roshanzamir, et al., Coronary artery disease detection using computational intelligence methods, Knowl. Based Syst., 109 (2016), 187-197. https://doi.org/10.1016/j.knosys.2016.07.004 doi: 10.1016/j.knosys.2016.07.004
    [19] Z. Arabasadi, R. Alizadehsani, M. Roshanzamir, H. Moosaei, A. A. Yarifard, Computer aided decision making for heart disease detection using hybrid neural network-Genetic algorithm, Comput. Methods Programs Biomed., 141 (2017), 19-26. https://doi.org/10.1016/j.cmpb.2017.01.004 doi: 10.1016/j.cmpb.2017.01.004
    [20] R. Alizadehsani, M. J. Hosseini, A. Khosravi, F. Khozeimeh, M. Roshanzamir, N. Sarrafzadegan, et al., Non-invasive detection of coronary artery disease in high-risk patients based on the stenosis prediction of separate coronary arteries, Comput. Methods Programs Biomed., 162 (2018), 119-127. https://doi.org/10.1016/j.cmpb.2018.05.009 doi: 10.1016/j.cmpb.2018.05.009
    [21] R. Alizadehsani, M. Roshanzamir, M. Abdar, A. Beykikhoshk, A. Khosravi, S. Nahavandi, et al., Hybrid genetic-discretized algorithm to handle data uncertainty in diagnosing stenosis of coronary arteries, Expert Syst., (2020), 1-17. https://doi.org/10.1111/exsy.12573
    [22] R. Alizadehsani, M. Roshanzamir, M. Abdar, A. Beykikhoshk, M. H. Zangooei, A. Khosravi, et al., Model uncertainty quantification for diagnosis of each main coronary artery stenosis, Soft Comput., 24 (2020), 10149-10160. https://doi.org/10.1007/s00500-019-04531-0 doi: 10.1007/s00500-019-04531-0
    [23] M. H. Nadimi-Shahraki, M. Banaie-Dezfouli, H. Zamani, S. Taghian, S. Mirjalili, B-MFO: a binary moth-flame optimization for feature selection from medical datasets, Computer, 10 (2021), 136. https://doi.org/10.3390/computers10110136 doi: 10.3390/computers10110136
    [24] D. E. Goldberg, J. H. Holland, Genetic algorithms and machine learning, Mach. Learn., 3 (1988), 95-99. https://doi.org/10.1023/A:1022602019183 doi: 10.1023/A:1022602019183
    [25] R. Eberhart, J. Kennedy, A new optimizer using particle swarm theory, in MHS'95. Proceedings of the Sixth International Symposium on Micro Machine and Human Science, (1995), 39-43. https://doi.org/10.1109/MHS.1995.494215
    [26] S. Mirjalili, SCA: a sine cosine algorithm for solving optimization problems, Knowl. Based Syst., 96 (2016), 120-133. https://doi.org/10.1016/j.knosys.2015.12.022 doi: 10.1016/j.knosys.2015.12.022
    [27] S. Mirjalili, Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm, Knowl. Based Syst., 89 (2015), 228-249. https://doi.org/10.1016/j.knosys.2015.07.006 doi: 10.1016/j.knosys.2015.07.006
    [28] S. Mirjalili, A. Lewis, The whale optimization algorithm, Adv. Eng. Softw., 95 (2016), 51-67. https://doi.org/10.1016/j.advengsoft.2016.01.008 doi: 10.1016/j.advengsoft.2016.01.008
    [29] S. Mirjalili, S. M. Mirjalili, A. Lewis, Grey wolf optimizer, Adv. Eng. Softw., 69 (2014), 46-60. https://doi.org/10.1016/j.advengsoft.2013.12.007 doi: 10.1016/j.advengsoft.2013.12.007
    [30] M. Abdar, W. Książek, U R. Acharya, R. Tan, V. Makarenkov, P. Plawiak, A new machine learning technique for an accurate diagnosis of coronary artery disease, Comput. Methods Programs Biomed., 179 (2019), 104992. https://doi.org/10.1016/j.cmpb.2019.104992 doi: 10.1016/j.cmpb.2019.104992
    [31] S. Abbas, Z. Jalil, A. R. Javed, I. Batool, M. Z. Khan, A. Noorwali, et al., BCD-WERT: a novel approach for breast cancer detection using whale optimization based efficient features and extremely randomized tree algorithm, PeerJ. Comput. Sci., 7 (2021), 390. https://doi.org/10.7717/peerj-cs.390 doi: 10.7717/peerj-cs.390
    [32] H. Zamani, M. H. Nadimi-Shahraki, Feature selection based on whale optimization algorithm for diseases diagnosis, Intl. J. Comput. Sci. Info. Sec., 14 (2016), 1243-1247. https://doi.org/10.13140/RG.2.2.29065.88161 doi: 10.13140/RG.2.2.29065.88161
    [33] S. Taghian, M. H. Nadimi-Shahraki, H. Zamani, Comparative analysis of transfer function-based binary metaheuristic algorithms for feature selection, in 2018 International Conference on Artificial Intelligence and Data Processing (IDAP), (2018), 1-6. https://doi.org/10.1109/IDAP.2018.8620828
    [34] E. Emary, H. M. Zawbaa, A. E. Hassanien, Binary grey wolf optimization approaches for feature selection, Neurocomputing, 172 (2016), 371-381. https://doi.org/10.1016/j.neucom.2015.06.083 doi: 10.1016/j.neucom.2015.06.083
    [35] S. Taghian, M. H. Nadimi-Shahraki, Binary sine cosine algorithms for feature selection from medical data, Adv. Comput.: An Intl. J., 10 (2019), 1-10. https://doi.org/10.5121/acij.2019.10501 doi: 10.5121/acij.2019.10501
    [36] M. M. Ali, B. K. Paul, K. Ahmed, F. M. Bui, J. M.W. Quinn, M. A. Moni, Heart disease prediction using supervised machine learning algorithms: Performance analysis and comparison, Comput. Biol. Med., 136 (2021), 104672. https://doi.org/10.1016/j.compbiomed.2021.104672 doi: 10.1016/j.compbiomed.2021.104672
    [37] M. W. Nadeem, H. G. Goh, M. A. Khan, M. Hussain, M. F. Mushtaq, P. Vasaki, Fusion-based machine learning architecture for heart disease prediction, Comput. Mater. Contin., 67 (2021), 2481-2496. https://doi.org/10.32604/cmc.2021.014649 doi: 10.32604/cmc.2021.014649
    [38] A. H. Shahid, M. P. Singh, A novel approach for coronary artery disease diagnosis using hybrid particle swarm optimization based emotional neural network, Biocybern. Biomed. Eng., 40 (2020), 1568-1585. https://doi.org/10.1016/j.bbe.2020.09.005 doi: 10.1016/j.bbe.2020.09.005
    [39] M. Mafarja, S. Mirjalili, Whale optimization approaches for wrapper feature selection, Appl. Soft Comput., 62 (2018), 441-453. https://doi.org/10.1016/j.asoc.2017.11.006 doi: 10.1016/j.asoc.2017.11.006
    [40] O. Terrada, B. Cherradi, A. Raihani, O. Bouattane, Classification and prediction of atherosclerosis diseases using machine learning algorithms, in 2019 5th International Conference on Optimization and Applications (ICOA), (2019), 1-5. https://doi.org/10.1109/ICOA.2019.8727688
    [41] V. K. Chauhan, K. Dahiya, A. Sharma, Problem formulations and solvers in linear SVM: a review, Artif. Intell. Rev., 52 (2019), 803-855. https://doi.org/10.1007/s10462-018-9614-6 doi: 10.1007/s10462-018-9614-6
    [42] M. M. Ghiasi, S. Zendehboudi, A. A. Mohsenipour, Decision tree-based diagnosis of coronary artery disease: CART model, Comput. Methods Programs Biomed., 192 (2020) 105400. https://doi.org/10.1016/j.cmpb.2020.105400 doi: 10.1016/j.cmpb.2020.105400
    [43] L. Breiman, Random Forests, Mach. Learn., 45 (2001), 5-32. https://doi.org/10.1023/A:1010933404324 doi: 10.1023/A:1010933404324
    [44] J. H. Friedman, Greedy function approximation: a gradient boosting machine, Ann. Statist., 29 (2000). https://doi.org/10.1214/aos/1013203451 doi: 10.1214/aos/1013203451
    [45] T. Chen, C. Guestrin, XGBoost: a scalable tree boosting system, in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '16), Assoc. Comput. Mach., New York, NY, USA, (2016), 785-794. https://doi.org/10.1145/2939672.2939785
    [46] K. Li, G. Zhou, J. Zhai, F. Li, M. Shao, Improved PSO_AdaBoost ensemble algorithm for imbalanced data, Sensors, 19 (2019), 1476. https://doi.org/10.3390/s19061476 doi: 10.3390/s19061476
    [47] D. H. Wolpert, Stacked generalization, Neural Networks, 5 (1992), 241-259. https://doi.org/10.1016/S0893-6080(05)80023-1 doi: 10.1016/S0893-6080(05)80023-1
    [48] O. B. Robert, L. M. Douglas, P. Z. Douglas, L. Peter, Braunwald's heart disease: a textbook of cardiovascular medicine, Adolf WK. Infective Endocarditis. Int. Ed., (2012), 1540-1556.
    [49] Z. Zhang, Z. P. Liu, Robust biomarker discovery for hepatocellular carcinoma from high-throughput data by multiple feature selection methods, BMC Med. Genomics, 14 (2021), 112. https://doi.org/10.1186/s12920-021-00957-4 doi: 10.1186/s12920-021-00957-4
    [50] R. Alizadehsani, J. Habibi, B. Bahadorian, H. Mashayekhi, A. Ghandeharioun, R. Boghrati, et al., Diagnosis of coronary arteries stenosis using data mining, J. Med. Signals Sens., 2 (2012), 153-159. https://doi.org/10.4103/2228-7477.112099 doi: 10.4103/2228-7477.112099
    [51] R. Detrano, V. A. Medical Center, Long beach and cleveland clinic foundation, 2022. Available from: https://archive.ics.uci.edu/ml/datasets/Heart+Disease.
    [52] R. Detrano, A. Janosi, W. Steinbrunn, M. Pfisterer, JJ. Schmid, S. Sandhu, et al., International application of a new probability algorithm for the diagnosis of coronary artery disease, Am. J. Cardiol., 64 (1989), 304-310. https://doi.org/10.1016/0002-9149(89)90524-9 doi: 10.1016/0002-9149(89)90524-9
    [53] S. Mohan, C. Thirumalai, G. Srivastava, Effective heart disease prediction using hybrid machine learning techniques, IEEE Access, 7 (2019), 81542-81554. https://doi.org/10.1109/ACCESS.2019.2923707 doi: 10.1109/ACCESS.2019.2923707
    [54] M. Elhoseny, M. A. Mohammed, S. A. Mostafa, K. H. Abdulkareem, M. S. Maashi, B. Garcia-Zapirain, et al., A new multi-agent feature wrapper machine learning approach for heart disease diagnosis, Comput. Mater. Contin., 67 (2021), 51-71. https://doi.org/10.32604/cmc.2021.012632 doi: 10.32604/cmc.2021.012632
    [55] K. V. V. Reddy, I. Elamvazuthi, A. A. Aziz, S. Paramasivam, H. N. Chua, S. Pranavanand, Heart disease risk prediction using machine learning classifiers with attribute evaluators, Appl. Sci., 11 (2021), 8352. https://doi.org/10.3390/app11188352 doi: 10.3390/app11188352
  • This article has been cited by:

    1. Lemin Peng, Caitou He, Ali Asghar Heidari, Qian Zhang, Huiling Chen, Guoxi Liang, Nojood O. Aljehane, Romany F. Mansour, Information sharing search boosted whale optimizer with Nelder-Mead simplex for parameter estimation of photovoltaic models, 2022, 270, 01968904, 116246, 10.1016/j.enconman.2022.116246
    2. Marcelo Becerra-Rozas, José Lemus-Romani, Felipe Cisternas-Caneo, Broderick Crawford, Ricardo Soto, Gino Astorga, Carlos Castro, José García, Continuous Metaheuristics for Binary Optimization Problems: An Updated Systematic Literature Review, 2022, 11, 2227-7390, 129, 10.3390/math11010129
    3. Huaijun Deng, Linna Liu, Jianyin Fang, Boyang Qu, Quanzhen Huang, A novel improved whale optimization algorithm for optimization problems with multi-strategy and hybrid algorithm, 2023, 205, 03784754, 794, 10.1016/j.matcom.2022.10.023
    4. Mohammad Hashemi, Seyedeh Somayeh Salehi Komamardakhi, Mohammad Maftoun, Omid Zare, Javad Hassannataj Joloudari, Mohammad Ali Nematollahi, Roohallah Alizadehsani, Pietro Sala, Juan M Gorriz, 2024, Chapter 11, 978-3-031-61139-1, 108, 10.1007/978-3-031-61140-7_11
    5. W.-T. Tang, C.-Q. Su, J. Lin, Z.-W. Xia, S.-S. Lu, X.-N. Hong, T2-FLAIR mismatch sign and machine learning-based multiparametric MRI radiomics in predicting IDH mutant 1p/19q non-co-deleted diffuse lower-grade gliomas, 2024, 79, 00099260, e750, 10.1016/j.crad.2024.01.021
    6. Daphin Lilda S, Jayaparvathy R, Effective cardiac disease classification using FS-XGB and GWO approach, 2024, 132, 13504533, 104239, 10.1016/j.medengphy.2024.104239
    7. Zaid Abdi Alkareem Alyasseri, Nabeel Salih Ali, Mohammed Azmi Al-Betar, Sharif Naser Makhadmeh, Norziana Jamil, Mohammed A. Awadallah, Malik Braik, Seyedali Mirjalili, 2024, 9780323953658, 9, 10.1016/B978-0-32-395365-8.00008-7
    8. Soumya Bajpai, Kapil Sharma, Brijesh Kumar Chaurasia, A Hybrid Meta-heuristics Algorithm: XGBoost-Based Approach for IDS in IoT, 2024, 5, 2661-8907, 10.1007/s42979-024-02913-2
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2703) PDF downloads(114) Cited by(8)

Figures and Tables

Figures(4)  /  Tables(16)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog