
The emergence of COVID-19 has broken the silence of humanity and people are gradually becoming concerned about pneumonia-related diseases; thus, improving the recognition rate of pneumonia-related diseases is an important task. Neural networks have a remarkable effectiveness in medical diagnoses, though the internal parameters need to be set in accordance to different data sets; therefore, an important challenge is how to further improve the efficiency of neural network models. In this paper, we proposed a learning exponential distribution optimizer based on chaotic evolution, and we optimized Resnet50 for COVID classification, in which the model is abbreviated as IEDO-net. The algorithm introduces a criterion for judging the distance of the signal-to-noise ratio, a chaotic evolution mechanism is designed according to this criterion to effectively improve the search efficiency of the algorithm, and a rotating flight mechanism is introduced to improve the search capability of the algorithm. In the computed tomography (CT) image data of COVID-19, the accuracy, sensitivity, specificity, precision, and F1 score of the optimized Resnet50 were 94.42%, 93.40%, 94.92%, 94.29% and 93.84%, respectively. The proposed network model is compared with other algorithms and models, and ablation experiments and convergence and statistical analyses are performed. The results show that the diagnostic performance of IEDO-net is competitive, which validates the feasibility and effectiveness of the proposed network.
Citation: Chengtian Ouyang, Huichuang Wu, Jiaying Shen, Yangyang Zheng, Rui Li, Yilin Yao, Lin Zhang. IEDO-net: Optimized Resnet50 for the classification of COVID-19[J]. Electronic Research Archive, 2023, 31(12): 7578-7601. doi: 10.3934/era.2023383
[1] | Islam Uddin, Salman A. AlQahtani, Sumaiya Noor, Salman Khan . Deep-m6Am: a deep learning model for identifying N6, 2′-O-Dimethyladenosine (m6Am) sites using hybrid features. AIMS Bioengineering, 2025, 12(1): 145-161. doi: 10.3934/bioeng.2025006 |
[2] | Shital Hajare, Rajendra Rewatkar, K.T.V. Reddy . Design of an iterative method for enhanced early prediction of acute coronary syndrome using XAI analysis. AIMS Bioengineering, 2024, 11(3): 301-322. doi: 10.3934/bioeng.2024016 |
[3] | Wallace Camacho Carlos, Alessandro Copetti, Luciano Bertini, Leonard Barreto Moreira, Otávio de Souza Martins Gomes . Human activity recognition: an approach 2D CNN-LSTM to sequential image representation and processing of inertial sensor data. AIMS Bioengineering, 2024, 11(4): 527-560. doi: 10.3934/bioeng.2024024 |
[4] | Rabia Aziz, C.K. Verma, Namita Srivastava . Dimension reduction methods for microarray data: a review. AIMS Bioengineering, 2017, 4(1): 179-197. doi: 10.3934/bioeng.2017.1.179 |
[5] | Kayode Oshinubi, Augustina Amakor, Olumuyiwa James Peter, Mustapha Rachdi, Jacques Demongeot . Approach to COVID-19 time series data using deep learning and spectral analysis methods. AIMS Bioengineering, 2022, 9(1): 1-21. doi: 10.3934/bioeng.2022001 |
[6] | Marianthi Logotheti, Eleftherios Pilalis, Nikolaos Venizelos, Fragiskos Kolisis, Aristotelis Chatziioannou . Development and validation of a skin fibroblast biomarker profile for schizophrenic patients. AIMS Bioengineering, 2016, 3(4): 552-565. doi: 10.3934/bioeng.2016.4.552 |
[7] | Abdulmajeed Alsufyani . Performance comparison of deep learning models for MRI-based brain tumor detection. AIMS Bioengineering, 2025, 12(1): 1-21. doi: 10.3934/bioeng.2025001 |
[8] | Lamia Fatiha KAZI TANI, Mohammed Yassine KAZI TANI, Benamar KADRI . Gas-Net: A deep neural network for gastric tumor semantic segmentation. AIMS Bioengineering, 2022, 9(3): 266-282. doi: 10.3934/bioeng.2022018 |
[9] | Kuna Dhananjay Rao, Mudunuru Satya Dev Kumar, Paidi Pavani, Darapureddy Akshitha, Kagitha Nagamaleswara Rao, Hafiz Tayyab Rauf, Mohamed Sharaf . Cardiovascular disease prediction using hyperparameters-tuned LSTM considering COVID-19 with experimental validation. AIMS Bioengineering, 2023, 10(3): 265-282. doi: 10.3934/bioeng.2023017 |
[10] | Xiao-Li Gu, Li Yu, Yu Du, Xiu-Peng Yang, Yong-Gang Xu . Identification and validation of aging-related genes and their classification models based on myelodysplastic syndromes. AIMS Bioengineering, 2023, 10(4): 440-465. doi: 10.3934/bioeng.2023026 |
The emergence of COVID-19 has broken the silence of humanity and people are gradually becoming concerned about pneumonia-related diseases; thus, improving the recognition rate of pneumonia-related diseases is an important task. Neural networks have a remarkable effectiveness in medical diagnoses, though the internal parameters need to be set in accordance to different data sets; therefore, an important challenge is how to further improve the efficiency of neural network models. In this paper, we proposed a learning exponential distribution optimizer based on chaotic evolution, and we optimized Resnet50 for COVID classification, in which the model is abbreviated as IEDO-net. The algorithm introduces a criterion for judging the distance of the signal-to-noise ratio, a chaotic evolution mechanism is designed according to this criterion to effectively improve the search efficiency of the algorithm, and a rotating flight mechanism is introduced to improve the search capability of the algorithm. In the computed tomography (CT) image data of COVID-19, the accuracy, sensitivity, specificity, precision, and F1 score of the optimized Resnet50 were 94.42%, 93.40%, 94.92%, 94.29% and 93.84%, respectively. The proposed network model is compared with other algorithms and models, and ablation experiments and convergence and statistical analyses are performed. The results show that the diagnostic performance of IEDO-net is competitive, which validates the feasibility and effectiveness of the proposed network.
The liver is a vital organ in the human body that maintains basic physiological activities necessary for health. The liver is essential for digestion, metabolism, detoxification, and immunity, highlighting its critical role in maintaining bodily equilibrium. Despite being one of the most diverse and intricate organs, the liver is susceptible to infections, traumas, and genetic disorders, jeopardizing human life and longevity [1],[2]. These disorders will adversely affect the liver and lead to severe consequences; thus, preventative actions and cures for general liver issues are essential. However, persistent injury to this organ may disrupt its activities and overall health despite its significant ability for renewal and adaptation. Consequently, understanding the factors that lead to liver function deterioration, inflammation, and jaundice, alongside pursuing timely diagnosis and prevention, are two essential steps toward enhancing the quality of life and mitigating mortality risks associated with liver diseases (LD).
The healthcare industry has seen significant transformations in its delivery systems and organizations as a result of data processing, primarily via the use of machine learning (ML) and artificial intelligence (AI) [3]. These innovations have transformed data collection, storage, and analysis methods, allowing clinical practitioners to integrate cutting-edge decision-making models into traditional healthcare procedures. This integration has significantly enhanced inspection and informed decision-making in the healthcare industry. It is important to note that, because of the progressive advancement of AI and ML applications, early illness prediction is feasible for conditions such as diabetes, hypertension, COPD, and cardiovascular diseases [4]. These technologies use big data to provide significant insights into a specific diagnosis, aiding illness prognosis management and treatment. Although closely linked with technology, we must recognize that preventative efforts promoting healthy behavior have enhanced the positive transformation this innovation brings. Integrating AI and ML into healthcare procedures is already underway, with potential advancements in precision medicine, medication discovery, and diagnostic tools. The following sections illustrate that these technological advancements include enhancements in healthcare delivery, including improved diagnostics, more focused medication, and overall population health outcomes.
Researchers have proposed many machine learning techniques in the literature for accurately and quickly classifying liver patients. In 2016, Babu et al. [5] introduced a K-means clustering methodology for identifying liver illness, using several classification models to assess its effectiveness. Their study evaluated the efficacy of the Naive Bayes (NB), C4.5 decision tree, and k-nearest neighbors (KNN) classifiers. The results indicated that NBC obtained an accuracy of 56%, KNN reached 64%, and C4.5 attained 69%. This approach illustrates the differential efficacy of several classifiers in conjunction with k-means clustering for identifying liver patients. Likewise, Gan et al. [6] investigated several categorization methods for predicting liver illness. They suggested that the AdaC-TANBN approach had an accuracy of 69.03%. The integrated strategy, which merges AdaBoost with a modified tree-augmented Naive Bayes (TANBN) model, outperformed other classification techniques evaluated in their research. The result highlights the efficacy of integrating boosting methods with probabilistic models to improve prediction accuracy in liver disease categorization. Subsequently, Amare et al. [7] proposed an innovative predictor for identifying LD via sophisticated machine-learning techniques. They created a classification model that combines the complement Naive Bayes (CNB) technique with other classifiers to improve predictive accuracy, comparing it to traditional Naive Bayes classifiers and other models. Their results demonstrated that the CNB approach achieved an accuracy of 71.36%, exceeding the performance of many classifiers analyzed in the literature. This method underscores the efficacy of the CNB technique in improving the accuracy of liver disease categorization. Sreejith and colleagues [8] evaluated classification performance using the ILPD, Pima Indian Diabetes (PID), and Thoracic Surgery (TSD) datasets. They used chaotic multi-verse optimization (CMVO) for feature selection and the synthetic minority oversampling technique (SMOTE), achieving an accuracy of 82.62% on the ILPD dataset with enhanced methodologies. P. Kumar et al. improved liver disease classification accuracy with the ILPD dataset. They used 10-fold cross-validation and utilized variable neighbor-weighted fuzzy KNN, neighbor-weighted K-NN (NWKNN), and fuzzy-neighbor-weighted K-NN classifiers. To address dataset imbalance, they used Tomek connections and redundancy-based under-sampling (TR-RUS), attaining accuracies of 72.31% with NWKNN, 76.61% with fuzzy NWKNN, and 87.71% with their developed variable-NWFKNN approach, which surpassed existing classifiers. Kuzhippallil et al. [9] enhanced the categorization of chronic liver disease by several data preparation and feature selection techniques. Their methodologies improved classification efficacy, achieving a maximum accuracy of 88%. Ruhul et al. [10] recently presented a sophisticated feature extraction and classification technique for predicting LD with the ILPD dataset. Their study integrated projection-based statistical methods to enhance feature selection with relevant machine learning approaches, including support vector machine, logistic regression, and random forest. The suggested technique improves the accuracy of liver disease diagnosis to 88.10%, facilitating early identification and treatment alternatives.
This research introduces a precise and robust learning approach to successfully differentiate liver disease. We first use the random oversampling strategy to address overfitting and class imbalance during model building. Subsequently, we employ two feature methodologies: ranking and statistical projection-based strategies. The Pearson correlation coefficient (PCC), gain ratio (GR), and ReliefF algorithm (RFA) are utilized to rank and evaluate each feature's contribution to the target class (liver illness). We use principal component analysis (PCA) and linear discriminant analysis (LDA) for statistical projection-based techniques. A hybrid feature vector is generated by combining these five feature vectors. We use a feature selection method, Shapley additive explanations (SHAP) [11], to improve computational efficiency, identifying the most significant characteristics by analyzing their impact on model predictions. Diverse hypothesis learners, such as logistic regression (LR), k-nearest neighbor (KNN), random forest (RF), support vector machine (SVM), and Naive Bayes (NB), are used to assess prediction rates. A thorough performance evaluation using measures such as Matthew's correlation coefficient, accuracy, sensitivity, specificity, and area under the curve (AUC) illustrates the enhanced efficacy of our proposed DNN model in predicting liver disease. This methodology offers significant insights into predicting early liver disease, improving diagnostic accuracy, and enabling prompt medical treatments.
This section delivers a detailed, illustrated representation of data transfers via the prototype and architecture of a suggested predictor model. Figure 1 illustrates essential design components to be entirely explored in subsequent sections.
The dataset under study, the ILPD dataset [10], comprises 583 observations and encompasses 10 distinct features. A comprehensive summary of the dataset ILPD is presented in Table 1. The dataset under consideration comprises gender distribution revealing 439 male participants, constituting approximately 75.30%, and 144 female participants, accounting for 24.70% of the total cohort. The dataset's primary target class revolves around diagnosing liver disease, distinguishing participants based on whether they have been diagnosed with this condition.
Feature | Description |
Gender | Participant's gender |
Albumin (ALB) | Albumin level |
Direct bilirubin (DB) | Direct bilirubin level |
Age (years) | Age of participants (4–90 years) |
Total bilirubin (TB) | Total bilirubin level |
Aspartate aminotransferase (SGOT) | Aspartate aminotransferase level |
Total protein (TP) | Total protein level |
Alkaline phosphatase (ALP) | Alkaline phosphatase level |
Alanine aminotransferase (SGPT) | Alanine aminotransferase level |
Albumin and globulin ratio (AGR) | Albumin and globulin ratio |
The long-term liver disease risk is addressed as a classification challenge with Liver-Disease (LD) or Non-Liver-Disease (non-LD) classes. The total dataset comprises 416 positive cases of liver disease and 167 controlled cases. Considering the significance of balanced datasets in machine learning analysis, we applied a data balancing technique, i.e., random oversampling, resulting in a dataset of 832 cases, evenly distributed between instances of liver disease and non-disease cases (i.e., LD = 416, non-LD = 416). Finally, Table 2 outlines the balanced dataset statistical properties of the features, highlighting essential measurements, i.e., standard deviation, mean, minimum, and maximum, for each variable. This analysis provides a complete analysis of the dataset's numerical distributions. Table 2 summarizes the statistical properties of features, including min, max, mean, and standard deviation, providing numerical distribution analysis.
Features | Min | Mean | Max | ± Stdv |
TB | 0.4 | 2.65 | 75 | ± 5.32 |
SGOT | 10 | 88.78 | 4929 | ± 245.07 |
Age | 4 | 43.55 | 90 | ± 16.28 |
ALP | 63 | 267.26 | 2110 | ± 212.62 |
SGPT | 10 | 66.78 | 2000 | ± 155.16 |
AGR | 0.3 | 0.98 | 2.8 | ± 0.30 |
DB | 0.1 | 1.16 | 19.7 | ± 2.42 |
ALB | 0.9 | 3.19 | 5.5 | ± 0.76 |
TP | 2.7 | 6.5 | 9.6 | ± 1.02 |
ML algorithms often face the obscenity of dimensionality, which arises when there are many data points but relatively few meaningful features or when the feature space contains irrelevant information. To address this challenge, we first implement three ranking strategies [12],[13], i.e., Pearson correlation coefficient (PCC), gain ratio (GR), and ReliefF algorithm (RFA), and statistical projection-based strategies such as principal component analysis (PCA) and linear discriminant analysis (LDA) to examine the involvement of each feature to the target class (i.e., liver disease). Subsequently, we use Shapley additive explanations (SHAP) to determine how much each feature matters. These techniques help simplify the dataset by focusing only on the most important parts, making it easier to analyze and reducing the computing power needed. These methods are crucial to finding the most valuable features when predicting liver disease, leading to more accurate and understandable diagnostic and prediction models [14],[15].
In this section, three ranking methods, i.e., PCC, GR, and RFA, are nominated to calculate the involvement of a feature (i.e., LD class). Table 3 provides an evaluation of feature contributions to predicting the target class.
Feature | PCC | GR | RFA |
SGPT | 0.2141 | 0.0701 | 0.2613 |
TB | 0.2874 | 0.1373 | 0.2848 |
Gender | 0.0857 | 0.0065 | 0.025 |
SGOT | 0.2017 | 0.1005 | 0.2393 |
AGR | 0.2046 | 0.0822 | 0.2599 |
DB | 0.3205 | 0.1421 | 0.2895 |
ALB | 0.1836 | 0.0408 | 0.2883 |
Age | 0.1596 | 0.0372 | 0.2625 |
ALP | 0.246 | 0.0867 | 0.1936 |
TP | 0.0443 | 0.00 | 0.2161 |
First, using PCC analysis, some significant relationships were detected, including highly significant ones, for example, 0.88, indicating a relationship between TB and DB. The percentage for this association was comparatively low, at 0.32. DB remained the most frequent feature with 32 DDR, followed by task-interdependent, task-supportive, and task-contrast with 14 DDR, 13 DDR, 11 DDR, and 5 DDR, respectively. Second, the GR method used in the “screening variable selection” highlighted DB, TB, and SGOT as prominent ones. Third, using the Gini impurity ReliefF algorithm suggested that gender had the least priority, whereas DB, ALP, and TB were the least essential features. These methods all emphasize the importance of all features in screening for liver diseases.
PCA reduces data dimensions using covariance matrices [16] with minimum loss of discriminative features. Let us consider a feature vector S with the dimension of i*j, where i represents the number of extracted features and j represents the number of samples. Let k represent the number of desired features. The value of k must be less than i. Let us consider the following input feature vector S using Eq. (1):
Through the PCA algorithm, we implemented the following steps to reduce the dimensionality of the feature vector:
Step 1: Compute the mean
Step 2: Subtract the mean
Step 3: Compute the covariance
Where
Step 4: Calculate the eigenvalues in ascending order. The first eigenvalue should be greater than the second and fourth.
Step 5: Compute the corresponding eigenvectors:
Last step: Select k eigenvectors corresponding to the largest eigenvalues to reconstruct a new set of feature vectors in the space. This selection process helps avoid dimensionality problems, keeping the most essential data. It also improves the computational speed and interpretability of various other tasks being performed in the analytical methods of the dataset, which is the main aim of PCA in the case of data preprocessing and data analysis.
Linear discriminant analysis (LDA) is a supervised dimensionality reduction technique that projects data onto a lower-dimensional space while maximizing class separability. The main goal of LDA is to find a transformation matrix W that maximizes the ratio of between-class variance to within-class variance, ensuring that different classes are well-separated in the projected space. Mathematically, the objective function of LDA is to maximize the following criterion.
Here, Sb is the between-class scatter matrix, Sw is the within-class scatter matrix, and W is the transformation matrix to be optimized.
The methodology proposed in this paper consists of five feature extraction mechanisms described in section 2.2. To develop a hybrid feature vector, we combine the five feature vectors as given in Eq. (9). Mathematically, the hybrid feature vector is also called the hybrid feature vector. Table 4 shows the overview of features contributed by each technique and the total number of features created by the hybrid vector.
Method | Number of features (dimensionality) |
Principal component analysis (PCA) | 10 |
Linear discriminant analysis (LDA) | 10 |
Gain ratio (GR) | 10 |
Pearson correlation coefficient (PCC) | 10 |
ReliefF algorithm (RFA) | 10 |
Hybrid features | 50 |
Feature selection is vital in the machine learning pipeline to identify the most relevant features to improve model performance, reduce overfitting, and lower computational costs [17]. This paper employs Shapley additive explanations (SHAP) for feature selection, which leverages cooperative game theory to assign importance to input features based on their contribution to predictions. SHAP quantifies each feature's impact, providing insights into its influence on model outcomes, as detailed in Eq. (10). Understanding the structure and size of the dataset is equally essential, as it informs preprocessing steps like data splitting, regularization, and feature selection. Properly shaping data ensures efficient processing and supports informed decision-making throughout the data science process.
Where φi, represents the SHAP value for the feature i, N is the set of all features, S is a subset of features excluding i, f(S) is the model's prediction given features in S, and
In this paper, we selected the vital features shown in Figure 2. Each row represents a feature, with dots indicating their corresponding SHAP values. Blue dots signify low values, while red dots denote high values. Positive SHAP values indicate a higher likelihood of liver disease, whereas negative values suggest a lower risk. Testing different feature groups revealed that the top 25 features significantly improved the proposed method's performance.
Deep neural networks (DNNs) mimic the human brain's functioning, consisting of an input layer for receiving data, an output layer for producing results, and multiple hidden layers that process information through interconnected tiers [18]. These layers transform input data, enabling the model to identify intricate patterns and representations. The deep structure of DNNs makes them highly effective in tasks such as image recognition and natural language processing, as each layer extracts increasingly complex features. However, while additional hidden layers improve the model's ability to learn intricate patterns, they also raise computational costs, increase the risk of overfitting, and complicate the model. DNNs excel in feature extraction from unstructured or unlabeled data and have proven valuable in biotechnology, natural language processing, image analysis, audio processing, and engineering [19],[20].
This study focuses on the early prediction of liver illness using benchmark data and a deep neural network model with four hidden layers, as illustrated in Figure 3. The DNN architecture comprises multiple neurons that process the input feature vector to produce output, as detailed in Eqs (11) and (12). The neurons' weights in each layer are initialized using the Xavier function to ensure equal variance across layers [14],[15]. A backpropagation method adjusts the weight matrices to minimize the discrepancies between predicted and target classes. The ReLU activation function is applied in the input and hidden layers to capture nonlinearity and patterns in the data, while the logistic sigmoid activation function is used in the output layer. This mechanism determines whether a neuron is activated or remains inactive based on its output. Additionally, the Softmax activation function in the output layer converts the generated values into probabilities between 0 and 1, representing the likelihood of the data belonging to a specific class.
To thoroughly evaluate the performance of the proposed models, five essential evaluation metrics were employed, i.e., sensitivity (SN), specificity (SP), overall accuracy (ACC), Matthew's correlation coefficient (MCC), and the area under the curve (AUC) [21]. These critical indicators provide insights into the predictor's effectiveness from multiple perspectives. The ACC represents the proportion of correct predictions relative to the total number of samples. The SN measures the proportion of actual positives correctly identified, while SP reflects the proportion of actual negatives accurately detected. The AUC assesses the model's classification ability across multiple classes, with the receiver operating characteristic (ROC) curve plotting the true positive rate against the false positive rate. The MCC evaluates the classifier's performance by accounting for all four confusion matrix elements. The corresponding calculation formulas are provided below.
Here, TP and FN represent the true positive and negative samples (i.e., liver and non-liver diseases). Similarly, TN and FP are errors as the model inaccurately predicted the samples to be true or false.
This section provides a detailed evaluation and discussion of the proposed model's effectiveness. The experiments were conducted using machine learning tests implemented with Python and TensorFlow, both open-source tools that facilitated efficient setup and testing of the prediction models. The tests were executed on a system equipped with an 11th Gen Intel Core i7-1165G7 processor (2.80 GHz), 16 GB of RAM, running Windows 11 Home on a 64-bit architecture.
This section aims to identify the optimal configuration for hyperparameters in DNN topology. To analyze the performance of the DNN on different hyperparameters, we used a grid search technique that applies different combinations of parameters. We considered only those hyperparameters that primarily affect the DNN model's performance. These parameters include activation function, learning rate, and the number of iterations [15],[18],[20],[22].
Parameters list | Optimize values |
Hidden layers | 4 |
Updater | ADAGRAD function |
Hidden layers neurons | 23-17-14-6 |
Seed | 12345L |
Epoch | 10 |
Momentum | 0.9 |
LR | 0.01 |
Dropout | 0.25 |
Regularization l2 | 0.001 |
Activation functions | ReLU & Softmax |
Optimizer | SGD Method |
Weight initialization function | XAVIER function |
First, experiments were conducted to evaluate the impact of activation functions and learning rates on the model's performance. As shown in Table 6, the DNN classifier achieved its highest accuracy of 90.12% while using the ReLU activation function and a learning rate of 0.01. Furthermore, a gradual reduction in the learning rate improved the model's performance; however, further decreases beyond this point did not result in substantial accuracy gains. Thus, a learning rate of 0.01 and the ReLU activation function were identified as optimal parameters for achieving high accuracy in the DNN model.
LR | ReLU | Tanh | Sigmoid |
0.01 | 90.12 | 88.10 | 89.44 |
0.02 | 89.91 | 88.09 | 89.23 |
0.03 | 89.45 | 87.71 | 89.01 |
0.04 | 89.16 | 87.58 | 88.80 |
0.05 | 88.49 | 87.38 | 88.17 |
0.06 | 87.82 | 87.19 | 87.54 |
0.07 | 87.55 | 86.99 | 86.91 |
0.08 | 87.09 | 86.80 | 86.28 |
0.09 | 87.05 | 86.60 | 85.65 |
0.10 | 87.03 | 86.41 | 85.02 |
Next, we conducted numerous experiments to evaluate the performance of the DNN model by varying the number of training epochs. Figure 4 illustrates the error loss observed when using the ReLU activation function. The results indicate a consistent decrease in the error rate with increased training epochs. For example, the error loss began at 0.892 during the initial epoch and progressively declined to 0.002 by the 50th epoch. These findings suggest that 50 epochs are optimal, as the error rates stabilize beyond this point. The optimal configuration determined from this analysis is detailed in Table 5.
This section examines the prediction results of the deep neural network (DNN) model using individual and hybrid feature extraction methods. Machine learning techniques such as k-fold cross-validation (CV) have been widely used in the literature to evaluate model performance. A 10-fold cross-validation approach was adopted in this study, offering an optimal balance between computational efficiency and comprehensive performance evaluation while minimizing the influence of data variability. Table 7 presents the DNN model's predicted outcomes on the balanced dataset using various feature vectors.
Methods | ACC (%) | SP (%) | SN (%) | MCC |
PCA | 72.37 | 70.61 | 74.13 | 0.481 |
LDA | 84.88 | 81.22 | 88.54 | 0.703 |
GR | 77.85 | 76.38 | 79.32 | 0.561 |
PCC | 82.38 | 79.12 | 85.64 | 0.662 |
RFA | 81.32 | 78.18 | 84.46 | 0.665 |
Hybrid feature (before SHAP) | 89.36 | 88.21 | 90.44 | 0.797 |
Hybrid feature (after SHAP) | 90.12 | 88.17 | 92.05 | 0.825 |
Table 7 demonstrates that the DNN model attained superior performance with a hybrid feature vector compared to the individual feature extraction technique. The DNN model achieved an average accuracy of 89.36%, sensitivity of 90.44%, specificity of 88.21%, and an MCC of 0.797 utilizing a hybrid feature vector. In order to further enhance the performance of the proposed DNN model, we apply the feature selection method, i.e., SHAP. SHAP-selected features attained an accuracy of 90.12%, sensitivity of 92.05%, specificity of 88.17%, and a Matthew's correlation coefficient of 0.825. The result demonstrates the effectiveness of feature selection in enhancing prediction accuracy and target result identification in deep neural networks (DNNs).
In this section, we compare the performance of various classification methods based on accuracy (ACC), sensitivity (SN), specificity (SP), and Matthew's correlation coefficient (MCC). The classifiers being assessed are random forest (RF) [23], support vector machine (SVM) [24], and k-nearest neighbor (KNN) [25]. RF is a popular ensemble learning method for classification and prediction, building multiple decision trees from random samples. SVM is widely used in life sciences for linear and nonlinear classification, as it finds the best boundary to separate classes [26]. KNN, commonly used in image processing, is a distance-based method that classifies by comparing examples. Table 8 shows the overall performance comparison of the different classifiers on the benchmark dataset.
Method | ACC (%) | SN (%) | SP (%) | MCC |
Proposed DNN | 90.12 | 92.05 | 88.17 | 0.825 |
SVM | 89.66 | 87.75 | 91.67 | 0.794 |
RF | 87.86 | 84.98 | 90.91 | 0.759 |
KNN | 85.96 | 84.98 | 86.96 | 0.719 |
The proposed DNN model achieved the highest accuracy of 90.12%, with a sensitivity of 92.05% and specificity of 88.17%, resulting in an MCC of 0.825. Comparatively, among the traditional machine learning models, the SVM demonstrated an accuracy of 89.66%, a sensitivity of 87.75%, a specificity of 91.67%, and an MCC of 0.794. Similarly, the RF and KNN models achieved lower accuracy, at 87.86% and 85.96%, respectively. Overall, the proposed DNN demonstrates the most balanced and robust performance, improving average accuracy by 2.29%. The superior performance of the DNN model compared to traditional machine learning algorithms can be attributed to its use of multi-stack processing layers. These layers enable DNN to effectively handle complex datasets with high nonlinearity, a task that traditional ML algorithms struggle with due to their reliance on single-stack processing layers, which are insufficient for such intricate data. These findings confirm that the DNN model outperforms conventional learning methods in predicting liver and non-liver diseases.
Furthermore, we also compared the performance of the DNN model in terms of AUC (area under the ROC curve) [27] with traditional learning algorithms, as shown in Figure 5. The figure shows that the DNN model achieved the highest AUC value compared to the other algorithms. Specifically, the DNN model had an AUC of 0.921, while the SVM, RF, and KNN algorithms had AUC values of 0.901, 0.889, and 0.872, respectively. The main reason for the DNN model's superior performance is its multi-layer structure (hidden layers) with backpropagation, which effectively optimizes weights. This allows the DNN to handle complex and highly nonlinear datasets better than other models, which use a single-layer processing approach and struggle with such data. Additionally, the confusion matrix analysis in Figure 6 shows that our proposed DNN model achieves balanced recognition of both positive and negative. This strongly confirms the model's accuracy and reliability in distinguishing between positive and negative samples, making it particularly suitable for predictive tasks in the specific problem domain
In this section, we compare the performance of the proposed model with the existing state-of-the-art models, as mentioned in [10],[28]–[30]. Table 9 shows the comparative findings. Moreover, we provide the proposed models from published research works based on the same dataset. Table 9 shows that the proposed model provided the best prediction accuracy among four previously published predictors. For example, K. Gupta et al. [29] achieved an accuracy level of 63%, whereas the recently published predictor, R. Amin et al. [10], achieved an accuracy of 88.10%. Similarly, D. Elias et al. [30] achieved 80.10%. These results confirm that the proposed model performed better than the existing predictors, with an average accuracy improvement of 13.93%.
Method | Classifiers | ACC (%) |
K. Gupta et al. [29] | Light GB | 63.00 |
Altaf et al. [28] | Voting | 73.56 |
D. Elias et al. [30] | Voting | 80.10 |
R. Amin et al. [10] | Random Forest | 88.10 |
Proposed DNN | DNN | 90.12 |
Liver disease (LD) is a severe condition that poses significant risks to human health and requires timely medical intervention. Healthcare professionals rely on neurological techniques for evaluating and diagnosing affected individuals. In this paper, we explored the prediction of chronic LD using a deep neural network algorithm enhanced with integrated and optimized features. Our model used several techniques to assess feature importance, including PCA, LDA, gain ratio, Pearson correlation coefficient, and ReliefF algorithm. We also employed SHAP analysis to interpret complex features and select the most relevant ones for accurate LD prediction. Our novel approach demonstrated superior performance compared to existing models, which can be attributed to the implementation of efficient feature extraction methods, advanced compression techniques, and an optimized feature selection process. The results show that our model enhances discriminative ability, offering a more reliable tool for the early detection of chronic liver disease. Integrating AI and machine learning into clinical settings holds great potential for advancing disease detection and improving patient outcomes.
In the future, we plan to create an accessible web platform for biologists to use this model. Additionally, we aim to expand the dataset, explore new features, and implement more advanced algorithms to refine further and validate our model's predictive capabilities.
[1] |
Q. Yuan, K. Chen, Y. Yu, N. Q. K. Le, M. C. H. Chua, Prediction of anticancer peptides based on an ensemble model of deep learning and machine learning using ordinal positional encoding, Briefings Bioinf., 24 (2023), bbac630. https://doi.org/10.1093/bib/bbac630 doi: 10.1093/bib/bbac630
![]() |
[2] |
N. Q. K. Le, Potential of deep representative learning features to interpret the sequence information in proteomics, Proteomics, 22 (2022), 2100232. https://doi.org/10.1002/pmic.202100232 doi: 10.1002/pmic.202100232
![]() |
[3] |
P. Aggarwal, N. K. Mishra, B. Fatimah, P. Singh, A. Gupta, S. D. Joshi, COVID-19 image classification using deep learning: Advances, challenges and opportunities, Comput. Biol. Med., 144 (2022), 105350. https://doi.org/10.1016/j.compbiomed.2022.105350 doi: 10.1016/j.compbiomed.2022.105350
![]() |
[4] |
O. S. Albahri, A. A. Zaidan, A. S. Albahri, B. B. Zaidan, K. H. Abdulkareem, Z. T. Al-qaysi, et al., Systematic review of artificial intelligence techniques in the detection and classification of COVID-19 medical images in terms of evaluation and benchmarking: Taxonomy analysis, challenges, future solutions and methodological aspects, J. Infect. Public Health, 13 (2020), 1381–1396. https://doi.org/10.1016/j.jiph.2020.06.028 doi: 10.1016/j.jiph.2020.06.028
![]() |
[5] |
T. W. Cenggoro, B. Pardamean, A systematic literature review of machine learning application in COVID-19 medical image classification, Procedia Comput. Sci., 216 (2023), 749–756. https://doi.org/10.1016/j.procs.2022.12.192 doi: 10.1016/j.procs.2022.12.192
![]() |
[6] |
Y. Hu, K. Liu, K. Ho, D. Riviello, J. Brown, A. R. Chang, et al., A simpler machine learning model for acute kidney injury risk stratification in hospitalized patients, J. Clin. Med., 11 (2022), 5688. https://doi.org/10.3390/jcm11195688 doi: 10.3390/jcm11195688
![]() |
[7] |
E. Hussain, M. Hasan, M. A. Rahman, I. Lee, T. Tamanna, M. Z. Parvez, CoroDet: A deep learning based classification for COVID-19 detection using chest X-ray images, Chaos, Solitons Fractals, 142 (2021), 110495. https://doi.org/10.1016/j.chaos.2020.110495 doi: 10.1016/j.chaos.2020.110495
![]() |
[8] |
M. A. Ozdemir, G. D. Ozdemir, O. Guren, Classification of COVID-19 electrocardiograms by using hexaxial feature mapping and deep learning, BMC Med. Inf. Decis. Making, 21 (2021), 170. https://doi.org/10.1186/s12911-021-01521-x doi: 10.1186/s12911-021-01521-x
![]() |
[9] |
A. M. Ismael, A. Şengür, Deep learning approaches for COVID-19 detection based on chest X-ray images, Expert Syst. Appl., 164 (2021), 114054. https://doi.org/10.1016/j.eswa.2020.114054 doi: 10.1016/j.eswa.2020.114054
![]() |
[10] |
K. Kc, Z. Yin, M. Wu, Z. Wu, Evaluation of deep learning-based approaches for COVID-19 classification based on chest X-ray images, Signal, Image Video Process., 15 (2021), 959–966. https://doi.org/10.1007/s11760-020-01820-2 doi: 10.1007/s11760-020-01820-2
![]() |
[11] |
G. Muhammad, M. S. Hossain, COVID-19 and non-COVID-19 classification using multi-layers fusion from lung ultrasound images, Inf. Fusion, 72 (2021), 80–88. https://doi.org/10.1016/j.inffus.2021.02.013 doi: 10.1016/j.inffus.2021.02.013
![]() |
[12] |
X. Wang, X. Deng, Q. Fu, Q. Zhou, J. Feng, H. Ma, et al., A weakly-supervised framework for COVID-19 classification and lesion localization from chest CT, IEEE Trans. Med. Imaging, 39 (2020), 2615–2625. https://doi.org/10.1109/TMI.2020.2995965 doi: 10.1109/TMI.2020.2995965
![]() |
[13] |
M. Riaz, M. Bashir, I. Younas, Metaheuristics based COVID-19 detection using medical images: A review, Comput. Biol. Med., 2022 (2022), 105344. https://doi.org/10.1016/j.compbiomed.2022.105344 doi: 10.1016/j.compbiomed.2022.105344
![]() |
[14] |
D. Zhu, S. Wang, C. Zhou, S. Yan, J. Xue, Human memory optimization algorithm: A memory-inspired optimizer for global optimization problems, Expert Syst. Appl., 237 (2024), 121597. https://doi.org/10.1016/j.eswa.2023.121597 doi: 10.1016/j.eswa.2023.121597
![]() |
[15] |
D. Zhu, S. Wang, J. Shen, C. Zhou, T. Li, S. Yan, A multi-strategy particle swarm algorithm with exponential noise and fitness-distance balance method for low-altitude penetration in secure space, J. Comput. Sci., 74 (2023), 102149. https://doi.org/10.1016/j.jocs.2023.102149 doi: 10.1016/j.jocs.2023.102149
![]() |
[16] | J. Kennedy, R. Eberhart, Particle swarm optimization, in Proceedings of ICNN'95-International Conference on Neural Networks, IEEE, (1995), 1942–1948. https://doi.org/10.1109/ICNN.1995.488968 |
[17] | K. V. Price, Differential evolution, in Handbook of Optimization: From Classical to Modern Approach, Springer, (2013), 187–214. https://doi.org/10.1007/978-3-642-30504-7_8 |
[18] | M. Dorigo, M. Birattari, T. Stutzle, Ant colony optimization, IEEE Comput. Intell. Mag., 1 (2006), 28–39. https://doi.org/10.1109/MCI.2006.329691 |
[19] | S. Mirjalili, S. M. Mirjalili, A. Lewis, Grey wolf optimizer, Adv. Eng. Software, 69 (2014), 46–61. https://doi.org/10.1016/j.advengsoft.2013.12.007 |
[20] |
S. Mirjalili, A. Lewis, The whale optimization algorithm, Adv. Eng. Software, 95 (2016), 51–67. https://doi.org/10.1016/j.advengsoft.2016.01.008 doi: 10.1016/j.advengsoft.2016.01.008
![]() |
[21] |
J. Xue, B. Shen, A novel swarm intelligence optimization approach: Sparrow search algorithm, Syst. Sci. Control Eng., 8 (2020), 22–34. https://doi.org/10.1080/21642583.2019.1708830 doi: 10.1080/21642583.2019.1708830
![]() |
[22] |
A. A. Heidari, S. Mirjalili, H. Faris, I. Aljarah, M. Mafarja, H. Chen, Harris hawks optimization: Algorithm and applications, Future Gener. Comput. Syst., 97 (2019), 849–872. https://doi.org/10.1016/j.future.2019.02.028 doi: 10.1016/j.future.2019.02.028
![]() |
[23] |
W. Zhao, Z. Zhang, L. Wang, Manta ray foraging optimization: An effective bio-inspired optimizer for engineering applications, Eng. Appl. Artif. Intell., 87 (2020), 103300. https://doi.org/10.1016/j.engappai.2019.103300 doi: 10.1016/j.engappai.2019.103300
![]() |
[24] |
M. Abdel-Basset, D. El-Shahat, M. Jameel, M. Abouhawwash, Exponential distribution optimizer (EDO): A novel math-inspired algorithm for global optimization and engineering problems, Artif. Intell. Rev., 56 (2023), 9329–9400. https://doi.org/10.1007/s10462-023-10403-9 doi: 10.1007/s10462-023-10403-9
![]() |
[25] |
A. Dixit, A. Mani, R. Bansal, CoV2-Detect-Net: Design of COVID-19 prediction model based on hybrid DE-PSO with SVM using chest X-ray images, Inf. Sci., 571 (2021), 676–692. https://doi.org/10.1016/j.ins.2021.03.062 doi: 10.1016/j.ins.2021.03.062
![]() |
[26] |
D. A. D. Júnior, L. B. da Cruz, J. O. B. Diniz, G. L. F. da Silva, G. B. Junior, A. C. Silva, et al., Automatic method for classifying COVID-19 patients based on chest X-ray images, using deep features and PSO-optimized XGBoost, Expert Syst. Appl., 183 (2021), 115452. https://doi.org/10.1016/j.eswa.2021.115452 doi: 10.1016/j.eswa.2021.115452
![]() |
[27] |
M. A. A. Albadr, S. Tiun, M. Ayob, F. T. AL-Dhief, Particle swarm optimization-based extreme learning machine for COVID-19 detection, Cognit. Comput., 2022 (2022), 1–16. https://doi.org/10.1007/s12559-022-10063-x doi: 10.1007/s12559-022-10063-x
![]() |
[28] |
M. A. Elaziz, K. M. Hosny, A. Salah, M. M. Darwish, S. Lu, A. T. Sahlol, New machine learning method for image-based diagnosis of COVID-19, PLoS One, 15 (2020), e0235187. https://doi.org/10.1371/journal.pone.0235187 \newpage doi: 10.1371/journal.pone.0235187
![]() |
[29] |
E. S. M. El-Kenawy, S. Mirjalili, A. Ibrahim, M. Alrahmawy, M. El-Said, R. M. Zaki, et al., Advanced meta-heuristics, convolutional neural networks, and feature selectors for efficient COVID-19 X-ray chest image classification, IEEE Access, 9 (2021), 36019–36037. https://doi.org/10.1109/ACCESS.2021.3061058 doi: 10.1109/ACCESS.2021.3061058
![]() |
[30] |
S. Pathan, P. C. Siddalingaswamy, P. Kumar, M. M. M. Pai, T. Ali, U. R. Acharya, Novel ensemble of optimized CNN and dynamic selection techniques for accurate COVID-19 screening using chest CT images, Comput. Biol. Med., 137 (2021), 104835. https://doi.org/10.1016/j.compbiomed.2021.104835 doi: 10.1016/j.compbiomed.2021.104835
![]() |
[31] |
A. Basu, K. H. Sheikh, E. Cuevas, R. Sarkar, COVID-19 detection from CT scans using a two-stage framework, Expert Syst. Appl., 193 (2022), 116377. https://doi.org/10.1016/j.eswa.2021.116377 doi: 10.1016/j.eswa.2021.116377
![]() |
[32] |
M. H. Nadimi-Shahraki, H. Zamani, S. Mirjalili, Enhanced whale optimization algorithm for medical feature selection: A COVID-19 case study, Comput. Biol. Med., 148 (2022), 105858. https://doi.org/10.1016/j.compbiomed.2022.105858 doi: 10.1016/j.compbiomed.2022.105858
![]() |
[33] | S. Elghamrawy, A. E. Hassanien, Diagnosis and prediction model for COVID-19 patient's response to treatment based on convolutional neural networks and whale optimization algorithm using CT images, preprint, MedRxiv, 2020. https://doi.org/10.1101/2020.04.16.20063990 |
[34] |
T. Goel, R. Murugan, S. Mirjalili, D. K. Chakrabartty, OptCoNet: An optimized convolutional neural network for an automatic diagnosis of COVID-19, Appl. Intell., 51 (2021), 1351–1366. https://doi.org/10.1007/s10489-020-01904-z doi: 10.1007/s10489-020-01904-z
![]() |
[35] |
T. Hu, M. Khishe, M. Mohammadi, G. Parvizi, S. H. T. Karim, T. A. Rashid, Real-time COVID-19 diagnosis from X-Ray images using deep CNN and extreme learning machines stabilized by chimp optimization algorithm, Biomed. Signal Process. Control, 68 (2021), 102764. https://doi.org/10.1016/j.bspc.2021.102764 doi: 10.1016/j.bspc.2021.102764
![]() |
[36] |
D. Singh, V. Kumar, M. Kaur, Classification of COVID-19 patients from chest CT images using multi-objective differential evolution–based convolutional neural networks, Eur. J. Clin. Microbiol. Infect. Dis., 39 (2020), 1379–1389. https://doi.org/10.1007/s10096-020-03901-z doi: 10.1007/s10096-020-03901-z
![]() |
[37] | M. S. Iraji, M. Feizi-Derakhshi, J. Tanha, Deep learning for COVID-19 diagnosis based feature selection using binary differential evolution algorithm, preprint, 2021, arXiv: PPR343118. |
[38] |
A. M. Sahan, A. S. Al-Itbi, J. S. Hameed, COVID-19 detection based on deep learning and artificial bee colony, Periodicals Eng. Nat. Sci., 9 (2021), 29–36. http://doi.org/10.21533/pen.v9i1.1774 doi: 10.21533/pen.v9i1.1774
![]() |
[39] | F. Sadeghi, O. Rostami, M. K. Yi, A deep learning approach for detecting COVID-19 using the chest X-ray images, CMC-Comput. Mater. Continua, 75 (2023), 751–768. |
[40] |
H. M. Balaha, E. M. El-Gendy, M. M. Saafan, CovH2SD: A COVID-19 detection approach based on Harris Hawks Optimization and stacked deep learning, Expert Syst. Appl., 186 (2021), 115805. https://doi.org/10.1016/j.eswa.2021.115805 doi: 10.1016/j.eswa.2021.115805
![]() |
[41] |
W. M. Bahgat, H. M. Balaha, Y. AbdulAzeem, M. M. Badawy, An optimized transfer learning-based approach for automatic diagnosis of COVID-19 from chest x-ray images, PeerJ Comput. Sci., 7 (2021), e555. https://doi.org/10.7717/peerj-cs.555 doi: 10.7717/peerj-cs.555
![]() |
[42] | K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, (2016), 770–778. |
[43] |
L. Wen, X. Li, L. Gao, A transfer convolutional neural network for fault diagnosis based on ResNet-50, Neural Comput. Appl., 32 (2020), 6111–6124. https://doi.org/10.1007/s00521-019-04097-w doi: 10.1007/s00521-019-04097-w
![]() |
[44] |
J. Yang, J. Yu, C. Huang, Adaptive multistrategy ensemble particle swarm optimization with Signal-to-Noise ratio distance metric, Inf. Sci., 612 (2022), 1066–1094. https://doi.org/10.1016/j.ins.2022.07.165 doi: 10.1016/j.ins.2022.07.165
![]() |
[45] |
D. Zhu, Z. Huang, S. Liao, C. Zhou, S. Yan, G. Chen, Improved bare bones particle swarm optimization for DNA sequence design, IEEE Trans. Nanobiosci., 22 (2022), 603–613. https://doi.org/10.1109/TNB.2022.3220795 doi: 10.1109/TNB.2022.3220795
![]() |
[46] |
H. T. Kahraman, S. Aras, E. Gedikli, Fitness-distance balance (FDB): A new selection method for meta-heuristic search algorithms, Knowledge-Based Syst., 190 (2020), 105169. https://doi.org/10.1016/j.knosys.2019.105169 doi: 10.1016/j.knosys.2019.105169
![]() |
[47] |
R. Zheng, A. G. Hussien, R. Qaddoura, H. Jia, L. Abualigah, S. Wang, A multi-strategy enhanced African vultures optimization algorithm for global optimization problems, J. Comput. Des. Eng., 10 (2023), 329–356. https://doi.org/10.1093/jcde/qwac135 doi: 10.1093/jcde/qwac135
![]() |
[48] |
D. Zhu, S. Wang, C. Zhou, S. Yan, Manta ray foraging optimization based on mechanics game and progressive learning for multiple optimization problems, Appl. Soft Comput., 145 (2023), 110561. https://doi.org/10.1016/j.asoc.2023.110561 doi: 10.1016/j.asoc.2023.110561
![]() |
[49] |
R. Murugan, T. Goel, S. Mirjalili, D. K. Chakrabartty, WOANet: Whale optimized deep neural network for the classification of COVID-19 from radiography images, Biocybern. Biomed. Eng., 41 (2021), 1702–1718. https://doi.org/10.1016/j.bbe.2021.10.004 doi: 10.1016/j.bbe.2021.10.004
![]() |
[50] | C. Szegedy, W. Liu, Y. Jia, Going deeper with convolutions, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, (2015), 1–9. |
[51] | K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, preprint, arXiv: 1409.1556. |
[52] |
L. Kong, J. Cheng, Classification and detection of COVID-19 X-Ray images based on DenseNet and VGG16 feature fusion, Biomed. Signal Process. Control, 77 (2022), 103772. https://doi.org/10.1016/j.bspc.2022.103772 doi: 10.1016/j.bspc.2022.103772
![]() |
[53] |
A. Karacı, VGGCOV19-NET: Automatic detection of COVID-19 cases from X-ray images using modified VGG19 CNN architecture and YOLO algorithm, Neural Comput. Appl., 34 (2022), 8253–8274. https://doi.org/10.1007/s00521-022-06918-x doi: 10.1007/s00521-022-06918-x
![]() |
[54] |
A. Krizhevsky, I. Sutskever, G. E. Hinton, Imagenet classification with deep convolutional neural networks, Commun. ACM, 60 (2017), 84–90. https://doi.org/10.1145/3065386 doi: 10.1145/3065386
![]() |
[55] |
G. Muhammad, M. S. Hossain, COVID-19 and non-COVID-19 classification using multi-layers fusion from lung ultrasound images, Inf. Fusion, 72 (2021), 80–88. https://doi.org/10.1016/j.inffus.2021.02.013 doi: 10.1016/j.inffus.2021.02.013
![]() |
[56] | I. Bankman, Handbook of Medical Image Processing and Analysis, Elsevier, 2008. |
[57] |
Y. Song, S. Zheng, L. Li, X. Zhang, X. Zhang, Z. Huang, et al., Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images, IEEE/ACM Trans. Comput. Biol. Bioinf., 18 (2021), 2775–2780. https://doi.org/10.1109/TCBB.2021.3065361 doi: 10.1109/TCBB.2021.3065361
![]() |
[58] |
Y. Pathak, P. K. Shukla, K. V. Arya, Deep bidirectional classification model for COVID-19 disease infected patients, IEEE/ACM Trans. Comput. Biol. Bioinf., 18 (2020), 1234–1241. https://doi.org/10.1109/TCBB.2020.3009859 doi: 10.1109/TCBB.2020.3009859
![]() |
[59] |
Y. Pathak, P. K. Shukla, A. Tiwari, S. Stalin, S. Singh, P. K. Shukla, Deep transfer learning based classification model for COVID-19 disease, IRBM, 43 (2022), 87–92. https://doi.org/10.1016/j.irbm.2020.05.003 doi: 10.1016/j.irbm.2020.05.003
![]() |
[60] |
X. Fan, X. Feng, Y. Dong, H. Hou, COVID-19 CT image recognition algorithm based on transformer and CNN, Displays, 72 (2022), 102150. https://doi.org/10.1016/j.displa.2022.102150 doi: 10.1016/j.displa.2022.102150
![]() |
[61] |
A. S. Ebenezer, S. D. Kanmani, M. Sivakumar, S. J. Priya, Effect of image transformation on EfficientNet model for COVID-19 CT image classification, Mater. Today Proc., 51 (2022), 2512–2519. https://doi.org/10.1016/j.matpr.2021.12.121 doi: 10.1016/j.matpr.2021.12.121
![]() |
[62] |
N. S. Shaik, T. K. Cherukuri, Transfer learning based novel ensemble classifier for COVID-19 detection from chest CT-scans, Comput. Biol. Med., 141 (2022), 105127. https://doi.org/10.1016/j.compbiomed.2021.105127 doi: 10.1016/j.compbiomed.2021.105127
![]() |
1. | Max C. W. Engelen, River Betting, Christos Strydis, SimHH: A Versatile, Multi-GPU Simulator for Extended Hodgkin-Huxley Networks, 2025, 13, 2169-3536, 46865, 10.1109/ACCESS.2025.3550444 |
Feature | Description |
Gender | Participant's gender |
Albumin (ALB) | Albumin level |
Direct bilirubin (DB) | Direct bilirubin level |
Age (years) | Age of participants (4–90 years) |
Total bilirubin (TB) | Total bilirubin level |
Aspartate aminotransferase (SGOT) | Aspartate aminotransferase level |
Total protein (TP) | Total protein level |
Alkaline phosphatase (ALP) | Alkaline phosphatase level |
Alanine aminotransferase (SGPT) | Alanine aminotransferase level |
Albumin and globulin ratio (AGR) | Albumin and globulin ratio |
Features | Min | Mean | Max | ± Stdv |
TB | 0.4 | 2.65 | 75 | ± 5.32 |
SGOT | 10 | 88.78 | 4929 | ± 245.07 |
Age | 4 | 43.55 | 90 | ± 16.28 |
ALP | 63 | 267.26 | 2110 | ± 212.62 |
SGPT | 10 | 66.78 | 2000 | ± 155.16 |
AGR | 0.3 | 0.98 | 2.8 | ± 0.30 |
DB | 0.1 | 1.16 | 19.7 | ± 2.42 |
ALB | 0.9 | 3.19 | 5.5 | ± 0.76 |
TP | 2.7 | 6.5 | 9.6 | ± 1.02 |
Feature | PCC | GR | RFA |
SGPT | 0.2141 | 0.0701 | 0.2613 |
TB | 0.2874 | 0.1373 | 0.2848 |
Gender | 0.0857 | 0.0065 | 0.025 |
SGOT | 0.2017 | 0.1005 | 0.2393 |
AGR | 0.2046 | 0.0822 | 0.2599 |
DB | 0.3205 | 0.1421 | 0.2895 |
ALB | 0.1836 | 0.0408 | 0.2883 |
Age | 0.1596 | 0.0372 | 0.2625 |
ALP | 0.246 | 0.0867 | 0.1936 |
TP | 0.0443 | 0.00 | 0.2161 |
Method | Number of features (dimensionality) |
Principal component analysis (PCA) | 10 |
Linear discriminant analysis (LDA) | 10 |
Gain ratio (GR) | 10 |
Pearson correlation coefficient (PCC) | 10 |
ReliefF algorithm (RFA) | 10 |
Hybrid features | 50 |
Parameters list | Optimize values |
Hidden layers | 4 |
Updater | ADAGRAD function |
Hidden layers neurons | 23-17-14-6 |
Seed | 12345L |
Epoch | 10 |
Momentum | 0.9 |
LR | 0.01 |
Dropout | 0.25 |
Regularization l2 | 0.001 |
Activation functions | ReLU & Softmax |
Optimizer | SGD Method |
Weight initialization function | XAVIER function |
LR | ReLU | Tanh | Sigmoid |
0.01 | 90.12 | 88.10 | 89.44 |
0.02 | 89.91 | 88.09 | 89.23 |
0.03 | 89.45 | 87.71 | 89.01 |
0.04 | 89.16 | 87.58 | 88.80 |
0.05 | 88.49 | 87.38 | 88.17 |
0.06 | 87.82 | 87.19 | 87.54 |
0.07 | 87.55 | 86.99 | 86.91 |
0.08 | 87.09 | 86.80 | 86.28 |
0.09 | 87.05 | 86.60 | 85.65 |
0.10 | 87.03 | 86.41 | 85.02 |
Methods | ACC (%) | SP (%) | SN (%) | MCC |
PCA | 72.37 | 70.61 | 74.13 | 0.481 |
LDA | 84.88 | 81.22 | 88.54 | 0.703 |
GR | 77.85 | 76.38 | 79.32 | 0.561 |
PCC | 82.38 | 79.12 | 85.64 | 0.662 |
RFA | 81.32 | 78.18 | 84.46 | 0.665 |
Hybrid feature (before SHAP) | 89.36 | 88.21 | 90.44 | 0.797 |
Hybrid feature (after SHAP) | 90.12 | 88.17 | 92.05 | 0.825 |
Method | ACC (%) | SN (%) | SP (%) | MCC |
Proposed DNN | 90.12 | 92.05 | 88.17 | 0.825 |
SVM | 89.66 | 87.75 | 91.67 | 0.794 |
RF | 87.86 | 84.98 | 90.91 | 0.759 |
KNN | 85.96 | 84.98 | 86.96 | 0.719 |
Feature | Description |
Gender | Participant's gender |
Albumin (ALB) | Albumin level |
Direct bilirubin (DB) | Direct bilirubin level |
Age (years) | Age of participants (4–90 years) |
Total bilirubin (TB) | Total bilirubin level |
Aspartate aminotransferase (SGOT) | Aspartate aminotransferase level |
Total protein (TP) | Total protein level |
Alkaline phosphatase (ALP) | Alkaline phosphatase level |
Alanine aminotransferase (SGPT) | Alanine aminotransferase level |
Albumin and globulin ratio (AGR) | Albumin and globulin ratio |
Features | Min | Mean | Max | ± Stdv |
TB | 0.4 | 2.65 | 75 | ± 5.32 |
SGOT | 10 | 88.78 | 4929 | ± 245.07 |
Age | 4 | 43.55 | 90 | ± 16.28 |
ALP | 63 | 267.26 | 2110 | ± 212.62 |
SGPT | 10 | 66.78 | 2000 | ± 155.16 |
AGR | 0.3 | 0.98 | 2.8 | ± 0.30 |
DB | 0.1 | 1.16 | 19.7 | ± 2.42 |
ALB | 0.9 | 3.19 | 5.5 | ± 0.76 |
TP | 2.7 | 6.5 | 9.6 | ± 1.02 |
Feature | PCC | GR | RFA |
SGPT | 0.2141 | 0.0701 | 0.2613 |
TB | 0.2874 | 0.1373 | 0.2848 |
Gender | 0.0857 | 0.0065 | 0.025 |
SGOT | 0.2017 | 0.1005 | 0.2393 |
AGR | 0.2046 | 0.0822 | 0.2599 |
DB | 0.3205 | 0.1421 | 0.2895 |
ALB | 0.1836 | 0.0408 | 0.2883 |
Age | 0.1596 | 0.0372 | 0.2625 |
ALP | 0.246 | 0.0867 | 0.1936 |
TP | 0.0443 | 0.00 | 0.2161 |
Method | Number of features (dimensionality) |
Principal component analysis (PCA) | 10 |
Linear discriminant analysis (LDA) | 10 |
Gain ratio (GR) | 10 |
Pearson correlation coefficient (PCC) | 10 |
ReliefF algorithm (RFA) | 10 |
Hybrid features | 50 |
Parameters list | Optimize values |
Hidden layers | 4 |
Updater | ADAGRAD function |
Hidden layers neurons | 23-17-14-6 |
Seed | 12345L |
Epoch | 10 |
Momentum | 0.9 |
LR | 0.01 |
Dropout | 0.25 |
Regularization l2 | 0.001 |
Activation functions | ReLU & Softmax |
Optimizer | SGD Method |
Weight initialization function | XAVIER function |
LR | ReLU | Tanh | Sigmoid |
0.01 | 90.12 | 88.10 | 89.44 |
0.02 | 89.91 | 88.09 | 89.23 |
0.03 | 89.45 | 87.71 | 89.01 |
0.04 | 89.16 | 87.58 | 88.80 |
0.05 | 88.49 | 87.38 | 88.17 |
0.06 | 87.82 | 87.19 | 87.54 |
0.07 | 87.55 | 86.99 | 86.91 |
0.08 | 87.09 | 86.80 | 86.28 |
0.09 | 87.05 | 86.60 | 85.65 |
0.10 | 87.03 | 86.41 | 85.02 |
Methods | ACC (%) | SP (%) | SN (%) | MCC |
PCA | 72.37 | 70.61 | 74.13 | 0.481 |
LDA | 84.88 | 81.22 | 88.54 | 0.703 |
GR | 77.85 | 76.38 | 79.32 | 0.561 |
PCC | 82.38 | 79.12 | 85.64 | 0.662 |
RFA | 81.32 | 78.18 | 84.46 | 0.665 |
Hybrid feature (before SHAP) | 89.36 | 88.21 | 90.44 | 0.797 |
Hybrid feature (after SHAP) | 90.12 | 88.17 | 92.05 | 0.825 |
Method | ACC (%) | SN (%) | SP (%) | MCC |
Proposed DNN | 90.12 | 92.05 | 88.17 | 0.825 |
SVM | 89.66 | 87.75 | 91.67 | 0.794 |
RF | 87.86 | 84.98 | 90.91 | 0.759 |
KNN | 85.96 | 84.98 | 86.96 | 0.719 |
Method | Classifiers | ACC (%) |
K. Gupta et al. [29] | Light GB | 63.00 |
Altaf et al. [28] | Voting | 73.56 |
D. Elias et al. [30] | Voting | 80.10 |
R. Amin et al. [10] | Random Forest | 88.10 |
Proposed DNN | DNN | 90.12 |