Loading [MathJax]/jax/output/SVG/jax.js
Research article

An efficient, lightweight MobileNetV2-based fine-tuned model for COVID-19 detection using chest X-ray images

  • Received: 25 September 2022 Revised: 01 January 2023 Accepted: 02 January 2023 Published: 02 March 2023
  • In recent years, deep learning's identification of cancer, lung disease and heart disease, among others, has contributed to its rising popularity. Deep learning has also contributed to the examination of COVID-19, which is a subject that is currently the focus of considerable scientific debate. COVID-19 detection based on chest X-ray (CXR) images primarily depends on convolutional neural network transfer learning techniques. Moreover, the majority of these methods are evaluated by using CXR data from a single source, which makes them prohibitively expensive. On a variety of datasets, current methods for COVID-19 detection may not perform as well. Moreover, most current approaches focus on COVID-19 detection. This study introduces a rapid and lightweight MobileNetV2-based model for accurate recognition of COVID-19 based on CXR images; this is done by using machine vision algorithms that focused largely on robust and potent feature-learning capabilities. The proposed model is assessed by using a dataset obtained from various sources. In addition to COVID-19, the dataset includes bacterial and viral pneumonia. This model is capable of identifying COVID-19, as well as other lung disorders, including bacterial and viral pneumonia, among others. Experiments with each model were thoroughly analyzed. According to the findings of this investigation, MobileNetv2, with its 92% and 93% training validity and 88% precision, was the most applicable and reliable model for this diagnosis. As a result, one may infer that this study has practical value in terms of giving a reliable reference to the radiologist and theoretical significance in terms of establishing strategies for developing robust features with great presentation ability.

    Citation: Shubashini Velu. An efficient, lightweight MobileNetV2-based fine-tuned model for COVID-19 detection using chest X-ray images[J]. Mathematical Biosciences and Engineering, 2023, 20(5): 8400-8427. doi: 10.3934/mbe.2023368

    Related Papers:

    [1] Jingxu Xiao, Chaowen Chang, Yingying Ma, Chenli Yang, Lu Yuan . Secure multi-path routing for Internet of Things based on trust evaluation. Mathematical Biosciences and Engineering, 2024, 21(2): 3335-3363. doi: 10.3934/mbe.2024148
    [2] Abhishek Savaliya, Rutvij H. Jhaveri, Qin Xin, Saad Alqithami, Sagar Ramani, Tariq Ahamed Ahanger . Securing industrial communication with software-defined networking. Mathematical Biosciences and Engineering, 2021, 18(6): 8298-8313. doi: 10.3934/mbe.2021411
    [3] Jinjing Huang, Xi Wang . Influenced node discovery in a temporal contact network based on common nodes. Mathematical Biosciences and Engineering, 2023, 20(8): 13660-13680. doi: 10.3934/mbe.2023609
    [4] Lei Jin, Sixiang Lin, Binglei Xie, Lin Liu . A vulnerability-based vehicle routing approach for solving capacitated arc routing problem in urban snow plowing operations. Mathematical Biosciences and Engineering, 2021, 18(1): 166-181. doi: 10.3934/mbe.2021009
    [5] Zhibin Zuo, Rongyu He, Xianwei Zhu, Chaowen Chang . A novel software-defined network packet security tunnel forwarding mechanism. Mathematical Biosciences and Engineering, 2019, 16(5): 4359-4381. doi: 10.3934/mbe.2019217
    [6] Jiushuang Wang, Ying Liu, Huifen Feng . IFACNN: efficient DDoS attack detection based on improved firefly algorithm to optimize convolutional neural networks. Mathematical Biosciences and Engineering, 2022, 19(2): 1280-1303. doi: 10.3934/mbe.2022059
    [7] Ningning Zhao, Shihao Cui . Study on 4D taxiing path planning of aircraft based on spatio-temporal network. Mathematical Biosciences and Engineering, 2023, 20(3): 4592-4608. doi: 10.3934/mbe.2023213
    [8] Tingting Yang, Yi He . Design of intelligent robots for tourism management service based on green computing. Mathematical Biosciences and Engineering, 2023, 20(3): 4798-4815. doi: 10.3934/mbe.2023222
    [9] Hao Yuan, Qiang Chen, Hongbing Li, Die Zeng, Tianwen Wu, Yuning Wang, Wei Zhang . Improved beluga whale optimization algorithm based cluster routing in wireless sensor networks. Mathematical Biosciences and Engineering, 2024, 21(3): 4587-4625. doi: 10.3934/mbe.2024202
    [10] Meng Jiang, Bo Zhou, Lei Chen . Identification of drug side effects with a path-based method. Mathematical Biosciences and Engineering, 2022, 19(6): 5754-5771. doi: 10.3934/mbe.2022269
  • In recent years, deep learning's identification of cancer, lung disease and heart disease, among others, has contributed to its rising popularity. Deep learning has also contributed to the examination of COVID-19, which is a subject that is currently the focus of considerable scientific debate. COVID-19 detection based on chest X-ray (CXR) images primarily depends on convolutional neural network transfer learning techniques. Moreover, the majority of these methods are evaluated by using CXR data from a single source, which makes them prohibitively expensive. On a variety of datasets, current methods for COVID-19 detection may not perform as well. Moreover, most current approaches focus on COVID-19 detection. This study introduces a rapid and lightweight MobileNetV2-based model for accurate recognition of COVID-19 based on CXR images; this is done by using machine vision algorithms that focused largely on robust and potent feature-learning capabilities. The proposed model is assessed by using a dataset obtained from various sources. In addition to COVID-19, the dataset includes bacterial and viral pneumonia. This model is capable of identifying COVID-19, as well as other lung disorders, including bacterial and viral pneumonia, among others. Experiments with each model were thoroughly analyzed. According to the findings of this investigation, MobileNetv2, with its 92% and 93% training validity and 88% precision, was the most applicable and reliable model for this diagnosis. As a result, one may infer that this study has practical value in terms of giving a reliable reference to the radiologist and theoretical significance in terms of establishing strategies for developing robust features with great presentation ability.



    Machine-learning algorithms are well-known approaches with a wide range of applications in biomedical signal processing, such as classification [1], regression [2], and optimization [3]. In medical applications, diagnosis, recognition, or prediction are the most important tasks that can be handled using classification techniques [4]. In literature, a wide range of applications can be found based on the classification techniques such as emotion recognition [5], text classification [6], activity recognition [7], and epileptic seizure classification [8]. In most classification techniques, the classification performance strongly depends on the input features of the classification algorithms. Hence, feature extraction, the process of summarizing the data into some indexes, and feature selection, the process of finding the optimum combination of the extracted features, play important roles in classification approaches [9].

    Among the diverse features that can be extracted from any data, fractal-based features are known as powerful nonlinear tools for measuring data complexity [10]. For example, it was shown that an individual's eye movements are closely related to fractal patterns [11]. Also, it was revealed that the fractality of the electroencephalography (EEG) signals is significantly reduced in schizophrenia [12]. Furthermore, it found that the memory content could increase the fractality of the EEG signals [13]. Another class of powerful features is the features derived from transformed data. For instance, the features based on the time-frequency transforms strongly help to predict atrial fibrillation [14], diagnosis of heart failure [15], and detection of mitral valve prolapse [16]. Such transforms are sometimes not directly used for feature extraction; however, they can be beneficial in the data preparation processes prior to feature extraction [17].

    The study of the brain's cognitive functions, such as working memory and over and covert visual attention, are interesting areas of research. Thus, many studies have been conducted to investigate how these top-down functions influence neural spiking activity in different areas of the brain [18,19,20]. Nonetheless, different studies claimed that no memory-related modulation could be found in the neural spiking activity of the middle temporal (MT) cortex [21,22,23]. On the other hand, a recent study showed that working memory increased the fractionality of the firing rate signals [24]. Many studies have shown the neural correlate of working memory as an increase in the firing rate of neurons in the prefrontal [19,25,26], parietal [27], and visual [25] cortices; however, no such memory-related change in the average spiking activity of neurons has been reported in extrastriate cortex including V4 [19] and the middle temporal (MT) cortex [19,25] even after applying machine learning techniques [25]. In this paper, we will focus on the maintenance of spatial information, which has been shown to be sent directly from frontal eye field to the extrastriate cortex through feedback connections in form of persistent spiking activity [19]. Although spatial working memory does not increase the average firing rate of extrastriate neurons, it significantly enhances the sensitivity of individual neurons to incoming visual stimuli [26], alters the correlated activity of the population of neurons [18], increases the power of the local field potential (LFP) in the frequency band of alpha-beta, enhances the spike-phase coherency of MT responses in the same frequency range [28] and increases the fractionality of MT spiking activity [24].

    As mentioned above, a recent study revealed that when the spiking activity of neurons is mapped to the fractal dimension feature space, the content of working memory can be captured [24]. In this study, we examined whether a set of linear and/or nonlinear features could reveal the deployment of spatial working memory from the spiking activity of neurons in the area MT. In this regard, we used two different learners, three feature selection algorithms, and two cross-validation methods to show the robustness of our results. The remaining parts of the paper are arranged as follows: Section 2 describes the studied data, extracted features, selection methods, classification algorithms, and classification assessment criteria. Section 3 presents the results, and Section 4 discusses the results and concludes the paper.

    All experimental procedures were performed under the National Institutes of Health Guide for the Care and Use of Laboratory Animals, the Society for Neuroscience Guidelines and Policies. The protocols for all experimental, surgical, and behavioral procedures were approved by the Montana State University Institutional Animal Care and Use Committee.

    In this study, the spiking activity of 131 neurons (stored at 32 kHz), recorded in 11 sessions using electrode arrays, was used. These signals were recorded from the area MT (Figure 1a) of the two male macaque monkeys' brains (five and seven years old). The monkeys, already acquainted with carrying out the memory-guided saccade (MGS) task, were positioned on a customized chair in front of a monitor (24 inches with 144 Hz refresh rate) at a distance of 28.5 cm from their eyes. Initially, during a surgery in which the monkeys were anesthetized, the recording chambers were mounted on the monkeys' skulls in the MT area. The recordings of single electrodes were used to confirm that the chambers are mounted in a desired (MT) area of the monkeys' brains, During the task, the monkeys' heads were restrained, and they received juice as a reward through a syringe pump. This reward delivery, as well as the visual stimulus presentation procedures, were controlled using the MonkeyLogic toolbox in MATLAB software. Moreover, a photodiode was used to record the actual time of the visual stimulus incidence. Then the recorded data were digitalized with a sampling frequency of 32 kHz and stored.

    Figure 1.  a) The area MT of the brain wherein the firing rate signals were recorded. b) Potential positions for the visual stimulus during the visual period of the MGS task. The red dots indicate the locations in the same hemifield (IN conditions), and the green dot indicates the location in the opposite hemifield (OUT condition) as the receptive field of the recorded neuron. The dashed line shows an imaginary receptive field of a sample neuron. c) The schematic representation of the MGS task, including the fixation (appearance of the FP), visual (appearance of visual stimulus), memory (disappearance of the visual stimulus), and saccade (after the disappearance of the FP) periods, as well as the corresponding response of a sample neuron. The average neural response during IN and OUT conditions are shown in green and pink, respectively.

    The MGS task commenced with the appearance of a fixation point (FP) in the center of the monitor. In this period, called the fixation period, the monkeys are required to fixate at the FP for 1000 ms. After that, while the monkeys were fixating on the FP, a visual stimulus appeared in one of the positions, whether IN (same visual hemifield as the neuron's receptive field; displayed with red dots) or OUT (opposite hemifield relative to the RF of the recorded neuron; displayed with a green dot) conditions shown in Figure 1b. The stimulus remained for 1000 ms and then disappeared (visual period). As soon as the visual cue disappeared, the memory period started and lasted for 1000 ms. During the memory period, the monkeys had to keep gazing at the FP and memorize the location of the disappeared cue. Finally, the monkeys were obliged to make a saccade to the remembered location after the disappearance of the FP to receive a reward. The MGS phases are simply portrayed in Figure 1c, along with the recorded spiking activity for a sample MT neuron during the MGS task (more details can be found in see [18,19]).

    Data preparation plays a vital role in obtaining the best results for machine-learning approaches. Here, in the first step, the average spiking activity of individual neurons across trials was obtained. In the next step, the average firing rate signals of each neuron in IN conditions and OUT were considered for extracting the features. In order to eliminate the spiking activity related to the disappearance of the visual stimulus in the memory period, the first 400 ms of this period were trimmed for the feature extraction step. In the final step, the smoothed signals were considered for the following processes.

    Extracting distinguishable features is an essential step in detecting the presence of the working memory. This section introduces the six most used fractal features and some statistical measures. Also, four frequently used transforms in signal processing are described.

    Fractal dimension (FD) is an index of complexity that refers to an object's non-integer dimension geometrically [29]. FD can be obtained using different algorithms; however, in general, it can be obtained based on the number of blocks forming a pattern or covering the data graph. Here, six famous algorithms for calculating the FD are described.

    Higuchi fractal dimension (HFD): HFD is an accurate box-counting method that is the main algorithm for calculating the FD of a graph [30]. Any time series xt:x(1),x(2),,x(N) with a finite number of samples (N) can be expressed as the k sets of xkm where

    xm(k)={x(m),x(m+k),x(m+2k),,x(m+kR)}. (1)

    Here, m=1,2,,k is the initial time, k=1,2,,km is the delay, and R=[Nmk]. Note that [] denotes the integer part of the internal number. Accordingly, the curve length of each subset can be defined as

    Lkm=N1Rk2Ri=1|xm+ikxm+(i1)k|, (2)

    Letting Lkm defines the average value of m curve lengths, it can be written that

    LkmkD, (3)

    where D is FD obtained using the Higuchi algorithm. In this paper, km=30 was determined based on trial and error.

    Katz fractal dimension (KFD): KFD is a distance-based technique for obtaining the FD value that uses the averaged Euclidean distance of two consecutive points in a time series [31]. If L defines the sum of Euclidean distance between every two successive samples (L=N1i=1distEuclidean(ni,ni+1)) and Lm denotes the maximum value of the Euclidean distance between the first and jth sample (Lm=max(distEuclidean(n1,nj))forj=2,,N), the FD can be obtained as

    D=log(N)log(NdL). (4)

    In the above equation, N is the number of samples in a given time series, and D is the FD value with the Katz technique.

    Generalized Hurst exponent (GHE): GHE is known as a fractal index indicating the long-rage dependence of time series that uses the qth-order moment of the distribution [32]. Assuming x(t) as the studied time series with N samples, the the qth-order moment of the distribution can be defined as

    Kτq=|x(t)x(tτ)|qt|x(t)|q, (5)

    where τ is the delay and t indicates the average value of the internal value over the total duration. Then, the fractal dimension of the time series can be obtained through

    Kτq(τTs)qH, (6)

    where Ts is the sampling time, and H refers to the GHE of the time series. In this study, the first-order moment was considered.

    Margaos and Sun fractal dimension (MSFD): MSFD is a morphological method for calculating FD since it tries to cover the data graph employing the morphological operators, i.e., erosion and dilation [33]. The support-limited erosions and dilations with the support set s (s=1,2,,N) and using the structuring element of b can be formulated as

    erosion:{xnsbk=max{xn1,xn,xn+1}fork=1xnsbk=max{xn1sbk,xn+1sbk}fork2,dilation:{xnsbk=min{xn1,xn,xn+1}fork=1xnsbk=min{xn1sbk,xn+1sbk}fork2, (7)

    where k=1,2,,km. Consequently, the morphological cover that surrounds the data graph can be obtained as

    Ck=Nn=1[(xnsbk)(xnsbk)]. (8)

    Here km is set according to the rule mentioned in [34]. Finally, the morphological FD of the time series can be acquired as the angular coefficient of linear regression of ln(Ck(2kN)2) vs. ln(12kN).

    Leibovich and Toth fractal dimension (LTD): LTD is the fast implementation of the box-counting algorithm [35]. The box-counting algorithm is based on the number of blocks to which the data graph can be split. So, the FD can be found as

    D=limϵ0log(nblocks(d))log(1d), (9)

    where d is the size of the blocks and D is the FD base on the original box-counting method. The LTD algorithm implements the box-counting method by withdrawing too small and too large blocks; therefore, it is faster than the original algorithm [36].

    Fractal Volatility (FV): FV computes the FD of a time series using the box-counting method with the rand-walk process [37]. In other words, it splits the data into blocks of size d by performing the random-walk process.

    Discrete Wavelet Transform (DWT): DWT describes any data by a set of weighted orthonormal wavelets called mother wavelets [38] according to the following relation

    Xm,n=Nn=1xn12mψ(2mtn), (10)

    where N is the total number of samples, Xm,n is the transformed data, ψ is the orthonormal wavelet. Moreover, m and n are two control parameters regarding the translation and dilation operation of the discrete wavelet and the data. DWT includes temporal information as well as frequency information in different scales. In this paper, the db4 was selected as the orthonormal wavelet for further analysis.

    Discrete Fourier Transform (DFT): DFT is the most fundamental transformation that describes the data based on some weighted sinusoidal functions. DFT can be defined

    Xk=Nn=1xnej2πNkn, (11)

    where k=1,2,,N and Xk is the transformes data for a specific frequency.

    Discrete Short-Time Fourier Transform (DSTFT): DSTFT is the time-frequency version of the DFT transform that contains only the frequency information. Thus, DSTFT can be helpful when DFT becomes insufficient for analysis, such as nonstationary signals [39]. DSTFT is defined as

    Xk=Nn=1xnω(nm)ej2πNkn, (12)

    where ω(L) is the temporal window of size L.

    Discrete Stockwell Transform (DST): DST is the extension of the DWT while having a close relationship with DSTFT. Also, it provides a frequency-dependent resolution since the sinusoidal functions are fixed in time, and a scalable Gaussian window operates the dilation and translation [40]. DST can be described as

    S(iTs,nNTs)=N1m=0H(m+nNTs)e2π2m2n2ej2πmiN (13)

    where Ts is the inverse sampling frequency, iTs defines the window τ, nNTs is the frequency domain, and H(n) is the DFT of the input data.

    Moments of distribution are basically statistical measures, the primary features that can be simply obtained from any data. The first-, second-, third-, and forth-order moments are called mean, variance, skewness, and kurtosis, which are described as follows

    M1=Nn=1xnN, (14)
    M2=Nn=1(xnM1)2N, (15)
    M3=Nn=1(xnμ1)3N.M322, (16)
    M4=Nn=1(xnμ1)4N.M22. (17)

    It should be noted that N is the number of data samples.

    Median is another statistical measure that can be helpful whenever the mean of that data is not a good measure of the distribution. The median can be defined as follows

    med={x(n2)n2k12(x(n12)+x(n+12))n2k1, (18)

    where k=1,2,,N. Maximum and minimum values of the samples within a specific period are the other two famous statistical measures used in this paper.

    In signal processing, feature selection is an optimization method leading to the optimum features for classification. Therefore, feature selection was used as a preprocessing step for machine-learning problems and is of particular importance when the data or extracted features are of high dimension [41]. In this subsection, three popular algorithms, namely, Genetic Algorithm (GA), Particle Swarm Optimization (PSO), and Ant Colony Optimization (ACO) algorithms, are briefly described.

    GA is an evolutionary algorithm that can be applied for search and optimization problems inspired by species' natural selection and evolution. In this algorithm, a population of chromosomes, which are, in fact, the solution to the problem, are directly selected for the next generation. Based on the selected species, the new generation, including a few numbers of new chromosomes, is also created using crossover and mutation techniques. This process is heuristically repeated until the solution converges to the optimal value [42]. In this article, five primary chromosomes and 100 repetitions with a crossover rate of 0.6 and a mutation rate of 0.001 were selected as the initial parameters of the GA algorithm for selecting the optimum features.

    PSO is an evolutionary computational algorithm that tries to find an optimized solution inspired by natural social behaviors, such as the group behavior observed in schools of fish and flocks of birds. According to this idea, some initial particles are localized in the search space, each having its own position and velocity. These particles move within the search space to find the best solution. In general, particles make their next moves based on their best-experienced position as well as the best-experienced position of the whole population of particles called a swarm. So, the movement of the particles in the search space is not dependent on the gradient, and therefore, it can be applied for differentiable and non-differentiable optimization problems [43]. In this paper, we set cognitive and social factors C1=C2=2 and inertia weight w=1 with five initial particles and 100 maximum iterations as the assumed parameters for the PSO algorithm used for obtaining the optimum features.

    ACO is a probabilistic and graph-based evolutionary algorithm that was first proposed based on the natural behavior of ants in hunting their food. In real life, ants leave pheromones on the pass to guide others to resources as they explore their surroundings, and thus, they can find the shortest pass to the food. Inspired by this cooperative-based technique, optimization problems can be solved and handled. First, some initial artificial ants are positioned in the parameter space that move to a solution stochastically. The pheromone trails, which specify the edges of the graph in the ACO algorithm, are obtained for each ant, and the best solution is selected. For the following steps, the edges of the graph become updated and guide the artificial ants toward the solution. This process is repeated iteratively until the solution converges to an optimal value [44]. In this paper, five initial artificial ants, 100 allowed iterations, α=τ=η=1, ρ=0.2, and β=0.1 are selected as the parameters needed for applying the ACO algorithm to find optimum features.

    Feature classification is the process of assigning or predicting the label of new data based on the trained model or information gained from the observed data [45]. The Support Vector Machine (SVM) and the K-Nearest Neighbor (KNN) classifiers are two supervised and most-used machine-learning algorithms for classification, which are briefly described below.

    The original SVM algorithm can be used for classifying two classes using a linear boundary; however, it has been extended for the classification of multi-class data. The SVM classifier builds an optimum hyperplane with the largest margin—with the maximum distance from the nearest data to the decision boundary—that can distinguish the data with the highest accuracy. If the data is not linearly distinguishable, using nonlinear kernels, the data will be implicitly mapped into the higher dimension wherein the data is distinguishable with a linear boundary [44]. This paper uses the SVM classifier with a three-order polynomial kernel function to classify the neuronal spiking activity in fixation and memory periods.

    The KNN algorithm classifies each new data based on voting between the class of k-closest observed data to the new sample. Therefore, to classify any new data, the KNN classifier needs to find the k-nearest training data by computing the distances of the new sample from all other samples in the parameter space. Therefore, although the KNN algorithm is simple to implement, it may be time-consuming due to its computational costs. This paper employs the KNN classifier with k=3 and standardized Euclidean distance to detect the presence of working memory using the firing rate data.

    To assess the classification performance, different criteria are introduced in the literature [16]. Accuracy, the most well-known evaluation criterion for classification, is defined based on the number of samples correctly labeled by the classifier versus the number of samples (see Eq (19)). Sensitivity and specificity are more specific assessment criteria since they respectively show the performance of the classifier in detecting and not detecting the target class. Therefore, sensitivity is defined as the number of samples correctly labeled as the target class by the classifiers versus the actual number of samples in the target class (see Eq (20)). On the other hand, specificity is defined as the number of samples correctly labeled as the non-target class by the classifiers versus the actual number of samples in the non-target class (see Eq (21)).

    Accuracy=tp+tntp+tn+fp+fn, (19)
    Sensitivity=tptp+fn, (20)
    Specificity=tntn+fp. (21)

    Here, tp, tn, fp, and fn refer to the true positive, true negative, false positive, and false negative indexes.

    Figure 2 shows the average normalized response of 131 MT neurons during the MGS task (see Methods). According to this figure, on average, no change in spiking activity is observed when comparing the response of neurons before and after the visual stimulus (i.e., fixation and memory periods, respectively). The inset bar graph in Figure 2 reveals no significant difference between the average spiking activity of MT neurons in memory vs. fixation periods in both IN and OUT memory conditions (pfixationINvs,memoryIN=0.385 and pfixationOUTvs,memoryOUT=0.385).

    Figure 2.  The normalized response of 131 MT neurons during the MGS task, including the fixation, visual, memory, and saccade periods. The average response in IN and OUT conditions are shown in green and pink, respectively. The inset bar graphs show the average firing rate of MT neurons in IN and OUT conditions during the fixation and memory periods.

    To detect the presence of the memory, we used the neural spiking activity in IN conditions during the fixation (when no working memory is involved) and memory (when deployment of top-down working memory signals is present) periods for feature extraction, selection, and classification. In the feature extraction step, 41 features are extracted from the IN conditions in fixation and memory periods. Therefore, the feature vector can be described as:

    - Index 1–6 are the fractal-based features, including HFD, KFD, GHE, MSFD, LTFD, and FV.

    - Index 7–34 are the transform-based features, including mean (index 7–10), variance (index 11–14), kurtosis (index 15–18), skewness (index 19–22), median (index 23–26), maximum (index 27–30), and minimum (index 31–34) of the components of DWT, DFT, DSTFT, and DST, respectively.

    - Index 35–41 are statistical features, including mean, variance, kurtosis, skewness, median, maximum, and minimum values of the firing rate signals within the fixation and memory periods.

    Finally, the selected features from the fixation IN and memory IN were classified using the SVM and KNN classifiers. Based on the selection method, classification was performed in four cases. To estimate the performance of the classifiers on the data, the k-fold cross-validation method (with k=10) and its balanced version called the A-test algorithm (with k=10 and ten iterations) were employed as the procedure for performing the classification. Unlike the K-fold cross-validation method, the A-test algorithm ensures that the same number of samples from each class is included in all training and testing folds. Therefore, the A-test method might be more reliable, especially when classes contain unequal data samples.

    In the first case, all extracted features, including 41 linear and nonlinear features, were used for the classification step. The results of the classification performance are shown in Figure 3. The details of the results can be found in Table 1.

    Figure 3.  The average performance of the classifiers SVM and KNN using the K-fold (with K=10) and A-test (with K=10 and 10 iterations) cross-validation methods. In this case, the classification was performed using all extracted features.
    Table 1.  Classification performance results (mean%±STD%) of fixation IN versus memory IN using all extracted features.
    Classifier Assessment criterion K-fold (K=10) A-test (K=10,iter=10)
    SVM Accuracy 98.85±2.60 98.77±0.16
    sensitivity 97.69±5.19 97.54±0.32
    specificity 100±0 100±0
    KNN Accuracy 98.46±1.99 97.92±0.41
    sensitivity 97.69±3.72 97±0.67
    specificity 99.23±2.43 98.85±0.54

     | Show Table
    DownLoad: CSV

    Figure 3 shows that the SVM and KNN classifiers SVM and KNN classifiers performed closely (AccuracySVM=98.85±2.6; AccuracyKNN=98.46±1.99) in distinguishing fixation IN and memory IN data. However, SVM performed slightly better than KNN (AccuracySVMAccuracyKNN<1%) in both cross-validation approaches. Table 1 reveals that the average accuracy is higher in the K-fold approach; however, the standard deviation mentioned in the A-test method is considerably lower. This shows that in different iterations, the average accuracy is not changed remarkably, and thus, the results are valid.

    Here, 20 features were selected using the GA feature selection method, including three fractal-based features (including the HFD, the KFD, and the FV), 14 transform-based features (including the mean of DFT, the mean of DST, the variance of DFT, the variance of DSTFT, the variance of DST, the kurtosis of DWT, the kurtosis of DSTFT, the skewness of DFT, the skewness of DSTFT, the median of DWT, the median of DFT, the median of DST, the maximum of DWT, and the minimum DWT), and three statistical features (including the skewness, the median, and the minimum values). Figure 4 and Table 2 demonstrate the classification results using the 20 GA-based selected features.

    Figure 4.  The average performance of the classifiers SVM and KNN using the K-fold (with K=10) and A-test (with K=10 and 10 iterations) cross-validation methods. In this case, the classification was performed using the feature selected by GA.
    Table 2.  Classification performance results (mean%±STD%) of fixation IN versus memory IN using the GA as the feature selection method.
    Classifier Assessment criterion K-fold (K=10) A-test (K=10,iter=10)
    SVM Accuracy 98.46±2.69 99.04±0.33
    sensitivity 96.92±5.38 98.08±0.65
    specificity 100±0 100±0
    KNN Accuracy 99.23±1.62 99.19±0.12
    sensitivity 98.46±3.24 98.38±0.24
    specificity 100±0 100±0

     | Show Table
    DownLoad: CSV

    According to Figure 4, unlike Figure 3, the KNN classifier performed slightly more effectively than the SVM classifier (AccuracyKNNAccuracySVM<1%). Table 2 shows that the GA-selected feature improved the classification performance, especially for the KNN classifier (AccuracyKNN=99.23±1.62). In addition, the KNN classifier has more reliable results since it generally has a lower standard deviation.

    Employing the PSO feature selection method, 17 features were selected, including three fractal-based features (including the HFD, the KFD, and the FV), 12 transform-based features (including the mean of DFT, the mean of DST, the variance of DFT, the variance of DSTFT, the kurtosis of DWT, the kurtosis of DSTFT, the skewness of DFT, the median of DWT, the median of DFT, the median of DST, the maximum of DWT, the minimum of DWT), and two statistical features (including the variance, and the maximum values). Similar to the previous subsections, Figure 5 and Table 3 illustrate the classification performance considering the PSO-based selected features.

    Figure 5.  The average performance of the classifiers SVM and KNN using the K-fold (with K=10) and A-test (with K=10 and 10 iterations) cross-validation methods. In this case, the classification was performed using the feature selected by PSO.
    Table 3.  Classification performance results (mean%±STD%) of fixation IN versus memory IN using the PSO as the feature selection method.
    Classifier Assessment criterion K-fold (K=10) A-test (K=10,iter=10)
    SVM Accuracy 99.23±2.43 99.50±0.26
    sensitivity 98.46±4.87 99.00±0.52
    specificity 100±0 100±0
    KNN Accuracy 98.85±1.86 98.77±0.24
    sensitivity 97.69±3.72 97.77±0.24
    specificity 100±0 99.77±0.37

     | Show Table
    DownLoad: CSV

    In contrast to Figure 3, in this case, Figure 5 shows that the PSO-selected features helped the SVM classifier improve its performance (AccuracySVM=99.50±0.26) more considerably than the KNN classifier (Accuracy=98.85±1.86). Besides, Table 3 reveals that the KNN classifier has more reliable results due to the lower standard deviation on average.

    Using the ACO algorithm for selecting the optimum features, a total of 20 features were selected, including three fractal-based features (including the HFD, the GHE, and the FV), 14 transform-based features (including the skewness of DSFTF, the variance of DWT, the kurtosis of DSTFT, the skewness of DWT, the kurtosis of DFT, the mean of DST, the skewness of DFT, the min of DFT, the mean of DWT, the variance of DSTFT, the variance of DST, the maximum of DFT), and five statistical features (including the skewness, the maximum, the variance, the kurtosis, and the minimum values). Figure 6, as well as Table 4, contains the results of performing the classification using the ACO-based selected features.

    Figure 6.  The average performance of the classifiers SVM and KNN using the K-fold (with K=10) and A-test (with K=10 and 10 iterations) cross-validation methods. In this case, the classification was performed using the feature selected by ACO.
    Table 4.  Classification performance results (mean%±STD%) of fixation IN versus memory IN using the ACO as the feature selection method.
    Classifier Assessment criterion K-fold (K=10) A-test (K=10,iter=10)
    SVM Accuracy 99.23±1.62 99.42±0.20
    sensitivity 98.46±3.24 98.85±0.41
    specificity 100±0 100±0
    KNN Accuracy 99.62±1.22 99.65±0.12
    sensitivity 99.23±2.43 99.31±0.24
    specificity 100±0 100±0

     | Show Table
    DownLoad: CSV

    Figure 6 illustrates that using the ACO-selected features, the average performance of both classifiers in detecting the presence of working memory is significantly enhanced compared to Figure 3, wherein all features were involved in the classification procedure (AccuracySVM=99.42±0.20; AccuracyKNN=99.65±0.12). From Table 4, it can be seen that the KNN classifier not only has slightly better performance (AccuracyKNNAccuracySVM<0.5%) but also, due to the lower standard deviation, has a more reliable performance.

    The brain is the most complex system in the human body. This complexity is reflected in the signals recorded from the brain. Thus, brain-associated data such as EEG or the spiking activity of neurons predominantly have nonlinear properties. This nonlinearity can be captured by FD, which is an index of complexity. To obtain FD of a time series, different algorithms have been proposed, such as HFD [30], KFD [31], GHE [32], MSFD [33], LTD [35], and FV [37]. However, transform-based features and statistical indexes are popular in signal processing. The main objective of this paper was to examine the ability of machine-learning methods to detect the presence of working memory using various linear and nonlinear features. Therefore, we used several algorithms to obtain the FD value (including HFD, KFD, GHE, MSFD, LTFD, and FV) and different transforms to obtain the frequency and/or time-frequency components (including statistical measures of DWT, DFT, DSTFT, and DST) of the average spiking activity of MT neurons. Also, we included some of the important statistical indexes (including mean, variance, kurtosis, skewness, median, maximum, and minimum values) in the feature set.

    Selecting the optimum features can be an optimization problem that, in most cases, results in improving the classifiers' performance. Such methods mainly focus on finding the best set of features that can lead to the best classification result. In this way, it can also help reduce the feature space's dimensionality and simplify the classification problem. Hence, after performing the classification with all extracted features, we examined whether the feature selection method could enhance the classification performance. Accordingly, we used GA, PSO, and ACO algorithms to select the optimum features for detecting the presence of working memory. In the classification step, the ability of two machine-learning algorithms, namely SVM and KNN classifiers, were employed to detect the presence of memory. KNN is typically considered a nonlinear classifier as it has a nonlinear decision boundary whilst SVM can be a nonlinear classifier if it uses a nonlinear kernel function. In general, the nonlinearity of the decision boundary enables a classifier to learn and distinguish the data classes more precisely and define the membership's probability to each data class for new data. For this reason, we used SVM with a three-order polynomial kernel function and KNN with three nearest neighbors. It should be noted that cross-validation methods can determine the validity of classification results, particularly when the number of samples is not too high. Therefore, K-fold (with 10 folds) and A-test (with 10 folds and 10 iterations) cross-validation methods were performed to show the results' robustness.

    The best classification performance of the SVM and KNN classifiers is summarized in Figure 7. According to this figure, when no selection algorithms were employed, the SVM classifiers led to the best performance (AccuracySVM=98.85±2.60). The same result can be seen in the case where the PSO algorithm was used to select the features (AccuracySVM=99.50±0.26). In contrast, when the GA- and ACO-selected features were used as the input of the classification algorithms, the KNN classifier reached higher average accuracy (AccuracyKNN=99.23±1.62 and 99.65±0.12, respectively). Moreover, Figure 7 shows that employing the selection method can improve the classification performance since, in all cases, the average accuracy was grown compared to the case where no selecting method was employed.

    Figure 7.  The best performance of KNN and SVM classifiers employing no selection algorithm as well as the GA, PSO, and ACO selection methods.

    In total, Figure 7 reveals that when the features were selected using the ACO algorithm, the KNN classifier could detect the presence of working memory with the accuracy of 99.65% and the standard deviation of 0.12 (using the A-test cross-validation method), which is the highest obtained average accuracy among the studied cases. In this case, three out of six FD-base features (including the HFD, the GHE, and the FV), 14 out of 28 transform-based features (including the skewness of DSFTF, the variance of DWT, the kurtosis of DSTFT, the skewness of DWT, the kurtosis of DFT, the mean of DST, the skewness of DFT, the min of DFT, the mean of DWT, the variance of DSTFT, the variance of DST, the maximum of DFT), and five out of seven statistical-based features (including the skewness, the maximum, the variance, the kurtosis, and the minimum values) were involved in the classification procedures.

    To obtain the neural code for spatial working memory represented in the firing rate of visual neurons, we compared the neural responses of MT neurons during the memory period (where the monkey is actively memorizing a location) with the neural activity during the fixation period (when no working memory is present). This approach could be questioned as one can argue that any differences between the neural responses during memory and fixation periods could occur due to other cognitive signals such as arousal or expectation than spatial working memory. Here we review a series of neurophysiological evidence revealing the dependence of these response changes (i.e., the differences of neural responses between memory and fixation period) on the content of working memory:

    - By measuring the neural responses of extrastriate neurons to visual probes presented during fixation and memory periods, a recent study showed a strong modulation of RF profile in V4 and MT neurons that were dependent on the content of working memory [19]. It was shown that during the maintenance of spatial information, only the neurons whose RFs during the fixation period were close to the remembered location, expanded and shifted their RFs towards that location during the memory period.

    - It was also found that the encoding of the visual probe's location by the population of MT neurons was enhanced during the memory period compared to the fixation period. In detail, this was measured by the ability of MT individual neurons' firing activity to discriminate two different visual probes (two-point discriminability). In fact, this memory-related enhancement in two-point discriminability only occurred to those visual probes that were presented near the locus of working memory.

    - At the level of LFP, it has been shown that the amount of information regarding the visual input conveyed by the alpha-beta phase of spike times increases during the memory period compared to the fixation period. This phenomenon only occurred to those visual probes which were presented near the remembered location.

    - Furthermore, it was shown that the discrimination between visual probes, based on the phase of each spike in the alpha-beta frequency range, is enhanced during the memory period compared to fixation. This discrimination enhancement was observed for the visual probes presented near the locus of working memory.

    As the differences between the neural responses during memory and fixation periods occur in a spatial-specific manner (i.e., near the locus of working memory), it would be very unlikely to relate these response changes to any other cognitive signals, such as arousal, than spatial working memory.

    Yaser Merrikhi collected this dataset in the laboratory of Dr. Behrad Noudoost at Montana State University, Bozeman, MT, USA. The lab's experiments were supported by MSU start-up fund, Whitehall 2014-5-18, NIH R01EY026924, and NSF143221 and 1632738 grants to Dr. Noudoost. We would like to thank Dr. Noudoost for sharing this dataset. This work is funded by the Centre for Nonlinear Systems, Chennai Institute of Technology, India, vide funding number CIT/CNS/2022/RP-006.

    The authors declare there is no conflict of interest.



    [1] A. A. Abdelhamid, E. Abdelhalim, M. A. Mohamed, F. Khalifa, Multi-classification of chest X-rays for COVID-19 diagnosis using deep learning algorithms, Appl. Sci., 12 (2022), 2080. https://doi.org/10.3390/app12042080 doi: 10.3390/app12042080
    [2] W. S. McCulloch, W. Pitts, A logical calculus of the ideas immanent in nervous activity, Bull. Math. Biophys., 5 (1943), 115–133.
    [3] Z. Li, F. Liu, W. Yang, S. Peng, J. Zhou, A survey of convolutional neural networks: Analysis, applications and prospects, IEEE Trans. Neural Netw. Learn Syst., 12 (2022), 6999–7019. https://doi.org/10.1109/TNNLS.2021.3084827 doi: 10.1109/TNNLS.2021.3084827
    [4] J. P. Cohen, L. Dao, K. Roth, P. Morrison, Y. Bengio, A. F. Abbasi, et al., Predicting COVID-19 pneumonia severity on chest X-ray with deep learning, Cureus, 12 (2020), e9448. https://doi.org/10.7759/cureus.9448 doi: 10.7759/cureus.9448
    [5] V. Ravi, H. Narasimhan, T. D. Pham, A cost‐sensitive deep learning‐based meta‐classifier for pediatric pneumonia classification using chest X‐rays, Expert Syst., (2020), e12966. https://doi.org/10.1111/exsy.12966 doi: 10.1111/exsy.12966
    [6] I. Borlea, R. Precup, A. Borlea, D. Iercan, A unified form of fuzzy C-means and K-means algorithms and its partitional implementation, Knowledge-Based Syst., 214 (2021), 106731. http://dx.doi.org/10.1016/j.knosys.2020.106731 doi: 10.1016/j.knosys.2020.106731
    [7] D. Varshni, K. Thakral, L. Agarwal, R. Nijhawan, A.Mittal, Pneumonia detection using CNN based feature extraction, in IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT), (2019), 1–7.
    [8] M. Taresh, N. Zhu, T. A. A. Ali, Transfer learning to detect COVID-19 automatically from X-ray images, using convolutional neural networks, Int. J. Biomed. Imaging, (2021), 8828404. https://doi.org/10.1155/2021/8828404 doi: 10.1155/2021/8828404
    [9] S. R. Velu, V. Ravi, K. Tabianan, Data mining in predicting liver patients using classification model, Health Technol. (Berl), 12 (2022), 1211–1235. https://doi.org/10.1007/s12553-022-00713-3 doi: 10.1007/s12553-022-00713-3
    [10] M. H. Alsharif, Y. H. Alsharif, K. Yahya, O. A. Alomari, M. A. Albreem, A. Jahid, Deep learning applications to combat the dissemination of COVID-19 disease: A review, Eur. Rev. Med. Pharmacol. Sci., 24 (2020), 11455–11460. https://doi.org/10.26355/eurrev_202011_23640 doi: 10.26355/eurrev_202011_23640
    [11] S. Sharma, Drawing insights from COVID-19-infected patients using CT scan images and machine learning techniques: A study on 200 patients, Environ. Sci. Pollut. Res., 27 (2020), 37155–37163. https://doi.org/10.1007/s11356-020-10133-3 doi: 10.1007/s11356-020-10133-3
    [12] A. Narin, C. Kaya, Z. Pamuk, Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks, Pattern Anal, Appl., 24 (2021), 1207–1220. https://doi.org/10.1007/s10044-021-00984-y doi: 10.1007/s10044-021-00984-y
    [13] H. Panwar, P. K. Gupta, M. K. Siddiqui, R. Morales-Menendez, V. Singh, Application of deep learning for fast detection of COVID-19 in X-Rays using nCOVnet, Chaos Solitons Fract., 138 (2020), 109944. https://doi.org/10.1016/j.chaos.2020.109944 doi: 10.1016/j.chaos.2020.109944
    [14] M. Singh, S. Bansal, S. Ahuja, R. K. Dubey, Panigrahi, B. K. Dey, Transfer learning–based ensemble support vector machine model for automated COVID-19 detection using lung computerized tomography scan data, Med. Biol. Eng. Comput., 59 (2021), 825–839. https://doi.org/10.1007/s11517-020-02299-2 doi: 10.1007/s11517-020-02299-2
    [15] A. M. Alqudah, S. Qazan, A. Alqudah, Automated systems for detection of COVID-19 using chest X-ray images and lightweight convolutional neural networks, Emerg. Radiol., 4 (2020). https://doi.org/10.1007/s13246-020-00865-4 doi: 10.1007/s13246-020-00865-4
    [16] I. D. Apostolopoulos, T. A. Mpesiana, COVID-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks, Phys. Eng. Sci. Med., 43 (2020), 635–640. https://doi.org/10.1016/j.eng.2020.04.010 doi: 10.1016/j.eng.2020.04.010
    [17] X. Xu, X.Jiang, C. Ma, P. Du, X. Li, S. Lv, et al., deep learning system to screen novel A Coronavirus Disease 2019 pneumonia, Engineering, 6 (2020), 1122–1129. https://doi.org/10.1016/j.eng.2020.04.010 doi: 10.1016/j.eng.2020.04.010
    [18] E. Hussain, M. Hasan, M. A. Rahman, I. Lee, T. Tamanna, M. Z. Parvez, CoroDet: A deep learning based classification for COVID-19 detection using chest X-ray images, Chaos Solitons Fract., 142 (2021), 110495. https://doi.org/10.1016/j.chaos.2020.110495 doi: 10.1016/j.chaos.2020.110495
    [19] S. Wang, B. Kang, J. Ma, X. Zeng, M. Xiao, J. Guo, et al., A deep learning algorithm using CT images to screen for Corona Virus Disease (COVID-19), Eur Radiol., 31 (2021), 6096–6104. https://doi.org/10.1007/s00330-021-07715-1 doi: 10.1007/s00330-021-07715-1
    [20] L. L, L. Qin, Z.Xu, Y. Yin, X. Wang, B. Kong, et al., Artificial intelligence distinguishes COVID-19 from community acquired pneumonia on chest CT, Radiology, 296 (2020). https://doi.org/10.1148/radiol.2020200905 doi: 10.1148/radiol.2020200905
    [21] A. N. J Raj, H. Zhu, A. Khan, Z. Zhuang, Z. Yang, V. G. V. Mahesh, et al., ADID-UNET—a segmentation model for COVID-19 infection from lung CT scans, PeerJ Comput. Sci., 7 (2021), e349. https://doi.org/10.7717/PEERJ-CS.349 doi: 10.7717/PEERJ-CS.349
    [22] H. Khalid, M. Hussain, M. A. Al Ghamdi, T. Khalid, K. Khalid, M. A. Khan, et al., A comparative systematic literature review on knee bone reports from MRI, X-rays and CT scans using deep learning and machine learning methodologies, Diagnostics, 10 (2020), 518. https://doi.org/10.3390/diagnostics10080518 doi: 10.3390/diagnostics10080518
    [23] G. Puneet, Pneumonia detection using convolutional neural networks, Int. J. Sci. Technol. Res., 7 (2021), 77–80. https://doi.org/10.46501/ijmtst070117 doi: 10.46501/ijmtst070117
    [24] X. Ding, Y. Guo, G. Ding, J. Han, Acnet: Strengthening the kernel skeletons for powerful CNN via asymmetric convolution blocks, in IEEE/CVF international conference on computer vision (ICCV), (2019), pp. 1911–1920. http://dx.doi.org/10.1109/ICCV.2019.00200
    [25] S. Kostadinov, What is deep transfer learning and why is it becoming so popular? Towards Data Science, (2019).
    [26] M. Lascu, Deep learning in classification of Covid-19 coronavirus, pneumonia and healthy lungs on CXR and CT images, J. Med. Biol. Eng., 41 (2021), 514–522. http://dx.doi.org/10.1007/s40846-021-00630-2 doi: 10.1007/s40846-021-00630-2
    [27] X. Ma, B. Zheng, Y. Zhu, F. Yu, R. Zhang, B. Chen, Covid-19 lesion discrimination and localization network based on multi-receptive field attention module on CT images, Optik, 241 (2021), 167100. http://dx.doi.org/10.1016/j.ijleo.2021.167100 doi: 10.1016/j.ijleo.2021.167100
    [28] R. Kundu, R. Das, Z. W. Geem, G. T. Han, R. Sarkar, Pneumonia detection in chest X-ray images using an ensemble of deep learning models, PLoS One, 16 (2021), e0256630. https://doi.org/10.1371/journal.pone.0256630 doi: 10.1371/journal.pone.0256630
  • This article has been cited by:

    1. Feng Qiu, Hui Xu, Fukui Li, Applying modified golden jackal optimization to intrusion detection for Software-Defined Networking, 2023, 32, 2688-1594, 418, 10.3934/era.2024021
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2687) PDF downloads(91) Cited by(2)

Figures and Tables

Figures(21)  /  Tables(3)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog