
Decision-making in a railway station regarding the compatibility of the positions of the switches of the turnouts and the indications (proceed/stop) of the railway colour light signals is a safety-critical issue that is considered very labor-intensive. Different authors have proposed alternative solutions to automate its supervision, which is performed by the so-called railway interlocking systems. The classic railway interlocking systems are route-based and their compatibility is predetermined (usually by human experts): only some chosen routes are simultaneously allowed. Some modern railway interlocking systems are geographical and make decisions on the fly, but are unsuitable if the station is very large and the number of trains is high. In this paper, we present a completely new algebraic model for decision-making in railway interlocking systems, based on other computer algebra techniques, that bypasses the disadvantages of the approaches mentioned above (its performance does not depend on the number of trains in the railway station and can be used in large railway stations). The main goal of this work is to provide a mathematical solution to the interlocking problems. We prove that our approach solves it in linear time. Although our approach is interesting from a theoretical perspective, it has a significant limitation: it can hardly be adopted in an actual interlocking implementation, mainly due to the heavy certification requirements for this kind of safety-critical application. Nevertheless, the results may be useful for simulations that do not require certification credit.
Citation: Antonio Hernando, Eugenio Roanes-Lozano, José Luis Galán-García, Gabriel Aguilera-Venegas. Decision making in railway interlocking systems based on calculating the remainder of dividing a polynomial by a set of polynomials[J]. Electronic Research Archive, 2023, 31(10): 6160-6196. doi: 10.3934/era.2023313
[1] | Darrak Moin Quddusi, Sandesh Athni Hiremath, Naim Bajcinca . Mutation prediction in the SARS-CoV-2 genome using attention-based neural machine translation. Mathematical Biosciences and Engineering, 2024, 21(5): 5996-6018. doi: 10.3934/mbe.2024264 |
[2] | Wenhan Guo, Yixin Xie, Alan E Lopez-Hernandez, Shengjie Sun, Lin Li . Electrostatic features for nucleocapsid proteins of SARS-CoV and SARS-CoV-2. Mathematical Biosciences and Engineering, 2021, 18(3): 2372-2383. doi: 10.3934/mbe.2021120 |
[3] | A. D. Al Agha, A. M. Elaiw . Global dynamics of SARS-CoV-2/malaria model with antibody immune response. Mathematical Biosciences and Engineering, 2022, 19(8): 8380-8410. doi: 10.3934/mbe.2022390 |
[4] | Akansha Singh, Krishna Kant Singh, Michal Greguš, Ivan Izonin . CNGOD-An improved convolution neural network with grasshopper optimization for detection of COVID-19. Mathematical Biosciences and Engineering, 2022, 19(12): 12518-12531. doi: 10.3934/mbe.2022584 |
[5] | A. M. Elaiw, Raghad S. Alsulami, A. D. Hobiny . Global dynamics of IAV/SARS-CoV-2 coinfection model with eclipse phase and antibody immunity. Mathematical Biosciences and Engineering, 2023, 20(2): 3873-3917. doi: 10.3934/mbe.2023182 |
[6] | Lin Feng, Ziren Chen, Harold A. Lay Jr., Khaled Furati, Abdul Khaliq . Data driven time-varying SEIR-LSTM/GRU algorithms to track the spread of COVID-19. Mathematical Biosciences and Engineering, 2022, 19(9): 8935-8962. doi: 10.3934/mbe.2022415 |
[7] | Ze-Yu Zhang, Zhu-Jun Mao, Ye-ping Ruan, Xin Zhang . Computational identification of Shenshao Ningxin Yin as an effective treatment for novel coronavirus infection (COVID-19) with myocarditis. Mathematical Biosciences and Engineering, 2022, 19(6): 5772-5792. doi: 10.3934/mbe.2022270 |
[8] | Xianfang Wang, Qimeng Li, Yifeng Liu, Zhiyong Du, Ruixia Jin . Drug repositioning of COVID-19 based on mixed graph network and ion channel. Mathematical Biosciences and Engineering, 2022, 19(4): 3269-3284. doi: 10.3934/mbe.2022151 |
[9] | Qinyan shen, Jiang wang, Liangying zhao . To investigate the internal association between SARS-CoV-2 infections and cancer through bioinformatics. Mathematical Biosciences and Engineering, 2022, 19(11): 11172-11194. doi: 10.3934/mbe.2022521 |
[10] | Sakorn Mekruksavanich, Wikanda Phaphan, Anuchit Jitpattanakul . Epileptic seizure detection in EEG signals via an enhanced hybrid CNN with an integrated attention mechanism. Mathematical Biosciences and Engineering, 2025, 22(1): 73-105. doi: 10.3934/mbe.2025004 |
Decision-making in a railway station regarding the compatibility of the positions of the switches of the turnouts and the indications (proceed/stop) of the railway colour light signals is a safety-critical issue that is considered very labor-intensive. Different authors have proposed alternative solutions to automate its supervision, which is performed by the so-called railway interlocking systems. The classic railway interlocking systems are route-based and their compatibility is predetermined (usually by human experts): only some chosen routes are simultaneously allowed. Some modern railway interlocking systems are geographical and make decisions on the fly, but are unsuitable if the station is very large and the number of trains is high. In this paper, we present a completely new algebraic model for decision-making in railway interlocking systems, based on other computer algebra techniques, that bypasses the disadvantages of the approaches mentioned above (its performance does not depend on the number of trains in the railway station and can be used in large railway stations). The main goal of this work is to provide a mathematical solution to the interlocking problems. We prove that our approach solves it in linear time. Although our approach is interesting from a theoretical perspective, it has a significant limitation: it can hardly be adopted in an actual interlocking implementation, mainly due to the heavy certification requirements for this kind of safety-critical application. Nevertheless, the results may be useful for simulations that do not require certification credit.
In December 2019, a new, human-infecting SARS-Coronavirus-2, Coronavirus was recognized in Wuhan, China [1,2]. It was conveyed that the virus transferred between humans by droplets or close contact [3,4,5]. As of March 2020, the new SARS-Coronavirus-2 has above than 98,000 cases through 88 countries beyond China [6].
The virus is a pathogenic human coronavirus under the Beta coronavirus genus. The other two pathogenic species, including Severe Acute Respiratory Syndrome Coronavirus (SARS-Coronavirus) and the Middle East Respiratory Syndrome Coronavirus (MERS-Coronavirus) were outbreaks in China and the Middle East in 2002 and 2012 respectively [1]. On January 10, the full genome sequence of this massive RNA virus (SARS-Coronavirus-2) was first uncovered in China's laboratory [5,7] and placed in the NCBI GENBAND. Coronaviruses have enveloped viruses that contain a positive single-stranded ribonucleic acid (RNA) virus that affects humans and animals without segmentation [5,8]. The genomes of the Coronaviruses are formed of base pairs ranging from 26 kilobase-pairs (kbps) to 31 kbps, with GC contents fluctuating from 43 to 32%, and human-infecting coronaviruses including MERS-Coronavirus, HCoV-OC43, SARS-Coronavirus, HCoronavirus-NL63, HCoronavirus-229E, and HCoronavirus-HKU1 [9].
Coronaviruses are characterized by the ability to quickly evolve and adapt to various epidemiological conditions [10]. Each replication cycle of Coronavirus presents new genetic mutations, and its normal developmental rate is nearly from 4 to 10 nucleotide substitutions per site/ year [1]. During genomic information reproduction, SARS-Coronavirus-2 evolves [11]. The mutation happens because of certain errors when copying RNA to a new cell. SARS-Coronavirus-2 analyses can produce false-positive results if they are not targeted especially to SARS-Coronavirus-2, as the virus is difficult to distinguish from other Coronavirus organisms because of their genetic similarity. Therefore, it is essential to use improved diagnostic tools in order to correctly classify SARS Coronavirus-2 from other Coronaviruses.
Currently, Molecular approaches are commonly used to classify pathogens, such as quantitative real-time RT-PCR and nucleic acid sequencing methods [12]. However, owing to their comparatively new evolution and still not entirely understood characteristics, they have an overall unsatisfactory identification rate for this particular virus. Since coronaviruses are genetic diseases, viral sequencing techniques have been employed to detect the virus depending on its genome sequence and therefore avoiding the downsides of the classical diagnostic methods.
Classification using viral sequencing methods depends on strategies of alignment such as FASTA [13] and Basic Local Alignment Search Tool (BLAST) [14] algorithms. These methods count on the assumption that DNA sequences have common features, and their order predominates among various sequences [15,16]. However, there are limitations to these methods, which include the need for base sequences for the detection [17]. In addition, because viruses have a high level of mutation and most genomes do not have proper reference sequences, next-generation sequencing (NGS) genomic samples might not be identified by BLAST [18]. This method has also a negative impact of neglecting part of the vital information contained in the input sequence if it cannot completely fill a DNA sequence of fixed size.
Traditional machine learning approach using genomic signal processing technique was introduced to discern between Covid-19 and other Coronaviruses [19,20] by using their genomic sequences reported in the NCBI GENBAND in order to increase the accuracy of disease detection in lesser time. Some characteristics were then extracted, such as Discrete Fourier transformation, Discrete Cosine transformation, and incorporated into a classifier. However, this method needs to extract pre-selected features in order to recognize or classify the viral DNA sequences.
The deep learning approach has been recently introduced as various alternatives for DNA sequence recognition, as these strategies do not require pre-selected features to recognize or classify DNA sequences. Deep learning approaches [21,22] have evolved rapidly since they can be employed in the processing of large-scale genetic data, especially in the field of bioinformatics. Today, the size of these datasets exceeds the size of 10 million [23]. Studies in this area are focused on the analysis and classification of DNA and RNA sequences.
Deep learning to use Convolution Neural Networks (CNN) [24,25,26] have been efficiently used for the classification of DNA sequences. However, these methods encode the RNA sequences data into an input layer of the CNN therefore, numeric values have to be assigned to the different bases. In addition, CNN learns to recognize patterns across space while RNN helps in resolving temporal data problems.
Other alternative approaches focusing on deep learning have also been explored by slicing sequences into bits of fixed lengths, from 300 [16] to 3000 bps [23]. However, these methods ignore part of the data contained in the input sequence if it does not fill a piece of fixed size.
Because of the difficulties of differentiating between the SARS-Coronavirus-2 virus and other Coronavirus organisms (SARS-Coronavirus and MERS-Coronavirus) and respiratory infections because of their genetic similarity; it was aimed to develop an advanced system to correctly classify the virus based on its genomic sequence. Apart from the classical approaches, the proposed Deep Learning Bidirectional Recurrent Neural Networks (BRNNs) approach eliminates the need for pre-selected features to identify or classify the viral DNA sequences and also works on the full DNA input sequence as a whole thereby overcoming the problem of ignoring any data in the input sequence. It has also the advantages of rapid and high accuracy of disease detection and classification.
The rest of this article is organized as follows. The contribution of this research is presented in the next section. A background overview of BRNNs with different types of cell units is introduced in Section 4. The methodology, preparation of the data, and proposed algorithms are demonstrated in Section 5. Experimental results and a discussion of the results are presented in Section 6. Finally, concluding remarks are given in Section 7.
In this study, Recurrent Neural Networks (RNN) based models with different cell units were implemented for the purpose of classification of the SARS-Coronavirus-2 virus from other Coronavirus and respiratory infections based on its DNA sequence. Apart from the classical approaches, the proposed Deep Learning Bidirectional Recurrent Neural Networks (BRNNs) approach eliminates the need for pre-selected features and also works on the full DNA input sequence as a whole thereby overcoming the problem of ignoring any data in the input sequence. A comparison between the presented deep network models based on LSTM and GRU unit cells was performed to attain the most efficient model for virus classification. Also, an investigation of the hyper-parameters that worked best with the BRNN models was presented to select the highest accuracy classification model.
In recent times, the RNN model has been an extremely favored architecture [27], especially for sequential data. The RNN structure is demonstrated in Figure 1. Each node at a time step comprises an input from the preceding node and uses a feedback loop to continue. By employing the current input and previous hidden state, each node produces a current hidden state and output as follows:
ht=f(Whhht−1+Wxhxt+bh) | (1) |
ot=f(Whyht+bo) | (2) |
where x is the input, y is the output sequence, h is the hidden vector sequence at each time Step (t). W denotes the weight matrices, and b represents the bias for hidden and output states. f is the activation function used at the hidden layers.
The shared downside of the classical RNN models is that as the time step increases, the network fails to draw context from time steps of previous states far behind, such a phenomenon is known as long-term dependence. Moreover, exploding and vanishing gradient problems are often found quite commonly due to the deep layers of a network and repetitive behavior of the standard RNN,
To resolve this problem, LSTM models were implemented by positioning memory cells with multiple gates in a hidden layer [28]. Figure 2 displays a hidden layer block with an LSTM cell unit, and the three functions of gate controllers are demonstrated as follows:
■ Forget gate (ft) (should be omitted): It chooses which fragment of the long-term state ct should be ignored.
■ Input gate (it): It controls which part of c should be added to the long-term statect.
■ Output gate (gt): It defines which fragment of c should be fed to htandot.
The following equations demonstrate the cell's long-term and short-term states and each layer's output in time steps.
ft=σ(WTxf⋅xt+WThf⋅ht−1+bf) | (3) |
it=σ(WTxi⋅xt+WThi⋅ht−1+bi) | (4) |
ot=σ(WTxo⋅xt+WTho⋅ht−1+bi) | (5) |
gt=tanh(WTxg⋅xt+WThg⋅ht−1+bi) | (6) |
ct=ft⊗ct−1+it⊗˜ct | (7) |
(ot,ht)=gt⊗tanh(ct) | (8) |
where Wxf,Wxi,Wxo,Wxg represent the weight matrices for the corresponding associated input vector and Whf,Whi,Who,Whg characterize the weight matrices of the short-term state of the previous time step, and bf,bi,bo,bg are bias.
GRUs presented in 2014 resemble LSTMs but they have fewer parameters [11]. Figure 3 shows the block of a hidden layer with a GRU cell unit. They also have gated units as LSTMs that regulate the flow of context within the device without having separate memory cells. GRU doesn't have an output barrier, unlike LSTM, thereby revealing its full content. Mathematically, GRU formulations are expressed by the following equations:
rt=σ(WTxr⋅xt+WTor⋅ot−1+br) | (9) |
zt=σ(WTxz⋅xt+WToz⋅ot−1+bz) | (10) |
˜ot=tanh(WTx˜o⋅xt+WTo˜o⋅(rt⊗ot−1)+b˜o) | (11) |
ot=zt⊗ot−1+(1−zt)⊗˜ot | (12) |
where Wxr,Wxz,Wx˜o indicate the weight matrices for the consistently associated input vector and Wor,Woz,Wo characterize the weight matrices of the preceding time step, br,bz,b are biasing.
Bidirectional Recurrent Neural Networks (BNN) is a version of the RNN that Schuster and Paliwal suggested in 1997 [17]. Standard RNN's use the previous context information, but Bidirectional RNNs explore also the future context. BRNN is a better alternative than plain RNN since the input sequence of bidirectional RNN processes has two RNN paths, one responsible for the forward states and the other for the backward states. Hence, Bidirectional RNN effectively increases the amount of information available to the network, improving the content available to the algorithm.
The DNA sequence information was derived from the Novel Coronavirus Resource 2019 repository (2019nCoVR) [29]. All the accessible sequences with the Query Nucleotide Completeness = "Complete", Sequence Quality = "High", Host = "Homo Sapiens", DNA Sequence lengths = (29,800–29,900 bps) nucleotides for each epidemic, and Data Source = "GenBank" for a total of 2500 complete genome samples. 500 random complete genomes for each epidemic (COVID-19, MERSA–CORONAVIRUS-2, Common cold, ARI, and SARS) were chosen.
To preprocess the data, all repeated sequences and sequences having a length that is out of the identified range (29,800–29,900 bps) were removed. Then the data was labelled and prearranged, as shown in Table 1. The data were grouped into 5 classes based on the disease type as follows:
Virus name | Disease | Label |
SARS Coronavirus -2 | COVID 19 | 0 |
MERS- Coronavirus | MERSA – COV-2 | 1 |
Human coronavirus -OC43 | Common cold | 2 |
Human coronavirus -229E | ||
Human coronavirus -4408 | ||
Human coronavirus -NL63 | ARI | 3 |
Human coronavirus -HKU1 | ||
SARS Coronavirus | SARS | 4 |
SARS Coronavirus - P2 | ||
SARS Coronavirus HKU-39849 | ||
SARS Coronavirus GDH-BJH01 |
■ First Class "COVID-19": Includes the genome virus [SARS-CoV-2]
■ Second Class "MERSA–CORONAVIRUS-2": Includes the genome virus [MERS-CoV]
■ Third Class "Common Cold": Includes the genome viruses [HCoV-OC43/HCoV-229E] which are placed in one category, as they are coronaviruses accountable for normal cold [30], and [HCoV-4408] was also added because it is a Beta Coronavirus as HCoV-OC43.
■ Fourth Class "ARI": Includes the genome viruses [HCoV-NL63/ HCoV-HKU1] in the same group as they are both accompanied by acute breathing infections [31].
■ Fifth Class "SARS": Includes the genome viruses [SARS-Coronavirus-P2/SARS-Coronavirus/SARS-Coronavirus GDH-BJH01/SARS-Coronavirus HKU-39849 [32]] which are categorized in the same class because they are all strains of SARS.
The samples were divided into two parts, 80% for training, 20% for testing, in tenfold cross-validation. The steps of implementation were summarized in Figure 4.
In this section, the proposed BRNN classifier model architectures were performed and shown in Figure 5. The proposed model depends on RNN with enhanced cell units, LSTM and GRU, in their hidden states, while the training procedure is applied in a bidirectional manner. The model comprises five layers that are described as follows:
Layer 1 (Input sequence): After selecting and preprocessing the DNA sequences as described in Section 5.1, it was fed into the presented RNN models.
Layer 2 (k-mers model): The DNA sequence S was split into overlapping k-meters of length k (k = 4 is employed in this work) by the use of a sliding window with stride s, the size of the window is the current k-mer. As illustrated in Figure 6. For DNA sequence S, the windows are located covering the first k bases, and it moves one right character at a time to generate the next k-mer. Given a sequence of length n,S=(S1,S2,…,Sn)whereSi∈{A,C,G,T}, S is converted into ˜n=(n−khigh+1) k-mer's numbers.
f(S)=(S1:k1,S2:2+k2,…S˜n:˜n+k˜n) | (13) |
It was observed that k-mer embedding for representing input sequences has several advantages over one-hot encoding, which was reported in previous works. Firstly, it improves the model performance, secondly; it reduces computational time and space required to execute models as compared to one-hot encoding because the word vector in the one-hot method may be a high dimensional vector since it encodes each word in a large number of text data.
Layer 3 (Word2vec algorithm): By using the word2vec algorithm [33], each k-mer is mapped into a d-dimensional vector space. Word2vec can utilize either Continuous Bag-Of-Words (CBOW) or Continuous Skip-Gram (CSG) to produce a distributed representation of words. A continuous bag-of-words (CBOW) was employed for all our experiments because the CBOW is faster to train than skip-gram and CBOW works well with a large amount of training data.
Layer 4 (Bidirectional RNN): It represents the suggested models based on Bidirectional RNN, where LSTM or GRU cells are used as the hidden blocks. The forward track traces the data section from left to right, while in both BLSTM and BGRU, the backward track traces the input from right to left. The forward recurrent sequence and backward hidden sequence can be formulated as follows:
→ht=f(Wx,→hxt+W→h,→h→ht−1+b→h) | (14) |
←ht=f(Wx,←hxt+W←h,←h←ht+1+b←h) | (15) |
yt=(W→h,y→ht+W←h,y←ht+by) | (16) |
where xt is the input feature vector,→ht(←ht) is the activation vector on the forward (backward) hidden layer, Wp,q is the weight matrix,br is the bias term,f(⋅) is the activation function on each node in the hidden layers, and y is the posterior probability vector of the output label.
Layer 5 (Output layer): A regular sigmoid function that is applied on the output layer in order to measure the expected probability of characters for each step of t and k in the alphabet. This performance is shown as follows:
yt=sigmoid(W→h,y→ht+W←h,y←ht+by) | (17) |
While training a model, it is important to know which hyper-parameters to select. Hyper-parameters refer to the parameters that are set before the start of Bidirectional RNN training. The time required to train and test a model is dependent on the types of hyper-parameters selected. The tune-ability of an algorithm, hyper-parameters, or interacting hyper-parameters is a measure of how much performance can be gained by tuning it. In this research, an investigation of the hyper-parameters that work best on the RNN models was performed in order to attain the highest accuracy for DNA sequence classification. The hyper-parameters that were focused on in this work are the optimizer type and the number of unit cells.
Four different types of gradient descent algorithm were employed on both the BRNN models based on LSTM and GRU cells including SGD, Adam, Adagrad, and RMSprop optimizers and three different unit numbers of cells ((50, 80 and 100). The performance of different parameter settings was recorded and compared to get the optimum parameter settings.
Gradient descent (GD) is a class of algorithms that aims to find the minimum point on a function by following the gradient. Vanilla gradient descent just follows the gradient (scaled by the learning rate). Two common tools to improve gradient descent are the sum of gradient (first moment) and the sum of the gradient squared (second moment). The Momentum method uses the first moment with a decay rate to gain speed. AdaGrad uses the second moment with no decay to deal with sparse features. RMSProp uses the second moment with a decay rate to speed up from AdaGrad. Adam uses both first and second moments and is generally the best choice.
In our research, didn't use a Particle Swarm Optimization (PSO) because a global optimization method PSO does not use the gradient of the problem being optimized, which means PSO was less effective compared to GD in our application. In addition to that, we analyzed the effect of different embedding dimensions (numbers of cells), including 150 and 200. The model complexity will be increased by more weight parameters, which are caused by a larger d and need to be learned in the embedding layer.
The performance of the proposed deep learning BRNN approach for SARS Coronavirus-2 virus classification was evaluated with the following classification performance measures which include accuracy, precision, recall, specificity, and F1-score and the corresponding confusion matrix was further developed. Confusion matrices represent a very important metric in machine learning evaluation since it provides further information on relations between classes and types of classification errors. The x-axis in the confusion matrix shows the predicted output (disease class) of the model and the y-axis shows the true class of the disease. The diagonal elements of the matrix represent the true positive values of the model, meaning all the correct classifications performed by the model and all other entries in the matrix show misclassifications. Off-diagonal elements for the row represent the probability that the disease of a certain class will be mis-classified as another class. The definition of the performance metrics are clarified as follows:
Accuracy: Represents the performance accuracy of the model
Accuracy = TP(NumberofTruePositives)+TN(Numberofthetruenegatives)Allpredictionsperformedbythemodel | (18) |
Precision: It identifies the proportion of correct positive classifications.
Precision = TPTP+FP | (19) |
Recall (Sensitivity): It measures the proportion of actual positives correctly classified by the model.
Recall = TPTP+FN | (20) |
Specificity: It identifies the proportion of correct negative classifications.
Specificity = TNTN+FP | (21) |
F1-score: It indicates the perfect balance between precision and recall values. It can be calculated as follows
F1 score = 2Precision∗RecallPrecisio+Recall | (22) |
Another significant metric is also assessed which is the area under the Receiver Operating Characteristic (ROC) curve. The Receiver Operating Characteristic (ROC) curve shows a relation between the True Positive Rate (TPR) and False Positive Rate (FPR) of the classification. The FPR is defined by:
FPR = FPFP+TN | (23) |
Therefore, the ROC Curve Region (AUC) tests the efficiency of all potential classification thresholds in a given model and therefore shows the accuracy of the findings independently of that model.
In the first part of this section, experiments were conducted on two proposed BRNN models depending on two different types of cell units (GRU) and (LSTM) in order to attain a high-efficiency classification model based on viral DNA sequence. The RNN models were implemented using four different optimizers (SGD, Adam, Adagrad, and RMSprop) and the number of unit cells was altered between random values of 50, 80 and 100 to obtain the most suitable optimizer and the optimum number of cells that will lead to the highest classification efficiency.
A comparative study was performed and shown in Table 2 showing the classification accuracies of both models employing the four different types of optimizers and changing the unit number of cells in each optimizer. It was clear from the table that the highest accuracies were obtained when applying the SGD optimizer and the number of unit cells of 80 in both models. The results also revealed that the GRU BRNN model has achieved higher efficiencies than its counterpart model as it achieved an overall accuracy of 96.8 % while the LSTM RNN model achieved 95.8 %. This result can be attributed to the fact that the GRU cell in the BRNNs deploys a reset gate and an update gate in a hidden layer which is computationally more efficient than a usual LSTM network.
Optimizer Type | Units number | LSTM BRNN Model (%) | GRU BRNN Model (%) |
SGD_50 | 50 | 94.52 | 95.89 |
Adam_50 | 50 | 95.89 | 95.53 |
Adagrad_50 | 50 | 95.29 | 95.32 |
RMSprop_50 | 50 | 95.21 | 95.43 |
SGD_80 | 80 | 95.3 | 96.88 |
Adam_80 | 80 | 94.37 | 96.4 |
Adagrad_80 | 80 | 94.42 | 96.17 |
RMSprop_80 | 80 | 94.22 | 96.58 |
SGD_100 | 100 | 94.14 | 95.51 |
Adam_100 | 100 | 94.52 | 95.32 |
Adagrad_100 | 100 | 94.32 | 95.22 |
RMSprop_100 | 100 | 94.34 | 95.66 |
Furthermore, a plot was drawn and shown in Figure 7. In order to visualize the overall performance of the two models on all optimizers and applying the different number of unit cells. It was obvious from the plot that the performance of both models was nearly the same when applying a unit number of cells of 50. The highest classification accuracy achieved by the LSTM RNN model was 95.8% when applying the ADAM optimizer with a LSTM number of 50 and the SGD optimizer with a cell number of 80. However, the higher classification ability of the GRU was seen when applying a cell number of 80 and 100 for all optimizers reaching the highest value of 96.8% with the SGD optimizer and a GRU number of 80. The accuracy of both models didn't increase by increasing the number of unit cells concluding that the optimum number of cells was 80 and employing the SGD optimizer to achieve an efficient classification.
To evaluate the effectiveness of the implemented BRNN models employing the best hyper-parameters (SGD optimizer with a cell number of 80) and its ability to discriminate between SARS-Coronavirus-2 and other viruses, the confusion matrices for the two models and the ROC curves were developed and shown in Figures 8 and 9. The resulting confusion matrix of the LSTM BRNN model reveals that 77 out of the 83 SARS-Coronavirus-2 sequences were correctly categorized, while the other 6 sequences were incorrectly classified in other groups. In the other hand, the GRU BRNN model correctly identified 37 of the 39 SARS-Coronavirus-2 sequences, while the remaining two were misclassified. By comparing the two matrices, it can be seen that the GRU has fewer false-positive and fewer false negative values than the LSTM, which reveals the higher accuracy of prediction of the GRU model. Another measure that shows successful classification is the area under ROC curve (AUC value). Generally, ROC curve is defined as a probability curve in classification studies and the AUC values express the degree of class separability within these possibilities. The higher the AUC value, the higher the classification predictions of the employed model. As showed from Figures 8 and 9 that the average AUC value (0.98) of the GRU model was higher than that of the LSTM (0.96) which also confirms the results of the present study that proved the higher ability of classification of the GRU as compared to the LSTM model. These promising results can help in the efficient detection and identification of the SARS Coronavirus-2 and discriminate it from other Coronaviruses with high accuracy results which can further help contain its spread.
Furthermore, Tables 3 and 4 show the classification performance metrics for all five classes of disease. It was clear from the tables that the accuracy of prediction of the presented models was higher than 97% for all classes of viruses, showing the efficient performance of both models. However, when comparing the ability of both models to classify class 0 (SARS-Coronavirus-2 virus), it was seen that the GRU achieved a higher classification performance than the LSTM since it attained an accuracy of 98.4% and precision, Recall, and F1 Score values of 0.95 which were higher than the LSTM performance values. Hence, it can be concluded that the presented GRU BRNN model has a better prediction ability for COVID 19, thus providing an efficient tool to help doctors achieve better clinical decisions with high precision.
Classes | Accuracy (%) | Precision | Recall | F1 Score |
0 | 97.6 | 0.93 | 0.93 | 0.93 |
1 | 98.4 | 0.94 | 0.98 | 0.96 |
2 | 98.2 | 0.96 | 0.95 | 0.95 |
3 | 97.4 | 0.91 | 0.96 | 0.93 |
4 | 97.2 | 0.97 | 0.91 | 0.94 |
Classes | Accuracy (%) | Precision | Recall | F1 score |
0 | 98.4 | 0.95 | 0.95 | 0.95 |
1 | 98.4 | 0.98 | 0.94 | 0.96 |
2 | 98.4 | 0.95 | 0.98 | 0.96 |
3 | 98.8 | 0.98 | 0.96 | 0.97 |
4 | 98.8 | 0.97 | 0.98 | 0.97 |
Developing efficient diagnostic machine learning tools that target the genome in order to identify different classes of Corona Virus can help reduce the death rates and the time to identify the infection for proper clinical decisions. A Deep Learning Approach employing Bidirectional Recurrent Neural Networks (BRNN) based models was implemented to classify the SARS-Coronavirus-2 virus from other Coronaviruses and other respiratory infections based on its DNA sequence. It was aimed to propose a system with high-accuracy results that can classify the type of disease based on genome sequence leading to a rapid diagnosis of COVID-19, avoiding the disadvantages of the traditional diagnostic methods. An evaluation of the BRNN models depending on Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) cells was investigated to determine the most effective model for SARS Coronavirus-2 sequence classification. A study of the hyper-parameters that work best on the BRNN models was also performed, including the optimizer type and the number of unit cells. The results revealed that the best hyper-parameters producing the highest performance for both models were got when applying the SGD optimizer and an optimum number of unit cells of 80 in both models. Moreover, the GRU BRNN model was able to discriminate between SARS Coronavirus-2 and other viruses with a higher overall classification accuracy of 96.8% as compared to that of the LSTM BRNN model having a 95.8% overall classification accuracy. When comparing the ability of both models to classify all classes of viruses, the accuracy of prediction were all higher than 97% for all classes, indicating the efficient performance of both models. However, when comparing the ability of both models to classify class 0 (SARS-Coronavirus-2 virus), it was seen that the GRU achieved a higher classification performance than the LSTM since it attained higher accuracies, precision, recall, and F1 score values. Hence, it was concluded that the proposed GRU BRNN model has a better prediction ability for COVID-19, thus providing an efficient tool to help achieve better clinical decisions with high precision.
All authors declare that there is no conflicts of interest in this paper.
[1] |
N. Bešinović, Resilience in railway transport systems: a literature review and research agenda, Transport Rev., 40 (2020), 457–478. https://doi.org/10.1080/01441647.2020.1728419 doi: 10.1080/01441647.2020.1728419
![]() |
[2] |
Y. Zhou, J. Wang, H. Yang, Resilience of transportation systems: concepts and comprehensive review, IEEE Trans. Intell. Transp. Syst., 20 (2019), 4262–4276. https://doi.org/10.1109/TITS.2018.2883766 doi: 10.1109/TITS.2018.2883766
![]() |
[3] | J. Pachl, Railway signalling principles, 2021. Available from: http://www.joernpachl.de/rsp.htm. |
[4] |
H. Lian, X. Wang, A. Sharma, M. A. Shah, Application and study of artificial intelligence in railway signal interlocking fault, Informatica, 46 (2022), 343–354. https://doi.org/10.31449/inf.v46i3.3961 doi: 10.31449/inf.v46i3.3961
![]() |
[5] | A. Fantechi, G. Gori, A. E. Haxthausen, C. Limbrée, Compositional verification of railway interlockings: comparison of two methods, in Reliability, Safety, and Security of Railway Systems. Modelling, Analysis, Verification, and Certification, (2022), 3–19. https://doi.org/10.1007/978-3-031-05814-1_1 |
[6] | G. Lukács, T. Bartha, Conception of a formal model-based methodology to support railway engineers in the specification and verification of interlocking systems, in 2022 IEEE 16th International Symposium on Applied Computational Intelligence and Informatics (SACI), (2022), 000283–000288. https://doi.org/10.1109/SACI55618.2022.9919532 |
[7] | K. Winter, W. Johnston, P. Robinson, P. Strooper, L. van den Berg, Tool support for checking railway interlocking designs, in 10th Australian Workshop on Safety Re-lated Programmable Systems (SCS'05), Australian Computer Society, Sydney, 55 (2006), 101–107. Available from: https://crpit.scem.westernsydney.edu.au/confpapers/CRPITV55Winter.pdf. |
[8] |
A. Borälv, Case study: formal verification of a computerized railway interlocking, Formal Aspects Comput., 10 (1998), 338–360. https://doi.org/10.1007/s001650050021 doi: 10.1007/s001650050021
![]() |
[9] | Anonymous, Proyecto y obra del enclavamiento electrónico de la estación de Madrid-Atocha, Proyecto Técnico, Siemens, Madrid, 1988. |
[10] | Anonymous, Microcomputer interlocking hilversum, Siemens, Munich, 1986. |
[11] | Anonymous, Microcomputer interlocking rotterdam, Siemens, Munich, 1989. |
[12] | Anonymous, Puesto de enclavamiento con microcomputadoras de la estación de Chiasso de los SBB, Siemens, Munich, 1989. |
[13] | L. Villamandos, Sistema informático concebido por Renfe para diseñar los enclavamientos, Vía Libre, 348 (1993), 65. |
[14] |
J. Peleska, N. Krafczyk, A. E. Haxthausen, R. Pinger, Efficient data validation for geographical interlocking systems, Formal Aspects Comput., 33 (2021), 925–955. https://doi.org/10.1007/s00165-021-00551-6 doi: 10.1007/s00165-021-00551-6
![]() |
[15] | A. Fantechi, A. E. Haxthausen, M. B. R. Nielsen, Model checking geographically distributed interlocking systems using UMC, in 2017 25th Euromicro International Conference on Parallel, Distributed and Network-based Processing (PDP), (2017), 278–286. https://doi.org/10.1109/PDP.2017.66 |
[16] |
A. Hernando, E. Roanes-Lozano, An algebraic model for implementing expert systems based on the knowledge of different experts, Math. Comput. Simul., 107 (2015), 92–107. https://doi.org/10.1016/j.matcom.2014.05.003 doi: 10.1016/j.matcom.2014.05.003
![]() |
[17] |
E. Roanes-Lozano, E. Roanes-Macías, L. Laita, Railway interlocking systems and Gröbner bases, Math. Comput. Simul., 51 (2000), 473–481. https://doi.org/10.1016/S0378-4754(99)00137-8 doi: 10.1016/S0378-4754(99)00137-8
![]() |
[18] |
A. Hernando, E. Roanes-Lozano, R. Maestre-Martínez, J. Tejedor, A logic-algebraic approach to decision taking in a railway interlocking system, Ann. Math. Artif. Intell., 65 (2012), 317–328. https://doi.org/10.1007/s10472-012-9321-y doi: 10.1007/s10472-012-9321-y
![]() |
[19] |
A. Hernando, R. Maestre, E. Roanes-Lozano, A new algebraic approach to decision making in a railway interlocking system based on preprocess, Math. Probl. Eng., 2018 (2018), 4982974. https://doi.org/10.1155/2018/4982974 doi: 10.1155/2018/4982974
![]() |
[20] | D. Bjørner, The FMERail/TRain Annotated Rail Bibliography, 1999. Available from: http://www2.imm.dtu.dk/db/fmerail/fmerail/. |
[21] | M. J. Morley, Modelling British Rail's Interlocking Logic: Geographic Data Correctness, LFCS, Department of Computer Science, University of Edinburgh, 1991. |
[22] | K. Nakamatsu, Y. Kiuchi, A. Suzuki, EVALPSN based railway interlocking simulator, in Knowledge-Based Intelligent Information and Engineering Systems, Berlin - Heidelberg, (2004), 961–967. https://doi.org/10.1007/978-3-540-30133-2_127 |
[23] | A. Janota, Using Z specification for railway interlocking safety, Period. Polytech. Transp. Eng., 28 (2000), 39–53. Available from: https://pp.bme.hu/tr/article/view/1963. |
[24] | K. M. Hansen, Formalising railway interlocking systems, in Nordic Seminar on Dependable Computing Systems, Lyngby, (1994), 83–94. |
[25] | M. Montigel, Modellierung und Gewährleistung von Abhängigkeiten in Eisenbahnsicherungsanlagen, Ph.D thesis, ETH Zurich, Zurich, 1994. Available from: http://www.inf.ethz.ch/research/disstechreps/theses. |
[26] | X. Chen, H. Huang, Y. He, Automatic generation of relay logic for interlocking system based on statecharts, in 2010 Second World Congress on Software Engineering, (2010), 183–188. https://doi.org/10.1109/WCSE.2010.31 |
[27] |
X. Chen, Y. He, H. Huang, An approach to automatic development of interlocking logic based on Statechart, Enterp. Inf. Syst., 5 (2011), 273–286. https://doi.org/10.1080/17517575.2011.575475 doi: 10.1080/17517575.2011.575475
![]() |
[28] |
X. Chen, Y. He, H. Huang, A component-based topology model for railway interlocking systems, Math. Comput. Simul., 81 (2011), 1892–1900. https://doi.org/10.1016/j.matcom.2011.02.007 doi: 10.1016/j.matcom.2011.02.007
![]() |
[29] |
B. Luteberget, C. Johansen, Efficient verification of railway infrastructure designs against standard regulations, Formal Methods Syst. Des., 52 (2018), 1–32. https://doi.org/10.1007/s10703-017-0281-z doi: 10.1007/s10703-017-0281-z
![]() |
[30] |
M. Bosschaart, E. Quaglietta, B. Janssen, R. M. P. Goverde, Efficient formalization of railway interlocking data in RailML, Inf. Syst., 49 (2015), 126–141. https://doi.org/10.1016/j.is.2014.11.007 doi: 10.1016/j.is.2014.11.007
![]() |
[31] |
E. Roanes-Lozano, L. M. Laita, An applicable topology-independent model for railway interlocking systems, Math. Comput. Simul., 45 (1998), 175–183. https://doi.org/10.1016/S0378-4754(97)00093-1 doi: 10.1016/S0378-4754(97)00093-1
![]() |
[32] | E. Roanes-Lozano, L. M. Laita, E. Roanes-Macías, An application of an AI methodology to railway interlocking systems using computer algebra, in Tasks and Methods in Applied Artificial Intelligence, (1998), 687–696. https://doi.org/10.1007/3-540-64574-8_455 |
[33] |
B. Buchberger, An algorithm for finding the basis elements of the residue class ring of a zero dimensional polynomial ideal, J. Symb. Comput., 41 (2006), 475–511. https://doi.org/10.1016/j.jsc.2005.09.007 doi: 10.1016/j.jsc.2005.09.007
![]() |
[34] | D. Cox, J. Little, D. O'Shea, Ideals, Varieties and Algorithms, Springer, 2015. Available from: https://link.springer.com/book/10.1007/978-3-319-16721-3. |
[35] |
E. Roanes-Lozano, A. Hernando, J. A. Alonso, L. M. Laita, A logic approach to decision taking in a railway interlocking system using Maple, Math. Comput. Simul., 82 (2011), 15–28. https://doi.org/10.1016/j.matcom.2010.05.024 doi: 10.1016/j.matcom.2010.05.024
![]() |
[36] |
M. Davis, G. Logemann, D. Loveland, A machine program for theorem-proving, Commun. ACM, 5 (1962), 394–397. https://doi.org/10.1145/368273.368557 doi: 10.1145/368273.368557
![]() |
[37] |
E. Roanes-Lozano, J. A. Alonso, A. Hernando, An approach from answer set programming to decision making in a railway interlocking system, Rev. R. Acad. Cienc. Exactas, Fis. Nat. Ser. A Mat., 108 (2014), 973–987. https://doi.org/10.1007/s13398-013-0155-1 doi: 10.1007/s13398-013-0155-1
![]() |
[38] | J. Abbott, A. M. Bigatti, CoCoALib: a C++ library for doing Computations in Commutative Algebra, 2019. Available from: http://cocoa.dima.unige.it/cocoalib. |
[39] | J. Abbott, A. M. Bigatti, CoCoA: a system for doing Computations in Commutative Algebra, 2010. Available from: http://cocoa.dima.unige.it. |
[40] | Ministerio de Transportes, Movilidad y Agenda Urbana, Estudio Informativo del Nuevo Complejo Ferroviario de la Estación de Madrid-Chamartín. Available from: https://www.mitma.gob.es/ferrocarriles/estudios-en-tramite/estudios-y-proyectos-entramite/chamartin. |
[41] | Anonympus, Adjudicada la Modificación de Las Instalaciones de Seguridad, ERTMS, Comunicaciones y Energía de Madrid-Chamartín, Vía Libre, 2021. Available from: https://www.vialibre-ffe.com/noticias.asp?not = 33047 & cs = infr. |
[42] |
G. Aguilera-Venegas, J. L. Galán-García, E. Mérida-Casermeiro, P. Rodríguez-Cielos, An accelerated-time simulation of baggage traffic in an airport terminal, Math. Comput. Simul., 104 (2014), 58–66. https://doi.org/10.1016/j.matcom.2013.12.009 doi: 10.1016/j.matcom.2013.12.009
![]() |
[43] |
G. Aguilera, J. L. Galán, J. M. García, E. Mérida, P. Rodríguez, An accelerated-time simulation of car traffic on a motorway using a CAS, Math. Comput. Simul., 104 (2014), 21–30. https://doi.org/10.1016/j.matcom.2012.03.010 doi: 10.1016/j.matcom.2012.03.010
![]() |
[44] |
J. L. Galán-García, G. Aguilera-Venegas, M. Á. Galán-García, P. Rodríguez-Cielos, A new Probabilistic Extension of Dijkstra's Algorithm to simulate more realistic traffic flow in a smart city, Appl. Math. Comput., 267 (2015), 780–789. https://doi.org/10.1016/j.amc.2014.11.076 doi: 10.1016/j.amc.2014.11.076
![]() |
[45] |
J. L. Galán-García, G. Aguilera-Venegas, P. Rodríguez-Cielos, An accelerated-time simulation for traffic flow in a smart city, J. Comput. Appl. Math., 270 (2014), 557–563. https://doi.org/10.1016/j.cam.2013.11.020 doi: 10.1016/j.cam.2013.11.020
![]() |
[46] | A. Iliasov, D. Taylor, L. Laibinis, A. B. Romanovsky, SafeCap: from formal verification of railway interlocking to its certification, 2021. |
1. | Rania E. Hammam, Hani Attar, Ayman Amer, Haitham Issa, Ioannis Vourganas, Ahmed Solyman, P. Venu, Mohammad R. Khosravi, Mohanad A. Deif, Xin Ning, Prediction of Wear Rates of UHMWPE Bearing in Hip Joint Prosthesis with Support Vector Model and Grey Wolf Optimization, 2022, 2022, 1530-8677, 1, 10.1155/2022/6548800 | |
2. | Nadiah Baghdadi, Ahmed S. Maklad, Amer Malki, Mohanad A. Deif, Reliable Sarcoidosis Detection Using Chest X-rays with EfficientNets and Stain-Normalization Techniques, 2022, 22, 1424-8220, 3846, 10.3390/s22103846 | |
3. | Mohanad A. Deif, Hani Attar, Ayman Amer, Ismail A. Elhaty, Mohammad R. Khosravi, Ahmed A. A. Solyman, Akshi Kumar, Diagnosis of Oral Squamous Cell Carcinoma Using Deep Neural Networks and Binary Particle Swarm Optimization on Histopathological Images: An AIoMT Approach, 2022, 2022, 1687-5273, 1, 10.1155/2022/6364102 | |
4. | Marion Adebiyi, Miracle Nmesomachi Enwere, Abenego Shekari, Ayodele Adebiyi, Francis B. Osang, 2023, Chapter 59, 978-981-19-7659-9, 653, 10.1007/978-981-19-7660-5_59 | |
5. | Hamed Dashti, Iman Dehzangi, Masroor Bayati, James Breen, Amin Beheshti, Nigel Lovell, Hamid R. Rabiee, Hamid Alinejad-Rokny, Integrative analysis of mutated genes and mutational processes reveals novel mutational biomarkers in colorectal cancer, 2022, 23, 1471-2105, 10.1186/s12859-022-04652-8 | |
6. | Rania E. Hammam, Ahmed A. A. Solyman, Mohammed H. Alsharif, Peerapong Uthansakul, Mohanad A. Deif, Design of Biodegradable Mg Alloy for Abdominal Aortic Aneurysm Repair (AAAR) Using ANFIS Regression Model, 2022, 10, 2169-3536, 28579, 10.1109/ACCESS.2022.3155645 | |
7. | Yuanyuan Pei, Yi Guo, Tong Wu, Huiying Liang, Quantifying the dynamic transmission of COVID-19 asymptomatic and symptomatic infections: Evidence from four Chinese regions, 2022, 10, 2296-2565, 10.3389/fpubh.2022.925492 | |
8. | Jiangsheng Pi, Peishun Jiao, Yang Zhang, Junyi Li, MDGNN: Microbial Drug Prediction Based on Heterogeneous Multi-Attention Graph Neural Network, 2022, 13, 1664-302X, 10.3389/fmicb.2022.819046 | |
9. | Ruiyi Liu, Yongle Zhan, Xuan Liu, Yifang Zhang, Luting Gui, Yimin Qu, Hairong Nan, Yu Jiang, Mehdi Gheisari, Stacking Ensemble Method for Gestational Diabetes Mellitus Prediction in Chinese Pregnant Women: A Prospective Cohort Study, 2022, 2022, 2040-2309, 1, 10.1155/2022/8948082 | |
10. | Ahmad Alqerem, Hani Attar, Waleed Alomoush, Mohand Deif, 2022, The Ability of Ultra Wideband to Differentiate Between Hematoma and Tumor Occur in The Brain, 979-8-3503-3274-2, 1, 10.1109/EICEEAI56378.2022.10050471 | |
11. | Bakr Ahmed Taha, Yousif Al Mashhadany, Abdulmajeed H. J. Al-Jumaily, Mohd Saiful Dzulkefly Bin Zan, Norhana Arsad, SARS-CoV-2 Morphometry Analysis and Prediction of Real Virus Levels Based on Full Recurrent Neural Network Using TEM Images, 2022, 14, 1999-4915, 2386, 10.3390/v14112386 | |
12. | Marion Adebiyi, Miracle Nmesomachi Enwere, Timothy Adeliyi, Abiodun Okunola, Ayodele Adebiyi, 2023, Chapter 56, 978-981-19-7659-9, 627, 10.1007/978-981-19-7660-5_56 | |
13. | Mohamed Kareem Al-Thiabi, Ali J Dawood Al-Alwani, 2022, The Prediction of COVID-19 Virus Mutation Using Long Short-Term Memory, 979-8-3503-3486-9, 113, 10.1109/ICCITM56309.2022.10032012 | |
14. | Mohanad A. Deif, Ahmed A. A. Solyman, Mohammed H. Alsharif, Seungwon Jung, Eenjun Hwang, A Hybrid Multi-Objective Optimizer-Based SVM Model for Enhancing Numerical Weather Prediction: A Study for the Seoul Metropolitan Area, 2021, 14, 2071-1050, 296, 10.3390/su14010296 | |
15. | Zuher Jahshan, Esteban Garzón, Leonid Yavits, ViRAL: Vision Transformer Based Accelerator for ReAL Time Lineage Assignment of Viral Pathogens, 2024, 12, 2169-3536, 28353, 10.1109/ACCESS.2024.3367801 | |
16. | Qais Ibrahim Ahmed, Hani Attar, Ayman Amer, Mohanad A. Deif, Ahmed A. A. Solyman, Development of a Hybrid Support Vector Machine with Grey Wolf Optimization Algorithm for Detection of the Solar Power Plants Anomalies, 2023, 11, 2079-8954, 237, 10.3390/systems11050237 | |
17. | Talal Ali Ahmed Abdullah, Mohd Soperi Mohd Zahid, Ahmad F. Turki, Waleed Ali, Ahmad A. Jiman, Mohammed J. Abdulaal, Nebras M. Sobahi, Eyad T. Attar, Sig-Lime: A Signal-Based Enhancement of Lime Explanation Technique, 2024, 12, 2169-3536, 52641, 10.1109/ACCESS.2024.3384277 | |
18. | Fatma Refaat Ahmed, Samira Ahmed Alsenany, Sally Mohammed Farghaly Abdelaliem, Mohanad A. Deif, Development of a hybrid LSTM with chimp optimization algorithm for the pressure ventilator prediction, 2023, 13, 2045-2322, 10.1038/s41598-023-47837-8 | |
19. | Narendra N Khanna, Manasvi Singh, Mahesh Maindarkar, Ashish Kumar, Amer M. Johri, Laura Mentella, John R Laird, Kosmas I. Paraskevas, Zoltan Ruzsa, Narpinder Singh, Mannudeep K. Kalra, Jose Fernandes E. Fernandes, Seemant Chaturvedi, Andrew Nicolaides, Vijay Rathore, Inder Singh, Jagjit S. Teji, Mostafa Al-Maini, Esma R. Isenovic, Vijay Viswanathan, Puneet Khanna, Mostafa M. Fouda, Luca Saba, Jasjit S. Suri, Polygenic Risk Score for Cardiovascular Diseases in Artificial Intelligence Paradigm: A Review, 2023, 38, 1011-8934, 10.3346/jkms.2023.38.e395 | |
20. | Shikha Sharma, Rinkle Rani, Nidhi Kalra, Genomic Characterization of Emerging SARS-CoV-2: A Systematic Review, 2023, 18, 15748936, 375, 10.2174/1574893618666230228115423 | |
21. | Mohammed Kareem Al-Thiabi, Ali J. Dawood Al-Alwani, 2024, 3009, 0094-243X, 020017, 10.1063/5.0190461 | |
22. | Mohanad A. Dief, Waleed Alomoush, Hani Atta, Osama A. Khashan, Ahmed Solyman, Rania ELGohary, 2024, PID Controller Tuning Using Multi-Objective Ant Colony Optimization for Blood Glucose Level of a Diabetic Patient, 979-8-3503-9496-2, 1, 10.1109/ICCR61006.2024.10532878 | |
23. | Matthieu Vilain, Stéphane Aris-Brosou, Machine Learning Algorithms Associate Case Numbers with SARS-CoV-2 Variants Rather Than with Impactful Mutations, 2023, 15, 1999-4915, 1226, 10.3390/v15061226 |
Virus name | Disease | Label |
SARS Coronavirus -2 | COVID 19 | 0 |
MERS- Coronavirus | MERSA – COV-2 | 1 |
Human coronavirus -OC43 | Common cold | 2 |
Human coronavirus -229E | ||
Human coronavirus -4408 | ||
Human coronavirus -NL63 | ARI | 3 |
Human coronavirus -HKU1 | ||
SARS Coronavirus | SARS | 4 |
SARS Coronavirus - P2 | ||
SARS Coronavirus HKU-39849 | ||
SARS Coronavirus GDH-BJH01 |
Optimizer Type | Units number | LSTM BRNN Model (%) | GRU BRNN Model (%) |
SGD_50 | 50 | 94.52 | 95.89 |
Adam_50 | 50 | 95.89 | 95.53 |
Adagrad_50 | 50 | 95.29 | 95.32 |
RMSprop_50 | 50 | 95.21 | 95.43 |
SGD_80 | 80 | 95.3 | 96.88 |
Adam_80 | 80 | 94.37 | 96.4 |
Adagrad_80 | 80 | 94.42 | 96.17 |
RMSprop_80 | 80 | 94.22 | 96.58 |
SGD_100 | 100 | 94.14 | 95.51 |
Adam_100 | 100 | 94.52 | 95.32 |
Adagrad_100 | 100 | 94.32 | 95.22 |
RMSprop_100 | 100 | 94.34 | 95.66 |
Classes | Accuracy (%) | Precision | Recall | F1 Score |
0 | 97.6 | 0.93 | 0.93 | 0.93 |
1 | 98.4 | 0.94 | 0.98 | 0.96 |
2 | 98.2 | 0.96 | 0.95 | 0.95 |
3 | 97.4 | 0.91 | 0.96 | 0.93 |
4 | 97.2 | 0.97 | 0.91 | 0.94 |
Classes | Accuracy (%) | Precision | Recall | F1 score |
0 | 98.4 | 0.95 | 0.95 | 0.95 |
1 | 98.4 | 0.98 | 0.94 | 0.96 |
2 | 98.4 | 0.95 | 0.98 | 0.96 |
3 | 98.8 | 0.98 | 0.96 | 0.97 |
4 | 98.8 | 0.97 | 0.98 | 0.97 |
Virus name | Disease | Label |
SARS Coronavirus -2 | COVID 19 | 0 |
MERS- Coronavirus | MERSA – COV-2 | 1 |
Human coronavirus -OC43 | Common cold | 2 |
Human coronavirus -229E | ||
Human coronavirus -4408 | ||
Human coronavirus -NL63 | ARI | 3 |
Human coronavirus -HKU1 | ||
SARS Coronavirus | SARS | 4 |
SARS Coronavirus - P2 | ||
SARS Coronavirus HKU-39849 | ||
SARS Coronavirus GDH-BJH01 |
Optimizer Type | Units number | LSTM BRNN Model (%) | GRU BRNN Model (%) |
SGD_50 | 50 | 94.52 | 95.89 |
Adam_50 | 50 | 95.89 | 95.53 |
Adagrad_50 | 50 | 95.29 | 95.32 |
RMSprop_50 | 50 | 95.21 | 95.43 |
SGD_80 | 80 | 95.3 | 96.88 |
Adam_80 | 80 | 94.37 | 96.4 |
Adagrad_80 | 80 | 94.42 | 96.17 |
RMSprop_80 | 80 | 94.22 | 96.58 |
SGD_100 | 100 | 94.14 | 95.51 |
Adam_100 | 100 | 94.52 | 95.32 |
Adagrad_100 | 100 | 94.32 | 95.22 |
RMSprop_100 | 100 | 94.34 | 95.66 |
Classes | Accuracy (%) | Precision | Recall | F1 Score |
0 | 97.6 | 0.93 | 0.93 | 0.93 |
1 | 98.4 | 0.94 | 0.98 | 0.96 |
2 | 98.2 | 0.96 | 0.95 | 0.95 |
3 | 97.4 | 0.91 | 0.96 | 0.93 |
4 | 97.2 | 0.97 | 0.91 | 0.94 |
Classes | Accuracy (%) | Precision | Recall | F1 score |
0 | 98.4 | 0.95 | 0.95 | 0.95 |
1 | 98.4 | 0.98 | 0.94 | 0.96 |
2 | 98.4 | 0.95 | 0.98 | 0.96 |
3 | 98.8 | 0.98 | 0.96 | 0.97 |
4 | 98.8 | 0.97 | 0.98 | 0.97 |