
Severe acute respiratory syndrome coronavirus 2 (SARS‑CoV‑2) has been evolving rapidly after causing havoc worldwide in 2020. Since then, it has been very hard to contain the virus owing to its frequently mutating nature. Changes in its genome lead to viral evolution, rendering it more resistant to existing vaccines and drugs. Predicting viral mutations beforehand will help in gearing up against more infectious and virulent versions of the virus in turn decreasing the damage caused by them. In this paper, we have proposed different NMT (neural machine translation) architectures based on RNNs (recurrent neural networks) to predict mutations in the SARS-CoV-2-selected non-structural proteins (NSP), i.e., NSP1, NSP3, NSP5, NSP8, NSP9, NSP13, and NSP15. First, we created and pre-processed the pairs of sequences from two languages using k-means clustering and nearest neighbors for training a neural translation machine. We also provided insights for training NMTs on long biological sequences. In addition, we evaluated and benchmarked our models to demonstrate their efficiency and reliability.
Citation: Darrak Moin Quddusi, Sandesh Athni Hiremath, Naim Bajcinca. Mutation prediction in the SARS-CoV-2 genome using attention-based neural machine translation[J]. Mathematical Biosciences and Engineering, 2024, 21(5): 5996-6018. doi: 10.3934/mbe.2024264
[1] | Mohanad A. Deif, Ahmed A. A. Solyman, Mehrdad Ahmadi Kamarposhti, Shahab S. Band, Rania E. Hammam . A deep bidirectional recurrent neural network for identification of SARS-CoV-2 from viral genome sequences. Mathematical Biosciences and Engineering, 2021, 18(6): 8933-8950. doi: 10.3934/mbe.2021440 |
[2] | Chongyi Tian, Longlong Lin, Yi Yan, Ruiqi Wang, Fan Wang, Qingqing Chi . Photovoltaic power prediction based on dilated causal convolutional network and stacked LSTM. Mathematical Biosciences and Engineering, 2024, 21(1): 1167-1185. doi: 10.3934/mbe.2024049 |
[3] | Wenbo Yang, Wei Liu, Qun Gao . Prediction of dissolved oxygen concentration in aquaculture based on attention mechanism and combined neural network. Mathematical Biosciences and Engineering, 2023, 20(1): 998-1017. doi: 10.3934/mbe.2023046 |
[4] | Tao Yang, Qicheng Yang, Yibo Zhou, Chuanbiao Wen . Glucose trend prediction model based on improved wavelet transform and gated recurrent unit. Mathematical Biosciences and Engineering, 2023, 20(9): 17037-17056. doi: 10.3934/mbe.2023760 |
[5] | Xianfang Wang, Qimeng Li, Yifeng Liu, Zhiyong Du, Ruixia Jin . Drug repositioning of COVID-19 based on mixed graph network and ion channel. Mathematical Biosciences and Engineering, 2022, 19(4): 3269-3284. doi: 10.3934/mbe.2022151 |
[6] | Natalya Shakhovska, Vitaliy Yakovyna, Valentyna Chopyak . A new hybrid ensemble machine-learning model for severity risk assessment and post-COVID prediction system. Mathematical Biosciences and Engineering, 2022, 19(6): 6102-6123. doi: 10.3934/mbe.2022285 |
[7] | Zhijun Yang, Wenjie Huang, Hongwei Ding, Zheng Guan, Zongshan Wang . Performance analysis of a two-level polling control system based on LSTM and attention mechanism for wireless sensor networks. Mathematical Biosciences and Engineering, 2023, 20(11): 20155-20187. doi: 10.3934/mbe.2023893 |
[8] | Wenhan Guo, Yixin Xie, Alan E Lopez-Hernandez, Shengjie Sun, Lin Li . Electrostatic features for nucleocapsid proteins of SARS-CoV and SARS-CoV-2. Mathematical Biosciences and Engineering, 2021, 18(3): 2372-2383. doi: 10.3934/mbe.2021120 |
[9] | Lin Feng, Ziren Chen, Harold A. Lay Jr., Khaled Furati, Abdul Khaliq . Data driven time-varying SEIR-LSTM/GRU algorithms to track the spread of COVID-19. Mathematical Biosciences and Engineering, 2022, 19(9): 8935-8962. doi: 10.3934/mbe.2022415 |
[10] | Giuseppe Ciaburro . Machine fault detection methods based on machine learning algorithms: A review. Mathematical Biosciences and Engineering, 2022, 19(11): 11453-11490. doi: 10.3934/mbe.2022534 |
Severe acute respiratory syndrome coronavirus 2 (SARS‑CoV‑2) has been evolving rapidly after causing havoc worldwide in 2020. Since then, it has been very hard to contain the virus owing to its frequently mutating nature. Changes in its genome lead to viral evolution, rendering it more resistant to existing vaccines and drugs. Predicting viral mutations beforehand will help in gearing up against more infectious and virulent versions of the virus in turn decreasing the damage caused by them. In this paper, we have proposed different NMT (neural machine translation) architectures based on RNNs (recurrent neural networks) to predict mutations in the SARS-CoV-2-selected non-structural proteins (NSP), i.e., NSP1, NSP3, NSP5, NSP8, NSP9, NSP13, and NSP15. First, we created and pre-processed the pairs of sequences from two languages using k-means clustering and nearest neighbors for training a neural translation machine. We also provided insights for training NMTs on long biological sequences. In addition, we evaluated and benchmarked our models to demonstrate their efficiency and reliability.
COVID-19 is known as the biggest pandemic of the 21st century caused by SARS-CoV-2 (severe acute respiratory syndrome coronavirus 2) which is responsible for more than 6.9 million deaths and more than 770 million infections worldwide since its onset [1]. The first-ever case was reported in November 2019 and, for almost three years, the world has suffered a lot because it is hard to contain the virus due to its frequently occurring mutations and complex mechanism of action.
Viruses are known as obligate intracellular parasites that neither belong to living nor non-living organisms due to their unique course of reproduction. They behave like living beings and reproduce only when they are inside their host but remain dormant otherwise like non-living beings. SARS-CoV-2 belongs to the category of ss(+)RNA (single-stranded) viruses having a size of 29.9 Kb [2], its genome contains 12 functional open reading frames (ORFs) constituting ∼30,000 nucleotide base pairs, 38% GC content with 11 protein-coding regions and 12 expressed proteins. It has four major structural proteins, i.e., spike (S), envelope (E), membrane (M), and nucleocapsid (N), appearing in an order of 5′ to 3′; considered as the foremost vaccine and drug targets. In the genome, ORF1a and ORF1ab hold the viral nucleotide content; ORF1a along with ORF1b also encode polyproteins pp1a and pp1ab upon a ribosomal frame-shift between them. These polyproteins are then processed by viral proteinases to produce 16 non-structural proteins which have been found well-conserved in the coronavirus family [3].
The probability of a genetic change being passed to the next generation is known as the mutation rate in organisms; whereas, in viruses, one generation refers to the host cell infection cycle which comprises attachment, penetration, uncoating, gene expression, replication, encapsidation, and release [4]. Apart from replication, mutations can also occur from genome editing or rapid RNA/DNA damage and when these changes are left uncorrected these are transferred to the next viral generation. Some mutations are beneficial, some are neutral, while others are deleterious but higher mutation rates result in higher genetic diversity. RNA viruses are more prone to genetic variations as compared to DNA viruses, therefore, a small escalation in their mutation rates can cause an RNA virus strain to wipe off locally [5]. Cellular organisms have exonucleases that are responsible for correcting the nucleotide misincorporations that occur during the process of replication. Most of the viral RNA replicases lack proofreading activity, therefore, the absence of 3′ prime exonucleases causes a rise in the mutation rate of RNA viruses as compared to the DNA viruses, contributing to the immense number of SARS-CoV-2 variants and its high infectivity [6].
Predicting viral mutations can help a lot in understanding the course of the pandemic by identifying potential drug targets and rapid vaccine development. Statistical learning and probabilistic models were used for mutation prediction like indels or substitutions to get insights into sequence evolution [7]. Over time, new methods have been designed for mutation prediction tasks in genomic sequences such as statistical relational learning where mutation data is associated with drug resistance information related to the nucleoside and nonnucleoside human immunodeficiency viruses (HIV) reverse transcriptase inhibitors [7]. Such methods are dependent on general engineering techniques such as "rational designs" to induce site-specific mutations. Mutation prediction has also been performed on the influenza virus genome; time-series mutation prediction model (Tempel) is one such example that makes use of LSTMs (long short-term memory units) with an attention mechanism. This study was further topped by a classification model that discriminates whether a mutation is imminent in the genomic sequences or at specific residue sites of the influenza virus genome for months or not [8]. In our current study, the data preprocessing part has been inspired by the work done in Tempel. In the literature, the mutation prediction task has been performed using a deep neural network, cause and effect relationship, and randomness [9]. Apart from using deep neural networks to predict new viral strains, rough set theory has been used to extract point mutation patterns [10]. Recently, there have been rapid advancements in the field of NMT which have also escalated the prediction of mutations and their assessment in diverse ways. This includes using large language models (LLMs) to: study viral escape [11,12], identify high-risk SARS-CoV-2 variants [13], predict mutations in the SARS-CoV-2 genome [14], predict antigenic evolution in the SARS-CoV-2 genome and revealing its evolutionary dynamics [15,16], predict evolutionary dynamics of proteins [17], predict protein structure over evolutionary information [18,19], decipher gene regulatory code as one gene can produce various proteins [20], and predict the molecular phenotype [21].
So far, NMT applications (seq2seq) related to mutation prediction focus on the prediction of the mutations in the influenza virus, Newcastle disease virus, and SARS-CoV-2. The previously proposed models related to machine translation are imprecise in two ways, their training data and their evaluation, i.e., the input and output language have common sequences which creates a bias while making predictions from the decoder, and there is no proper evaluation measure based on which models are evaluated as accuracy is not an appropriate measure to evaluate NMT models. Furthermore, their models do not utilize attention mechanisms which suggests that their models are not information-rich enough to predict diverse mutations [22,23,24] like ours. Finally, these models are trained on the protein sequences which means that silent mutations cannot be predicted by these models. Contrary to seq2seq modeling, LLMs have been used not only to study the evolution of viral genomes, but also to study the evolutionary dynamics of eukaryotic genomes and proteomes [17,18,19,20,21]. However, LLMs used for mutation predictions are more complicated than seq2seq models as they have a wider range of parameters to be tuned and are more computationally expensive. Therefore, in this study, we are using bidirectional gated recurrent units (GRUs) with an attention mechanism to predict mutations in the SARS-CoV-2 genome specifically targeting only seven non-structural proteins (NSPs), i.e., NSP1, NSP3, NSP5, NSP8, NSP9, NSP13, and NSP15. Our proposed model (stacked biGRU with an attention mechanism) is not only less computationally expensive but is also simpler in terms of fine-grained control over architecture and training than LLMs alongside providing good predictions.
We first tried three unidirectional models (RNNs, LSTMs, and GRUs) to predict mutations. Based on slightly better convergence than RNNs and LSTMs, we opted for the GRU model and applied the attention mechanism. Since uni-directional models had convergence issues, we incorporated a bidirectional feature to obtain the proposed model which shows superior performance.
The paper is structured as follows: First, we explain neural machine translation (NMT) and recurrent neural networks (RNNs). Next, we describe data processing by the use of machine learning algorithms for the creation of sequence pairs. Following this, the workings of the proposed models are explained. The significance of the results is elaborated in the discussion section and, lastly, the conclusion marks the end of the paper.
This section describes the protocol adopted for the current study. Figure 1 gives an overview of the workflow. The integrants of this figure are explained in detail.
Neural machine translation is a famous approach for machine translation problems where an end-to-end trained model learns the meaningful information from the source text and uses it to output its correct translation. Neural networks are used for this purpose which require very little supervision. Unlike the conventional MT models, NMT works by training a single and large neural network that translates the source text into target text correctly/suitably. The NMT model comprises an encoder network and a decoder network. The encoder reads a source string and encodes it into a vector of fixed length known as a context vector. The context vector is taken by the decoder to provide a translation. Encoder and decoder networks are trained together to maximize the probability of a correct translation [25]. One way to perform machine translation tasks is through sequence-to-sequence (seq2seq) models.
Sequence-to-sequence models are mainly used for language processing tasks, usually implemented with recurrent neural networks (RNNs) or their modified variants like LSTMs or GRUs, etc. Upon giving a sequence of inputs (s′1,s′2,s′3,...,s′T), sequence-to-sequence models generate a sequence of outputs (s1,s2,s3,...,sT) known as source and target sequences, respectively. Seq2seq models can be demonstrated as conditional language models:
P(s1,s2,...,sT,∣s′)=T∏t=1p(st∣s<t,s′), | (2.1) |
where p(st|s<t,s′) denotes the probability for the output token, given the input sequence s′ and prior outputs s<t. These models comprise encoder and decoder parts. The encoder is a recurrent neural network and its task is to learn meaningful patterns from the source language and pass it to the decoder. The decoder is a recurrent neural network as well, but it takes a hidden state vector from the encoder and constructs a sequence from the target language, see Figure 2. One drawback is that sequence-to-sequence models suffer from information loss if the length of the input sequences is very long. Thus, to cater to this issue, attention mechanism was proposed [25]. Attention allows the model to focus on those parts of the source sequence where precise relevant information is present by using context vectors. The conditional probability changes to
P(s1,s2,...,sT,∣s′)=softmax(Wp[ˆht,ct]), | (2.2) |
where Wp represents the learnable weight matrix, [ˆht,ct] represents the concatenated vector, ˆht is the hidden state of the decoder vector at time point t, and ct denotes the context vector.
This was extended by incorporating local and global attention mechanisms by Luong et al. [26]. According to Luong's attention, the context vector ct at each time step t is computed for the output sequence and depends on the attention scores eti associated with it. It is the weighted sum of hidden states of encoder hi where attention weights ati are used as weights:
ct=m∑i=1αtihi, | (2.3) |
where hi represents the hidden state of the encoder at time step i, and attention weights are computed by applying softmax to the attention scores,
ati=exp(eti)∑mk=1exp((etk), | (2.4) |
where m is the length of the source sequence and k represents the individual position in the source sequence, and eti=ˆhTthi where ˆhTt represents the transpose of the hidden state of the decoder at time step t. The context vector is concatenated with the hidden state of the decoder and used as input for the next layer:
[ˆht,ct]=concat(ˆht,ct). | (2.5) |
Recurrent neural networks (RNNs) are neural networks used to model sequential or time series data and are adapted from standard feed-forward networks. The most important component of an RNN is its hidden state which integrates information over several time points aiding in accurate and precise predictions. For each sequence and time point, it works recursively by taking the input (xo,x1,…,xt), updating the hidden state (ho,h1,…,ht), and predicting an output (yo,y1,…,yt). So for each time point from 0 to t, RNN works by iterating:
ht=tanh(whxxt+whhht−1+bh), | (2.6) |
yt=wyhht+by, | (2.7) |
where, whx, whh, and wyh is input to hidden, hidden to hidden, and hidden to output weight matrices, respectively. bh and by represent the bias vectors. An RNN unit is shown in Figure 2.
RNNs are trained by computing gradients via backpropagation through time, but sometimes gradients become too large (exploding gradients) or too small (vanishing gradients) ceasing the model to learn. The former can be fixed by simply truncating the gradients; whereas, the latter makes RNNs unsuitable for learning long-term dependencies [27].
Long short-term memory units (LSTMs) were proposed to resolve the vanishing gradient problem of RNNs, by introducing a cell state (incorporating long-term memory) with three logic gates (forget, input/update, and output gates) alongside a hidden state (which refers to short-term memory) for previous and current time points. For each time point, LSTM iterates by updating three gates. Forget gate ft determines the extent up to which existing memory is forgotten:
ft=σ(wf⋅[ht−1,xt]+bf). | (2.8) |
Input gate it regulates the amount of new content to be stored in the cell state ct, followed by the generation of a candidate cell state ˜ct which ultimately leads to the updation of the old cell state to a new one, i.e., ct:
it=σ(wi⋅[ht−1,xt]+bi), | (2.9) |
˜ct=tanh(wc⋅[ht−1,xt]+bc), | (2.10) |
ct=ftct−1+it˜ct. | (2.11) |
Output gate ot determines the amount of information required to be output in the form of hidden state ht, computed as:
ot=σ(wo⋅[ht−1,xt]+bo), | (2.12) |
ht=ot(tanh(ct)). | (2.13) |
An LSTM unit is shown in Figure 2. Unlike RNNS, LSTMs keep updating their context vector values by incorporating a cell state which makes them suitable for learning long-term dependencies [28].
Gated recurrent units (GRUs) are simpler versions of LSTMs that capture the dependencies of various time scales adaptively. Unlike LSTMs, GRUs have two gates, i.e., the update and reset gate. The update gate is simply the coalescence of the input and forget gates of LSTM. It also merges the hidden state and cell state of LSTM into one state ht. GRU generates ht for each time point by updating these two gates, where update gate zt determines which new information to keep and which past information to get rid of, whereas, reset gate rt ascertains how much past information to forget [28]. See the GRU unit in Figure 2. For a given time point t, both gates are computed as:
zt=σ(wzhht−1+wzxxt), | (2.14) |
rt=σ(wrhht−1+wrxxt). | (2.15) |
Activation of candidate state ˜ht is done as:
˜ht=tanh(wh(rt⋅ht−1)+wxxt). | (2.16) |
Generation of the new state ht is a linear interpolation of the previous and candidate state, as:
ht=ht−1(1−zt)+zt˜ht). | (2.17) |
SARS-CoV-2 is composed of structural and non-structural proteins. The former, as the name indicates, participate in making the viral structure while the latter are responsible for viral replication and assembly processes. For mutation prediction, we chose non-structural proteins as their tendency to mutate is few and far between as compared to structural proteins. Out of 16 NSPs, we chose 7 depending on their functionality; namely NSP1, NSP3, NSP5, NSP8, NSP9, NSP13, and NSP15, see Figure 3. NSP1 is mainly responsible for suppressing host innate immune functions and promoting viral protein translation. Therefore, it is known as the host shut-off factor [29]. NSP3 also disrupts the host immunity by inhibiting IFN (interferons) production and modifying the endoplasmic reticulum to double-membrane vesicles (DMVs) [30]. NSP5 works as the main protease, cleaving ORF1ab to generate more NSPs [31]. NSP8 promotes virus replication by making complexes with NSP6 and NSP7. It is also responsible for the identification of virus inhibitors [32]. NSP9 is a replicase enzyme blocking mRNA export of the host, and also dampens cytokine and interleukin-1α/β production to avoid activation of the host immune system [33]. NSP13 blocks interferon activation, downregulates IFIT1 protein expression, and reduces NF-κβ activation in the host [34]. NSP15 works as an RNA endonuclease that evades the host's innate immune response and antagonizes interferon [35].
The SARS-CoV-2 nucleotide sequences used in this study were downloaded from the GISAID repository [36]. GISAID (Global Initiative on Sharing All Influenza Data) was established initially to provide influenza virus genomic data worldwide which was further expanded to the provision of SARS-CoV-2 genomic data after the pandemic in late 2019. The selected NSPs' nucleotide sequences were downloaded in the fasta format limited to Germany only.
Data was preprocessed separately for each NSP as their sequence lengths vary. Following the removal of redundant sequences, nucleotide sequences were further subjected to the removal of sequences with inappropriate lengths or those carrying unidentified nucleotides, e.g., 'NN', 'XX', etc. Then sequences were separated based on the months they were reported in, and in this study, we have taken sequences from January 2020 until May 2021 for training and from June 2021 until March 2022 for evaluation purposes. There were a total of 876,554 sequences and we were left with 20,079 training sequences after removing duplicates. Contrary to this, for evaluation, we had 2,164,077 sequences which were reduced to 38,372 after removing duplicates. For further insights into the data, please see Table S1. Since a single nucleotide in the sequence itself does not carry meaningful information and appears just as a character, each nucleotide sequence was transformed into a sequence of codons, i.e., a fusion of three consecutive nucleotides.
For seq2seq translation, we need a pair of two sequences where the first one belongs to the source language, and the other belongs to the target language, i.e., mutated/evolved sequences of SARs-CoV-2. To create the dataset for our seq2seq machine translation task, we applied k-means clustering on all sequences for every month and generated 'N' number of clusters in each month. We computed immediate neighbors of sequences using nearest neighbors, i.e., cluster Cj,j∈{1,...,N} from month mi can have two immediate neighbors (clusters Cl,l∈{1,...,N}) from month mi+1. To consider only biologically realistic evolutionary behaviors, we computed Euclidean distance for each sequence s from cluster Cj of month mi to sequences from the two immediate neighbors (clusters) from month mi+1, i.e.,
0.5≤‖sj,i−sl,i+1‖≤3, | (2.18) |
where ‖⋅‖ represents the Euclidean norm and threshold values 0.5 and 3 are chosen for the selection of sequences in a pair based on this assumption that a virus can have only a few mutations in a month. The distances that are beyond the threshold may depict unrealistic mutations. The process is repeated for all of the sequences of each selected NSP, thus forming a diverse dataset containing various mutations. Please find the script in the supplementary information.
All of the models have been trained on Nvidia V100 series GPUs. The training time varies according to the length of the sequences and size of the batches. We have used Adam optimizer for all of our models with a learning rate of 0.0001. We configured the optimizer with default values of decay (0.9, 0.999) and eps (1e-08). For all seven datasets (selected target proteins), the baseline models have been trained for 8000 epochs. The train and test split ratio has been 75 and 25, respectively, for each protein. For smaller sequence lengths, the models show convergence of validation loss but, for longer sequence lengths, models are prone to overfitting and poor learning capabilities, not ensuring any convergence in terms of validation loss even at a higher number of epochs. Figure 4 shows the training of the baseline models for NSP1 ((a)–(c)) and NSP3 ((g)–(i)) being the smaller and larger protein length sequences from our targeted proteins, respectively. Training of the baseline models with the other five datasets is shown in Figures S1–S3. For stacked bidirectional GRUs, models for individual proteins have been trained for 2200 epochs. The shorter the length of the protein, the faster the validation loss of the model converges. Figure 4 (m), (n) shows model loss convergence for NSP1 and NSP3, and Figure S4 shows the remaining five proteins. Negative log-likelihood is used as the loss function for all of the baseline models. Bilingual evaluation understudy (BLEU) scores have been incorporated to evaluate the models.
The BLEU score has been used as an evaluation metric that measures the similitude between a reference and predicted sequence in machine translation tasks. It is a standard method for language translation evaluation ranging between 1.0 and 0.0. A score closer to 1.0 means higher similarity while a score closer to 0.0 means higher dissimilarity. It is language-independent, efficient, computationally inexpensive, and shows a high correlation with human evaluation [44]. The BLEU score for given and reference text corpus can be calculated by multiplying the brevity penalty (BP) with the geometric average of n-gram precision scores (pn):
BLEU=BP⋅exp(N∑n=1wnlogpn), | (2.19) |
where w represents positive weights for n-grams, and BP can be computed as:
BP={1ifc>re1−rcifc≤r. | (2.20) |
Here, c shows the length of the predicted sequence, whereas r is the length of the reference sequence. We set the BLEU score for 11 different n-grams (1 to 11). 1 to 10 are the usual n-grams, where the number shows the pair length, but the last one is a mixture of all of the first 10-grams.
We have proposed three baseline models for seq2seq translation to predict mutations in the SARS-CoV-2 genome. They are, namely, vanilla RNNs, simple LSTMs, and simple GRUs, and all baseline models share the same architecture. They comprise of an encoder and a decoder similar to a usual NMT-based sequence-to-sequence translation model. The encoder is comprised of a single unidirectional layer of RNN and an embedding layer. The embedding layer takes tokens of linearly encoded nucleotide trigrams and converts them into 128-sized 1-dimensional vectors. These vectors are passed through the recurrent layers with a hidden state of size 256. The hidden states from the RNN layer are passed to the decoder of the model. The decoder comprises an embedding and an RNN layer. The RNN layer is coupled with a linear layer having a softmax function as the activation function. The softmax function helps in identifying the most meaningful word (a trigram of nucleotides) at a certain time stamp. In addition, the decoder is fed with the hidden states of the encoder such that the most useful words can be retrieved during the process of translation by computing the attention scores and context vector. The decoder-encoder architecture is trained with the help of teacher forcing to have appropriate convergence.
Due to the simplicity and inability of baseline models to capture complex feature distribution, we created stacked bidirectional GRUs by stacking the identical layers of the model to make a deep model. Such a model is capable enough to learn the complex feature patterns existing in the training data. The model is comprised of an encoder and a decoder, where the encoder consists of one embedding layer and two layers of bidirectional GRUs containing 256 neurons in each layer. Whereas, the decoder contains embedding, attention, GRU, and linear layers. The embedding size is the same as the encoder where a dropout of 0.2 is used. The regularized vectors are passed through linear layers inspired by Luong's attention. Followed by computation of the attention, these values are passed to the GRU layers and ultimately to the linear layer where log softmax has been used as an activation function. This makes the model solve a classification problem where the goal is to find the most probable word in a specific time stamp. According to our experiments, the regularization does not affect the training process a lot, but the length of the hidden size for the GRUs and the embedding size can affect the overfitting of the model by reducing the size of the vectors.
According to our experiments, the models perform quite well with the use of bidirectionality and only 2 stacked layers. All of the models have been implemented using Pytorch and Scikit-learn with Python as the base programming language. We evaluate our models based on the validation and training loss and BLEU scores. The baseline models work in the usual way as the RNN-based encoder-decoder architecture works, i.e., by simply passing the hidden states from the encoder to the decoder. In the case of bidirectional models, the two hidden states, i.e., those based on past and future events, are concatenated with each other from the deep layer of RNNs and then passed to the decoder. This way, the stacked models can learn more about the hidden state vectors. The predictions from the baseline models are not adequate for biological analyses as the models take longer to converge and show poor BLEU scores. Training results for all of the baseline models with each non-structural protein are given in Figures S1–S3.
The computational complexity for stacked bidirectional GRUs is O(2nd2) and for the attention mechanism is O(nencndecd), which makes the total complexity of our model O((2nd2)+(nencndecd)). Here, n is the sequence length, d is the size of the hidden states, nenc represents the length of the encoder sequences, and ndec represents the length of the decoder sequences. The 2 in the notation depicts that the complexity has been doubled because the input sequence is processed in both directions, i.e., forward and backward. Since we have chosen seven non-structural proteins in the current study and trained the models separately, the nenc and ndec will be different for each protein as per the protein length. However, for each protein, the length of these variables will be the same. The complexity of all seven stacked biGRU models is given in Table S2. Computational complexity increases with an increase in the length of the input sequences, however, the computational scalability in terms of resource utilization and model efficiency remains the same. In terms of time, it decreases. The longer the sequences, the greater the time required to execute them.
For the evaluation of the models, we created a new dataset by collecting nucleotide sequences of a future (concerning the period considered for the training set) time duration, i.e., from June 2021 until March 2022. Thus, obtained data not only consists of new unseen sequences but also consists of mutations that were not present in the period, i.e., January 2020 until May 2021 considered for the training set. Consequently, the evaluation sequence truly represents the evolved/mutated RNA sequences of SARs-CoV-2. The collected data represent raw evaluation sequences which are then subjected to the preprocessing and clustering steps outlined in Sections 2.5.3 and 2.5.4, respectively. Based on this, we obtain a labeled evaluation dataset consisting of a source sequence and its associated mutated sequence for seven non-structural proteins, namely NSP1, NPS3, NSP5, NSP8, NSP9, NSP13, and NPS15. Following this, we feed the (protein-specific) stacked-BiGRU model with the corresponding protein input sequence to obtain a prediction of a most likely mutated nucleotide sequence. The obtained predictions are compared with the ground truth sequence using the BLEU score metrics. The obtained scores are shown in Table 1 and also depicted as bar plots in Figure 5. Based on these scores, we see that the prediction of the stacked BiGRUs is good for all of the proteins. For proteins NSP8 and NSP9, the 9 and 10-gram scores are relatively poor with NSP8 showing below-acceptable performance having values below 0.6 for 9 and 10-gram scores.
Protein | 1-gram | 2-gram | 3-gram | 4-gram | 5-gram | 6-gram | 7-gram | 8-gram | 9-gram | 10-gram | 11-gram |
NSP1 | 0.997 | 0.986 | 0.976 | 0.963 | 0.957 | 0.951 | 0.945 | 0.94 | 0.934 | 0.928 | 0.957 |
NSP3 | 0.998 | 0.996 | 0.991 | 0.982 | 0.971 | 0.96 | 0.951 | 0.944 | 0.937 | 0.93 | 0.966 |
NSP5 | 0.992 | 0.982 | 0.966 | 0.938 | 0.917 | 0.901 | 0.887 | 0.874 | 0.861 | 0.847 | 0.915 |
NSP8 | 0.976 | 0.953 | 0.914 | 0.861 | 0.785 | 0.702 | 0.648 | 0.607 | 0.57 | 0.542 | 0.74 |
NSP9 | 0.969 | 0.933 | 0.904 | 0.855 | 0.813 | 0.781 | 0.754 | 0.731 | 0.71 | 0.693 | 0.809 |
NSP13 | 0.999 | 0.998 | 0.997 | 0.996 | 0.995 | 0.994 | 0.993 | 0.992 | 0.991 | 0.99 | 0.994 |
NSP15 | 0.999 | 0.998 | 0.997 | 0.996 | 0.995 | 0.994 | 0.992 | 0.991 | 0.99 | 0.989 | 0.994 |
After performing the above direct evaluation between the predicted and the ground truth sequence, we next want to evaluate the qualitative difference in the translated amino-acid (AA) sequence. To this end, for the above-mentioned seven proteins, we translate the predicted nucleotide sequence to the corresponding AA sequence corresponding to the respective protein. Next, we compare the so-obtained mutated AA sequence (ˆApseq) with that of the reference AA sequence (Apseq) obtained from the base reference (protein specific indexed by p) nucleotide sequence of SARS-CoV-2 and aim to obtain frequency/number of mutations at a given amino acid position of the reference genome. To this end, we perform a string matching of the two sequences, i.e., Apseq with ˆApseq, and determine the positions where there are differences between the two. This provides the count/frequency f of mismatch at a particular position x of the reference AA sequence Apseq. Thus, by summing the mutation counts at a given position across all predicted sequence samples, we obtain the mutation frequency/density f(t,x) for some future time t>t0 and position x of an AA chain representing the protein. Accordingly, we have that f(t0,x) denotes the mutation frequency corresponding to the source/input nucleotide chain, f(t,x) for t>t0 denotes the mutation frequency corresponding to the target/output/ground-truth/measured nucleotide chain, and ˆf(t,x) for t>t0 denotes the mutation frequency of the predicted nucleotide sequence. It is worth noting that the source/input nucleotide sequence (i.e., the sequence fed to the network as input) may have non-zero mutation frequency since the obtained data is that of a mutated virus, thus having differences with respect to the reference nucleotide sequence. Figure 6 depicts stem plots for the mutation frequency of the predicted sequences and the ground truth nucleotide sequences for all seven proteins. Based on this, we see that the obtained predictions are able to detect some of the point mutations in the amino-acid chain. In the case of NSP3, we see that the location of several point mutations, i.e., x∈{183,646,890,1159,1412,1682,1867}, were correctly predicted although the counts (i.e., the actual frequency) vary a bit from the ground truth. Similarly, for NSP5, the location of the point mutations matches for x∈{157,213,242}, for NSP15 at x∈{11,132,264}, and so on. Predicted frequently occurring point mutations for all targeted proteins are shown in Figures S5–S11.
Correspondingly, the precision and recall scores depicted in Table 2 indicate the high reliability of the model within 2–3 units of positional error. This is confirmed by the precision score, where we see above 80% precision for detecting the point mutations within the 2 units of accuracy and 100% precision for detecting within 3 units of accuracy for all of the target proteins. On the other hand, the recall scores (for detecting point mutations) are pretty low with less than 20% for proteins with an AA sequence length of more than 200. However, NSP8 and NSP9, which have AA sequence lengths less than 200, have recall scores up to 87% and 73% (for point mutations of at most 3 units of error). This indicates that the model is highly selective, especially for proteins with higher sequence lengths (>300) of its corresponding AA (indicated by the Selectivity column of Table 2), and only detects mutations that are positionally precise.
Protein | AA sequence length | Mutation count | Predicted mutations | True mutations | Precision | Recall | Selectivity |
NSP1 | 181 | 85 | 4 | [2, 4, 4, 4] | [0.5, 1, 1, 1] | [0.0223, 0.046, 0.046, 0.046] | 0.978 |
NSP3 | 1946 | 360 | 23 | [7, 12, 19, 23] | [0.3, 0.52, 0.83, 1.0] | [0.02, 0.033, 0.052, 0.063] | 0.99 |
NSP5 | 307 | 65 | 11 | [3, 8, 11, 11] | [0.27, 0.72, 1.0, 1.0] | [0.04, 0.12, 0.17, 0.17] | 0.967 |
NSP8 | 199 | 39 | 34 | [9, 20, 31, 34] | [0.26, 0.52, 0.91, 1.0] | [0.23, 0.51, 0.79, 0.87] | 0.844 |
NSP9 | 114 | 19 | 14 | [2, 9, 14, 14] | [0.14, 0.64, 1.0, 1.0] | [0.11, 0.47, 0.73, 0.73] | 0.87 |
NSP13 | 602 | 123 | 2 | [0, 1, 2, 2] | [0.0, 0.5, 1.0, 1.0] | [0.0, 0.008, 0.016, 0.016] | 0.99 |
NSP15 | 347 | 104 | 4 | [3, 4, 4, 4] | [0.75, 0.5, 1.0, 1.0] | [0.029, 0.0385, 0.0385, 0.0385] | 0.99 |
Based on these observations, we can conclude that the stacked BiGRUs do a fairly good job at predicting practically relevant predictions of plausible mutations. Furthermore, whenever the model detects a mutation, it seems to be of high precision, with at most 3 units of positional error, thus providing a good indication of the validity of the model and its usefulness.
Nucleotide changes in the genome are known as mutations. In this study, we are focusing on point mutations, which are changes in the position of one nucleotide that can either be an insertion, deletion, or transition/transversion. In viruses, mutations play a huge role in the emergence of various viral strains, which either render them weak and in return wipe the entire viral strain or strengthen them. As a result of favorable mutations, viruses can adapt to a wide range of hosts, increase their virulence, better mask, and evade host immune responses. SARS-CoV-2 has created havoc since its emergence and, due to its high mutation rate, it is hard to contain. Therefore, in this current work, we have used recurrent neural networks to predict mutations in its genome. Predicted mutations will help in designing inhibitory molecules to obstruct viral transmission in the future. We have selected non-structural proteins for the mutation prediction task as these proteins tend to change their genomic sequences less as compared to structural ones, thus proffering the designed respective inhibitory molecules to hold their efficacy for a longer time.
In our study, mutation prediction has been defined as a machine translation task. We started with vanilla RNNs, but due to their lack of ability to capture long-term dependencies in data, we tried solving this problem with simple LSTMs and simple GRUs. Both performed better than vanilla RNNs but their losses took longer to converge. Moreover, obtained BLEU scores were also not up to the mark. Since simple GRUs performed slightly better than simple LSTMS, we chose to stack the two bidirectional layers of GRUs and applied the attention mechanism to further exhilarate the predictions.
We trained our models using genomic sequence data of seven targeted proteins–NSP1, NSP3, NSP5, NSP8, NSP9, NSP13, and NSP15–after finding nearest neighbors for sequences and clustering the data. A separate model has been trained for each protein due to their varying sequence lengths. For model training, the sequence data comes from 17 months (i.e., January 2020 to May 2021) and for model evaluation, the sequence data comes from 10 months (i.e., June 2021 to March 2022). The BLEU score has been used as a metric to evaluate the model outcomes. It shows quite good translation results with an average of 0.9 for all of the targeted proteins except NSP8 and NSP9, where 9 and 10-gram scores lie in the range of 0.5 and 0.7. Differences in the metric scores lie in the availability of data for training since the training data for both NSP8 and NSP9 are smaller as compared to the other selected proteins. Contrary to this, other protein models have performed very well, which is depicted in the BLEU scores as well as in precision and recall scores.
In our study, we have made sure that the input and output languages should have some evolutionary relation to eliminate the bias in predictions, unlike other proposed studies where both languages have common sequences. We achieved this by performing k-means clustering and nearest neighbor-joining methods. Moreover, we have used attention mechanisms to enrich the mutation prediction models for this task. We have focused on evaluating the predictions and model performance by using BLEU scores. Implementing all of these features in this study will advance the scope of this work beyond that of previously published work.
We have predicted some new mutations in the seven targeted proteins that did not exist in the training and evaluation data before, see Table 3. Mutation statistics for the whole study are given in Table 4. Further downstream analysis of predicted mutations will help in selecting suitable targeting protein domains to design inhibitory molecules against viruses. This study adds a slightly new direction in fighting the pandemics of the future by preparing beforehand. Our models are generic and will work with any kind of genomic sequence coming from different sources.
Protein | Predicted new mutations |
NSP1 | W160T, T169G, G179C |
NSP3 | P511R, R645S, S691T, S696T, T762P, T969P, S1037T, T1045P, T1183P, S1205P, R1340S, |
P1376R, P1557R, S1614T, S1681T, T1778P, S1806T, R1819S | |
NSP5 | I151S, H162V, G173H, T195G, I212S, R221T, L241R, I248S, L249R, I280S |
NSP8 | Q23D, V25Q, N27V, G28N, V32Q, V33Q, N42V, D51G, Q72D, V82Q, Q87D, N99V, N108V, |
G112N, V114Q, V129Q, N139V, G143N, Q157D, V158Q, N175V, N178V, N191V, Q197D | |
NSP13 | S262C |
NSP15 | D131E |
Protein | t_seq | t_mut | e_seq | e_mut | p_seq | p_mut | pt_mut | pe_mut | n_mut |
NSP1 | 864 | 375 | 2024 | 767 | 288 | 178 | 70 | 175 | 3 |
NSP3 | 13,347 | 2888 | 23,999 | 4195 | 840 | 559 | 295 | 541 | 18 |
NSP5 | 1002 | 456 | 2137 | 495 | 227 | 248 | 174 | 118 | 10 |
NSP8 | 577 | 394 | 852 | 384 | 125 | 110 | 40 | 83 | 24 |
NSP13 | 2674 | 680 | 6921 | 1063 | 472 | 218 | 111 | 217 | 1 |
NSP15 | 1232 | 512 | 2578 | 858 | 297 | 215 | 94 | 214 | 1 |
This study focused on the mutation prediction in the SARS-CoV-2 genome using neural machine translation. The sequence data was downloaded from GISAID comprising of sequences from Germany only. The choice of non-structural proteins was made to prolong the efficacy of the designed inhibitory molecules against viral proteins as non-structural proteins rear fewer mutations as compared to structural proteins. To perform machine translation tasks, three models were used as baseline models, i.e., vanilla RNNs, simple LSTMs, and simple GRUs. Since simple GRUs performed much and slightly better than vanilla RNNs and simple LSTMs, respectively, they were chosen to be stacked and implemented with an attention mechanism to further improve the prediction results. The results showed that stacking the layers and attention mechanisms has indeed enhanced the performance of the models. The prediction results highlighted the significance of certain mutations in each of the seven proteins and also produced some new mutations that did not exist in the data before. These findings may contribute to fighting the pandemics of the future by identifying the critical mutations. Although our model can predict mutations in individual proteins, it is unlikely to be any better at doing the same task on the whole genome. In the future, it will be interesting to explore this problem using transformer models on larger protein sequences or whole viral genomes, as transformers are more computationally expensive than stacked biGRUs and can perform such tasks more effectively.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work was supported by the German Federation of Industrial Research Associations (AiF) within the scope of the AIMPID project (AI-based Mutation Predictions and relevant Protein Inhibitor Development in SARS-CoV-2).
The authors declare there is no conflict of interest.
[1] | World Health Organization, WHO Coronavirus (COVID-19) Dashboard, 2023. Available from: https://covid19.who.int. |
[2] |
R. Lu, X. Zhao, J. Li, P. Niu, B. Yang, H. Wu, et al., Genomic characterisation and epidemiology of 2019 novel coronavirus: implications for virus origins and receptor binding, The Lancet, 395 (2020), 565–574. https://doi.org/10.1016/S0140-6736(20)30251-8 doi: 10.1016/S0140-6736(20)30251-8
![]() |
[3] |
A. Naqvi, K. Fatima, T. Mohammad, U. Fatima, I. Singh, A. Singh, et al., Insights into SARS-CoV-2 genome, structure, evolution, pathogenesis and therapies: structural genomics approach, Biochim. Biophys. Acta, Mol. Basis Dis., 1866 (2020), 165878. https://doi.org/10.1016/j.bbadis.2020.165878 doi: 10.1016/j.bbadis.2020.165878
![]() |
[4] |
R. Sanjuán, M. Nebot, N. Chirico, L. Mansky, R. Belshaw, Viral mutation rates, J. Virol., 84 (2010), 9733–9748. https://doi.org/10.1128/jvi.00694-10 doi: 10.1128/jvi.00694-10
![]() |
[5] |
S. Duffy, Why are RNA virus mutation rates so damn high, PLoS Biol., 16 (2018), e3000003. https://doi.org/10.1371/journal.pbio.3000003 doi: 10.1371/journal.pbio.3000003
![]() |
[6] |
R. Carrasco-Hernandez, R. Jácome, Y. L. Vidal, S. P. de León, Are RNA viruses candidate agents for the next global pandemic? A review, ILAR J., 58 (2017), 343–358. https://doi.org/10.1093/ilar/ilx026 doi: 10.1093/ilar/ilx026
![]() |
[7] |
E. Cilia, S. Teso, S. Ammendola, T. Lenaerts, A. Passerini, Predicting virus mutations through statistical relational learning, BMC Bioinf., 15 (2014), 309. https://doi.org/10.1186/1471-2105-15-309 doi: 10.1186/1471-2105-15-309
![]() |
[8] |
R. Yin, E. Luusua, J. Dabrowski, Y. Zhang, C. Kwoh, Tempel: time-series mutation prediction of influenza A viruses via attention-based recurrent neural networks, Bioinformatics, 36 (2020), 2697–2704. https://doi.org/10.1093/bioinformatics/btaa050 doi: 10.1093/bioinformatics/btaa050
![]() |
[9] |
G. Wu, S. Yan, Prediction of mutations engineered by randomness in H5N1 neuraminidases from influenza A virus, Amino Acids, 34 (2008), 81–90. https://doi.org/10.1007/s00726-007-0579-z doi: 10.1007/s00726-007-0579-z
![]() |
[10] |
M. Salama, A. Hassanien, A. Mostafa, The prediction of virus mutation using neural networks and rough set techniques, EURASIP J. Bioinf. Syst. Biol., 2016 (2016), 10. https://doi.org/10.1186/s13637-016-0042-0 doi: 10.1186/s13637-016-0042-0
![]() |
[11] |
B. Hie, E. Zhong, B. Berger, B. Bryson, Learning the language of viral evolution and escape, Science, 371 (2021), 284–288. https://doi.org/10.1126/science.abd7331 doi: 10.1126/science.abd7331
![]() |
[12] |
N. Thadani, S. Gurev, P. Notin, N. Youssef, N. Rollins, D. Ritter, et al., Learning from prepandemic data to forecast viral escape, Nature, 622 (2023), 818–825. https://doi.org/10.1038/s41586-023-06617-0 doi: 10.1038/s41586-023-06617-0
![]() |
[13] |
K. Beguir, M. Skwark, Y. Fu, T. Pierrot, N. Carranza, A. Laterre, et al., Early computational detection of potential high-risk SARS-CoV-2 variants, Comput. Biol. Med., 155 (2023), 106618. https://doi.org/10.1016/j.compbiomed.2023.106618 doi: 10.1016/j.compbiomed.2023.106618
![]() |
[14] |
B. Zhou, H. Zhou, X. Zhang, X. Xu, Y. Chai, Z. Zheng, et al., TEMPO: a transformer-based mutation prediction framework for SARS-CoV-2 evolution, Comput. Biol. Med., 152 (2023), 106264. https://doi.org/10.1016/j.compbiomed.2022.106264 doi: 10.1016/j.compbiomed.2022.106264
![]() |
[15] |
W. Han, N. Chen, X. Xu, A. Sahil, J. Zhou, Z. Li, et al., Predicting the antigenic evolution of SARS-COV-2 with deep learning, Nat. Commun., 14 (2023), 3478. https://doi.org/10.1038/s41467-023-39199-6 doi: 10.1038/s41467-023-39199-6
![]() |
[16] |
M. Zvyagin, A. Brace, K. Hippe, Y. Deng, B. Zhang, C. Bohorquez, et al., GenSLMs: genome-scale language models reveal SARS-CoV-2 evolutionary dynamics, Int. J. High Perform. Comput. Appl., 37 (2023), 683–705. https://doi.org/10.1177/10943420231201154 doi: 10.1177/10943420231201154
![]() |
[17] |
B. Hie, K. Yang, P. Kim, Evolutionary velocity with protein language models predicts evolutionary dynamics of diverse proteins, Cell Syst., 13 (2022), 274–285. https://doi.org/10.1016/j.cels.2022.01.003 doi: 10.1016/j.cels.2022.01.003
![]() |
[18] |
Z. Lin, H. Akin, R. Rao, B. Hie, Z. Zhu, W. Lu, et al., Evolutionary-scale prediction of atomic-level protein structure with a language model, Science, 379 (2023), 1123–1130. https://doi.org/10.1126/science.ade2574 doi: 10.1126/science.ade2574
![]() |
[19] |
A. Elnaggar, M. Heinzinger, C. Dallago, G. Rehawi, Y. Wang, L. Jones, et al., Prottrans: toward understanding the language of life through self-supervised learning, IEEE Trans. Pattern Anal. Mach. Intell., 44 (2021), 7112–7127. https://doi.org/10.1109/TPAMI.2021.3095381 doi: 10.1109/TPAMI.2021.3095381
![]() |
[20] |
Y. Ji, Z. Zhou, H. Liu, R. Davuluri, DNABERT: pre-trained Bidirectional Encoder Representations from Transformers model for DNA-language in genome, Bioinformatics, 37 (2021), 2112–2120. https://doi.org/10.1093/bioinformatics/btab083 doi: 10.1093/bioinformatics/btab083
![]() |
[21] | H. Dalla-Torre, L. Gonzalez, J. Mendoza-Revilla, N. Carranza, A. Grzywaczewski, F. Oteri, et al., The nucleotide transformer: building and evaluating robust foundation models for human genomics, preprint, 2023. https://doi.org/10.1101/2023.01.11.523679 |
[22] |
P. Pushkar, C. Ananth, P. Nagrath, J. Al-Amri, Vividha, A. Nayyar, Mutation prediction for coronaviruses using genome sequence and recurrent neural networks, CMC-Comput. Mater., 73 (2022), 1601–1619. https://doi.org/10.32604/cmc.2022.026205 doi: 10.32604/cmc.2022.026205
![]() |
[23] |
T. Mohamed, S. Sayed, A. Salah, E. Houssein, Long short-term memory neural networks for RNA viruses mutations prediction, Math. Probl. Eng., 2021 (2021), 9980347. https://doi.org/10.1155/2021/9980347 doi: 10.1155/2021/9980347
![]() |
[24] |
S. Tasnim, K. Talukder, A. Asfi, Next mutation prediction of SARS-COV-2 spike protein using encoder-decoder based long short term memory (LSTM) method, Khulna Univ. Stud., 2022 (2022), 803–816. https://doi.org/10.53808/KUS.2022.ICSTEM4IR.0142-se doi: 10.53808/KUS.2022.ICSTEM4IR.0142-se
![]() |
[25] | D. Bahdanau, K. Cho, Y. Bengio, Neural machine translation by jointly learning to align and translate, preprint, arXiv: 1409.0473. |
[26] | M. Luong, H. Pham, C. Manning, Effective approaches to attention-based neural machine translation, preprint, arXiv: 1508.04025. |
[27] | I. Sutskever, J. Martens, G. Hinton, Generating text with recurrent neural networks, in Proceedings of the 28th International Conference on Machine Learning (ICML-11), (2011), 1017–1024. |
[28] | J. Chung, C. Gulcehre, K. Cho, Y. Bengio, Empirical evaluation of gated recurrent neural networks on sequence modeling, preprint, arXiv: 1412.3555. |
[29] |
K. Schubert, E. Karousis, A. Jomaa, A. Scaiola, B. Echeverria, L. Gurzeler, et al., SARS-CoV-2 NSP1 binds the ribosomal mRNA channel to inhibit translation, Nat. Struct. Mol. Biol., 27 (2020), 959–966. https://doi.org/10.1038/s41594-020-0511-8 doi: 10.1038/s41594-020-0511-8
![]() |
[30] |
B. Qin, Z. Li, K. Tang, T. Wang, Y. Xie, S. Aumonier, et al., Identification of the SARS-unique domain of SARS-CoV-2 as an antiviral target, Nat. Commun., 14 (2023), 3999. https://doi.org/10.1038/s41467-023-39709-6 doi: 10.1038/s41467-023-39709-6
![]() |
[31] |
Y. Zheng, J. Deng, L. Han, M. Zhuang, Y. Xu, J. Zhang, et al., SARS-CoV-2 NSP5 and N protein counteract the RIG-I signaling pathway by suppressing the formation of stress granules, Signal Transduction Targeted Ther., 7 (2022), 22. https://doi.org/10.1038/s41392-022-00878-3 doi: 10.1038/s41392-022-00878-3
![]() |
[32] |
S. Reshamwala, V. Likhite, M. Degani, S. Deb, S. Noronha, Mutations in SARS-CoV-2 NSP7 and NSP8 proteins and their predicted impact on replication/transcription complex structure, J. Med. Virol., 93 (2021), 4616–4619. https://doi.org/10.1002/jmv.26791 doi: 10.1002/jmv.26791
![]() |
[33] | G. Yeo, J. Xiang, J. Mueller, E. Luo, B. Yee, D. Schafer, et al., Discovery and functional interrogation of SARS-CoV-2 protein-RNA interactions, preprint, 2022. https://doi.org/10.21203/rs.3.rs-1394331/v1 |
[34] |
C. Vazquez, S. Swanson, S. Negatu, M. Dittmar, J. Miller, H. Ramage, et al., SARS-CoV-2 viral proteins NSP1 and NSP13 inhibit interferon activation through distinct mechanisms, PLoS One, 16 (2021), e0253089. https://doi.org/10.1371/journal.pone.0253089 doi: 10.1371/journal.pone.0253089
![]() |
[35] |
M. Pillon, M. Frazier, L. Dillard, J. Williams, S. Kocaman, J. Krahn, et al., Cryo-EM structures of the SARS-CoV-2 endoribonuclease Nsp15 reveal insight into nuclease specificity and dynamics, Nat. Commun., 12 (2021), 636. https://doi.org/10.1038/s41467-020-20608-z doi: 10.1038/s41467-020-20608-z
![]() |
[36] |
S. Khare, C. Gurry, L. Freitas, M. Schultz, G. Bach, A. Diallo, et al., Perspectives: GISAID's role in pandemic response, China CDC Weekly, 3 (2021), 1049–1051. https://doi.org/10.46234/ccdcw2021.255 doi: 10.46234/ccdcw2021.255
![]() |
[37] |
L. Clark, T. Green, C. Petit, Structure of nonstructural protein 1 from SARS-CoV-2, J. Virol., 95 (2021), 4. https://doi.org/10.1128/jvi.02019-20 doi: 10.1128/jvi.02019-20
![]() |
[38] |
V. Srinivasan, H. Brognaro, P. Prabhu, E. Souza, S. Günther, P. Reinke, et al., Antiviral activity of natural phenolic compounds in complex at an allosteric site of SARS-CoV-2 papain-like protease, Commun. Biol., 5 (2022), 805. https://doi.org/10.1038/s42003-022-03737-7 doi: 10.1038/s42003-022-03737-7
![]() |
[39] |
A. Ebrahim, B. Riley, D. Kumaran, B. Andi, M. Fuchs, S. McSweeney, et al., The temperature-dependent conformational ensemble of SARS-CoV-2 main protease (Mpro), IUCrJ, 9 (2022), 682–694. https://doi.org/10.1107/S2052252522007497 doi: 10.1107/S2052252522007497
![]() |
[40] |
M. Biswal, S. Diggs, D. Xu, N. Khudaverdyan, J. Lu, J. Fang, et al., Two conserved oligomer interfaces of NSP7 and NSP8 underpin the dynamic assembly of SARS-CoV-2 RdRP, Nucleic Acids Res., 49 (2021), 5956–5966. https://doi.org/10.1093/nar/gkab370 doi: 10.1093/nar/gkab370
![]() |
[41] |
C. Zhang, Y. Chen, L. Li, Y. Yang, J. He, C. Chen, et al., Structural basis for the multimerization of nonstructural protein NSP9 from SARS-CoV-2, Mol. Biomed., 1 (2020), 5. https://doi.org/10.1186/s43556-020-00005-0 doi: 10.1186/s43556-020-00005-0
![]() |
[42] |
J. Chen, Q. Wang, B. Malone, E. Llewellyn, Y. Pechersky, K. Maruthi, et al., Ensemble cryo-EM reveals conformational states of the NSP13 helicase in the SARS-CoV-2 helicase replication–transcription complex, Nat. Struct. Mol. Biol., 29 (2022), 250–260. https://doi.org/10.1038/s41594-022-00734-6 doi: 10.1038/s41594-022-00734-6
![]() |
[43] |
Y. Kim, R. Jedrzejczak, N. Maltseva, M. Wilamowski, M. Endres, A. Godzik, et al., Crystal structure of NSP15 endoribonuclease NendoU from SARS-CoV-2, Protein Sci., 29 (2020), 1596–1605. https://doi.org/10.1002/pro.3873 doi: 10.1002/pro.3873
![]() |
[44] | K. Papineni, S. Roukos, T. Ward, W. Zhu, Bleu: a method for automatic evaluation of machine translation, in Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, (2002), 311–318. https://doi.org/10.3115/1073083.1073135 |
![]() |
![]() |
1. | Xia Zhan, Innovative Research on the Teaching Mode of English Translation Course in Colleges and Universities with the Support of Deep Learning, 2024, 9, 2444-8656, 10.2478/amns-2024-2824 |
Protein | 1-gram | 2-gram | 3-gram | 4-gram | 5-gram | 6-gram | 7-gram | 8-gram | 9-gram | 10-gram | 11-gram |
NSP1 | 0.997 | 0.986 | 0.976 | 0.963 | 0.957 | 0.951 | 0.945 | 0.94 | 0.934 | 0.928 | 0.957 |
NSP3 | 0.998 | 0.996 | 0.991 | 0.982 | 0.971 | 0.96 | 0.951 | 0.944 | 0.937 | 0.93 | 0.966 |
NSP5 | 0.992 | 0.982 | 0.966 | 0.938 | 0.917 | 0.901 | 0.887 | 0.874 | 0.861 | 0.847 | 0.915 |
NSP8 | 0.976 | 0.953 | 0.914 | 0.861 | 0.785 | 0.702 | 0.648 | 0.607 | 0.57 | 0.542 | 0.74 |
NSP9 | 0.969 | 0.933 | 0.904 | 0.855 | 0.813 | 0.781 | 0.754 | 0.731 | 0.71 | 0.693 | 0.809 |
NSP13 | 0.999 | 0.998 | 0.997 | 0.996 | 0.995 | 0.994 | 0.993 | 0.992 | 0.991 | 0.99 | 0.994 |
NSP15 | 0.999 | 0.998 | 0.997 | 0.996 | 0.995 | 0.994 | 0.992 | 0.991 | 0.99 | 0.989 | 0.994 |
Protein | AA sequence length | Mutation count | Predicted mutations | True mutations | Precision | Recall | Selectivity |
NSP1 | 181 | 85 | 4 | [2, 4, 4, 4] | [0.5, 1, 1, 1] | [0.0223, 0.046, 0.046, 0.046] | 0.978 |
NSP3 | 1946 | 360 | 23 | [7, 12, 19, 23] | [0.3, 0.52, 0.83, 1.0] | [0.02, 0.033, 0.052, 0.063] | 0.99 |
NSP5 | 307 | 65 | 11 | [3, 8, 11, 11] | [0.27, 0.72, 1.0, 1.0] | [0.04, 0.12, 0.17, 0.17] | 0.967 |
NSP8 | 199 | 39 | 34 | [9, 20, 31, 34] | [0.26, 0.52, 0.91, 1.0] | [0.23, 0.51, 0.79, 0.87] | 0.844 |
NSP9 | 114 | 19 | 14 | [2, 9, 14, 14] | [0.14, 0.64, 1.0, 1.0] | [0.11, 0.47, 0.73, 0.73] | 0.87 |
NSP13 | 602 | 123 | 2 | [0, 1, 2, 2] | [0.0, 0.5, 1.0, 1.0] | [0.0, 0.008, 0.016, 0.016] | 0.99 |
NSP15 | 347 | 104 | 4 | [3, 4, 4, 4] | [0.75, 0.5, 1.0, 1.0] | [0.029, 0.0385, 0.0385, 0.0385] | 0.99 |
Protein | Predicted new mutations |
NSP1 | W160T, T169G, G179C |
NSP3 | P511R, R645S, S691T, S696T, T762P, T969P, S1037T, T1045P, T1183P, S1205P, R1340S, |
P1376R, P1557R, S1614T, S1681T, T1778P, S1806T, R1819S | |
NSP5 | I151S, H162V, G173H, T195G, I212S, R221T, L241R, I248S, L249R, I280S |
NSP8 | Q23D, V25Q, N27V, G28N, V32Q, V33Q, N42V, D51G, Q72D, V82Q, Q87D, N99V, N108V, |
G112N, V114Q, V129Q, N139V, G143N, Q157D, V158Q, N175V, N178V, N191V, Q197D | |
NSP13 | S262C |
NSP15 | D131E |
Protein | t_seq | t_mut | e_seq | e_mut | p_seq | p_mut | pt_mut | pe_mut | n_mut |
NSP1 | 864 | 375 | 2024 | 767 | 288 | 178 | 70 | 175 | 3 |
NSP3 | 13,347 | 2888 | 23,999 | 4195 | 840 | 559 | 295 | 541 | 18 |
NSP5 | 1002 | 456 | 2137 | 495 | 227 | 248 | 174 | 118 | 10 |
NSP8 | 577 | 394 | 852 | 384 | 125 | 110 | 40 | 83 | 24 |
NSP13 | 2674 | 680 | 6921 | 1063 | 472 | 218 | 111 | 217 | 1 |
NSP15 | 1232 | 512 | 2578 | 858 | 297 | 215 | 94 | 214 | 1 |
Protein | 1-gram | 2-gram | 3-gram | 4-gram | 5-gram | 6-gram | 7-gram | 8-gram | 9-gram | 10-gram | 11-gram |
NSP1 | 0.997 | 0.986 | 0.976 | 0.963 | 0.957 | 0.951 | 0.945 | 0.94 | 0.934 | 0.928 | 0.957 |
NSP3 | 0.998 | 0.996 | 0.991 | 0.982 | 0.971 | 0.96 | 0.951 | 0.944 | 0.937 | 0.93 | 0.966 |
NSP5 | 0.992 | 0.982 | 0.966 | 0.938 | 0.917 | 0.901 | 0.887 | 0.874 | 0.861 | 0.847 | 0.915 |
NSP8 | 0.976 | 0.953 | 0.914 | 0.861 | 0.785 | 0.702 | 0.648 | 0.607 | 0.57 | 0.542 | 0.74 |
NSP9 | 0.969 | 0.933 | 0.904 | 0.855 | 0.813 | 0.781 | 0.754 | 0.731 | 0.71 | 0.693 | 0.809 |
NSP13 | 0.999 | 0.998 | 0.997 | 0.996 | 0.995 | 0.994 | 0.993 | 0.992 | 0.991 | 0.99 | 0.994 |
NSP15 | 0.999 | 0.998 | 0.997 | 0.996 | 0.995 | 0.994 | 0.992 | 0.991 | 0.99 | 0.989 | 0.994 |
Protein | AA sequence length | Mutation count | Predicted mutations | True mutations | Precision | Recall | Selectivity |
NSP1 | 181 | 85 | 4 | [2, 4, 4, 4] | [0.5, 1, 1, 1] | [0.0223, 0.046, 0.046, 0.046] | 0.978 |
NSP3 | 1946 | 360 | 23 | [7, 12, 19, 23] | [0.3, 0.52, 0.83, 1.0] | [0.02, 0.033, 0.052, 0.063] | 0.99 |
NSP5 | 307 | 65 | 11 | [3, 8, 11, 11] | [0.27, 0.72, 1.0, 1.0] | [0.04, 0.12, 0.17, 0.17] | 0.967 |
NSP8 | 199 | 39 | 34 | [9, 20, 31, 34] | [0.26, 0.52, 0.91, 1.0] | [0.23, 0.51, 0.79, 0.87] | 0.844 |
NSP9 | 114 | 19 | 14 | [2, 9, 14, 14] | [0.14, 0.64, 1.0, 1.0] | [0.11, 0.47, 0.73, 0.73] | 0.87 |
NSP13 | 602 | 123 | 2 | [0, 1, 2, 2] | [0.0, 0.5, 1.0, 1.0] | [0.0, 0.008, 0.016, 0.016] | 0.99 |
NSP15 | 347 | 104 | 4 | [3, 4, 4, 4] | [0.75, 0.5, 1.0, 1.0] | [0.029, 0.0385, 0.0385, 0.0385] | 0.99 |
Protein | Predicted new mutations |
NSP1 | W160T, T169G, G179C |
NSP3 | P511R, R645S, S691T, S696T, T762P, T969P, S1037T, T1045P, T1183P, S1205P, R1340S, |
P1376R, P1557R, S1614T, S1681T, T1778P, S1806T, R1819S | |
NSP5 | I151S, H162V, G173H, T195G, I212S, R221T, L241R, I248S, L249R, I280S |
NSP8 | Q23D, V25Q, N27V, G28N, V32Q, V33Q, N42V, D51G, Q72D, V82Q, Q87D, N99V, N108V, |
G112N, V114Q, V129Q, N139V, G143N, Q157D, V158Q, N175V, N178V, N191V, Q197D | |
NSP13 | S262C |
NSP15 | D131E |
Protein | t_seq | t_mut | e_seq | e_mut | p_seq | p_mut | pt_mut | pe_mut | n_mut |
NSP1 | 864 | 375 | 2024 | 767 | 288 | 178 | 70 | 175 | 3 |
NSP3 | 13,347 | 2888 | 23,999 | 4195 | 840 | 559 | 295 | 541 | 18 |
NSP5 | 1002 | 456 | 2137 | 495 | 227 | 248 | 174 | 118 | 10 |
NSP8 | 577 | 394 | 852 | 384 | 125 | 110 | 40 | 83 | 24 |
NSP13 | 2674 | 680 | 6921 | 1063 | 472 | 218 | 111 | 217 | 1 |
NSP15 | 1232 | 512 | 2578 | 858 | 297 | 215 | 94 | 214 | 1 |