Research article Special Issues

Sentiment analysis for e-commerce product reviews by deep learning model of Bert-BiGRU-Softmax

  • Sentiment analysis of e-commerce reviews is the hot topic in the e-commerce product quality management, from which manufacturers are able to learn the public sentiment about products being sold on e-commerce websites. Meanwhile, customers can know other people's attitudes about the same products. This paper proposes the deep learning model of Bert-BiGRU-Softmax with hybrid masking, review extraction and attention mechanism, which applies sentiment Bert model as the input layer to extract multi-dimensional product feature from e-commerce reviews, Bidirectional GRU model as the hidden layer to obtain semantic codes and calculate sentiment weights of reviews, and Softmax with attention mechanism as the output layer to classify the positive or negative nuance. A series of experiments are conducted on the large-scale dataset involving over 500 thousand product reviews. The results show that the proposed model outperforms the other deep learning models, including RNN, BiGRU, and Bert-BiLSTM, which can reach over 95.5% of accuracy and retain a lower loss for the e-commerce reviews.

    Citation: Yi Liu, Jiahuan Lu, Jie Yang, Feng Mao. Sentiment analysis for e-commerce product reviews by deep learning model of Bert-BiGRU-Softmax[J]. Mathematical Biosciences and Engineering, 2020, 17(6): 7819-7837. doi: 10.3934/mbe.2020398

    Related Papers:

    [1] Quan Zhu, Xiaoyin Wang, Xuan Liu, Wanru Du, Xingxing Ding . Multi-task learning for aspect level semantic classification combining complex aspect target semantic enhancement and adaptive local focus. Mathematical Biosciences and Engineering, 2023, 20(10): 18566-18591. doi: 10.3934/mbe.2023824
    [2] Jun Wu, Xinli Zheng, Jiangpeng Wang, Junwei Wu, Ji Wang . AB-GRU: An attention-based bidirectional GRU model for multimodal sentiment fusion and analysis. Mathematical Biosciences and Engineering, 2023, 20(10): 18523-18544. doi: 10.3934/mbe.2023822
    [3] Dezheng Zhang, Shuai Chen, Na Zhou, Pu Shi . Location optimization of fresh food e-commerce front warehouse. Mathematical Biosciences and Engineering, 2023, 20(8): 14899-14919. doi: 10.3934/mbe.2023667
    [4] Qiao Xiang, Tianhong Huang, Qin Zhang, Yufeng Li, Amr Tolba, Isack Bulugu . A novel sentiment analysis method based on multi-scale deep learning. Mathematical Biosciences and Engineering, 2023, 20(5): 8766-8781. doi: 10.3934/mbe.2023385
    [5] Roman Gumzej . Intelligent logistics systems in E-commerce and transportation. Mathematical Biosciences and Engineering, 2023, 20(2): 2348-2363. doi: 10.3934/mbe.2023110
    [6] Feng Li, Mingfeng Jiang, Hongzeng Xu, Yi Chen, Feng Chen, Wei Nie, Li Wang . Data governance and Gensini score automatic calculation for coronary angiography with deep-learning-based natural language extraction. Mathematical Biosciences and Engineering, 2024, 21(3): 4085-4103. doi: 10.3934/mbe.2024180
    [7] Luqi Li, Yunkai Zhai, Jinghong Gao, Linlin Wang, Li Hou, Jie Zhao . Stacking-BERT model for Chinese medical procedure entity normalization. Mathematical Biosciences and Engineering, 2023, 20(1): 1018-1036. doi: 10.3934/mbe.2023047
    [8] Chao Che, Chengjie Zhou, Hanyu Zhao, Bo Jin, Zhan Gao . Fast and effective biomedical named entity recognition using temporal convolutional network with conditional random field. Mathematical Biosciences and Engineering, 2020, 17(4): 3553-3566. doi: 10.3934/mbe.2020200
    [9] Shuguang Wang . Point of Interest recommendation for social network using the Internet of Things and deep reinforcement learning. Mathematical Biosciences and Engineering, 2023, 20(9): 17428-17445. doi: 10.3934/mbe.2023775
    [10] Jiajia Jiao, Haijie Wang, Ruirui Shen, Zhuo Lu . Word distance assisted dual graph convolutional networks for accurate and fast aspect-level sentiment analysis. Mathematical Biosciences and Engineering, 2024, 21(3): 3498-3518. doi: 10.3934/mbe.2024154
  • Sentiment analysis of e-commerce reviews is the hot topic in the e-commerce product quality management, from which manufacturers are able to learn the public sentiment about products being sold on e-commerce websites. Meanwhile, customers can know other people's attitudes about the same products. This paper proposes the deep learning model of Bert-BiGRU-Softmax with hybrid masking, review extraction and attention mechanism, which applies sentiment Bert model as the input layer to extract multi-dimensional product feature from e-commerce reviews, Bidirectional GRU model as the hidden layer to obtain semantic codes and calculate sentiment weights of reviews, and Softmax with attention mechanism as the output layer to classify the positive or negative nuance. A series of experiments are conducted on the large-scale dataset involving over 500 thousand product reviews. The results show that the proposed model outperforms the other deep learning models, including RNN, BiGRU, and Bert-BiLSTM, which can reach over 95.5% of accuracy and retain a lower loss for the e-commerce reviews.


    With the rapid growth of e-commerce, online shopping can bring convenience and lower price products for consumers. Due to the information inconsistency between the real product quality and description provided by the seller, more and more consumers seek product information from e-commerce reviews involving many features (dimensions) of price, service, logistics and so on. E-commerce reviews plays a key role for customers in shopping decision process, meanwhile the manufacturers and E-commerce platforms can improve their products’ quality and on-line services based on the public opinions [1]. Pavlou and Dimoka (2006) utilized content analysis to quantify over 10,000 publicly available feedback text comments of 420 sellers in eBay’s online auction marketplace, which can be differentiated among sellers and prevent a market of lemon sellers for online market [2]. The sentiment analysis is the content analysis technique to distinguish the positive or negative attitudes of consumer from the related information appearing in different forms like BBS, blogs, Wiki or forum websites (Abbasi et al., 2008) [3].

    The challenges of sentiment analysis (SA) of E-commerce product reviews lie in dimensions mapping and sentiment word disambiguation. The dimension mapping problem refers to mapping opinioned text blocks with right dimensions. The sentiment word disambiguation problem refers the situation that a sentiment word may be connected with two or more dimensions. Therefore, sentiment analysis of E-commerce reviews is considered as a multi-dimensional classification process. As shown in Figure 1, it has three main classification levels: document-level, sentence-level, and aspect-level tasks. Document-level SA aims to classify opinion document as expressing positive or negative sentiment. Sentence-level SA aims to classify sentiment expressed in each sentence. Aspect-level SA aims to classify the sentiment with respect to the specific aspects of entities. Different approaches have been proposed to realize relative tasks. The first approach is called semantic-based approach (Zhang and Ye, 2010) [4], which performs sentiment analysis based on rules and sentiment polarity lexicons. The second approach is machine learning, which regards sentiment analysis as a binary or multi-class classification task and uses common classification methods(Chang and Lin, 2011) [5]. The latest sentiment analysis method is the deep learning models including the Convolutional Neural Network(CNN), Recurrent Neural Network(RNN) and other models (Hu and Li, 2017) [6].

    Figure 1.  Sentiment analysis process of E-commerce product reviews.

    The first step of sentiment analysis is to extract text features. The main feature selection methods are as following.

    Terms presence and frequency: It either gives the words binary weighting (zero when the word appears, or one at other condition) or uses term frequency weights to indicate the relative importance of features, which is based on the frequency counts of individual words or word n-grams [7].

    Parts of speech (POS): It is to find out important indicators of opinions such as adjectives, adverb and so on.

    Opinion words and phrases: These words are usually used to express opinions including positive or negative. On the other hand, some phrases express opinions without any emotional words.

    Feature selection attempts to reduce the dimension of data by picking from the original attribute set, which can be divided into lexicon-based methods that need human annotation, and statistical methods that are more frequently used. Lexicon-based approaches usually begin with a small set of ‘seed’ words and bootstrap this set through synonym detection or on-line resources to obtain larger lexicon [8]. The mutual information measure provides a formal way to model the mutual information between the features and the classes. Turney and Littman (2003) use the point wise mutual information (PMI) to extend the commendatory and derogatory emotional vocabulary in order to analyses the emotional tendency of English text. Yu and Wu [9] outperforms the PMI-based expansion methods by developing contextual entropy model to expand the set of seed words generated from a small corpus of stock market news. Their method combines both co-occurrence strength and contextual distribution, and need more useful emotion words and fewer noisy words.

    Feature transformation methods create the smaller set of features as the function of the original feature set. Latent Semantic Indexing (LSI) method transforms the text space into a new axis system which is the linear combination of original word features. Principal Component Analysis techniques (PCA) are used to achieve optimal goal [10]. There are other statistical approaches which could be used in Feature Selection (FS) like Hidden Markov Model (HMM) and Latent Dirichlet Allocation (LDA). Duric and Song [11] apply it to separate the entities in a review document from the subjective expressions that describe those entities in terms of polarities. HMM-LDA is a topic model that simultaneously models topics and syntactic structures in the collection of documents [12].

    The second step of sentiment analysis is sentiment classification. Sentiment classification can be divided into machine learning, lexicon based approach and deep learning approaches [13].

    Kim and Hovy (2004) analyzed the emotional tendency of text based on the synonyms, antonyms and hierarchies of the WordNet dictionary [14]. The acquisition of sentiment tendency about Chinese text mainly depended on HowNet [15]. YL Zhu and J Min (2006) used the semantic similarity and correlation of HowNet to calculate the similarity between the new word and the datum word, thereby to distinguish the sentiment tendency of the text [16]. G. Somprasertsri (2006) extracted the product feature and opinion of online customer reviews based on the consideration of syntactic and semantic information [17]. Sentiment analysis model of polarity lexicons is commonly used, but the existing sentiment vocabulary lexicons is limited (B Liu, 2014) [18].

    Compared with sentiment polarity lexicon, Machine learning methods have better advantages in the nonlinear and high dimensional pattern recognition problem. Pang (2002) firstly applied the machine learning method of N-Gram into the emotional analysis field, the experiment result show the N-Gram reached highest classification accuracy of 81.9% [19]. Go and Bhayani (2009) proposed the supervised learning method to classify the positive or negative emotional reviews extracted from Twitter by the Naive Bayes, maximum entropy and SVM analysis algorithms, which the experiment show 80% analysis accuracy [20]. Marco Guerini (2013) applied the sentiment analysis methods such as Naive Bayesian (NB), K-Nearest neighbor (kNN), Maximum entropy(ME), support vector machine (SVM) to analyses the reviews emotional tendency, which the support vector machine (SVM) is obviously superior to other methods in large training set and reach the highest accuracy of 83% [21]. Due to the feature selection will affect the performance of the machine learning method, Abinash Tripathy (2016) analyzed online comment reviews by the N-gram model combined machine learning methods, experimental results show that the SVM combined unigram, bigram and trigram features achieved the best classification results [22].

    With the development of deep learning research, the deep neural network has outstanding performance in natural language processing. Kim (2014) used Convolutional Neural Network (CNN) to solve sentiment classification problem [23]. Cho (2014) proposed Gated Recurrent Unit (GRU) to analyze long dependent contexts, which has significant improvements in various tasks [24]. Qu and Wang (2018) proposed the sentiment analysis model based on hierarchical attention network, which improved by 5% accuracy compared with the recurrent neural network [25]. Due to the reviews become larger, the traditional neural network models can’t completely capture the entire context of sentence or review, which the novel neural networks are needed by changing the different types of word embedding (i.e., GloVe or BERT) [26]. There are variants of RNNs exploited for sentiment analysis, such as Bidirectional- LSTM(BiLSTM) and Bidirectional-GRU(BiGRU) [27]. Trofimovich (2016) used LSTM (long short-term memory) to solve sentiment analysis problem [28]. The LSTM model was offered to make phrase-level sentiment classification centered on the regularization containing linguistics such as negativity, intensity and polarity [29]. The bidirectional LSTM is used to train the tagging text in order to handle syntactic and semantics of the Japanese language [30]. The BiGRU model with multi-input and output was proposed for phrase-level polarity classification [31].

    This paper researches the deep learning model of Bert-BiGRU-Softmax for sentiment analysis of the E-commerce product reviews, in which the Bert model makes the features exaction by mapping each sentence into the proper dimension, Bi-GRU model solves the sentiment word disambiguation problems by using the multiple clause-rules recognition, and the Softmax model with attention mechanism makes the multi-dimensional sentiment analysis by calculating the sentiment polarity at specific dimension of the E-commerce product reviews.

    The BERT (Bidirectional Encoder Representations from Transformers) is the transfer learning -pre-training neural network model (Devlin, 2019) [32]. The BERT applies the bidirectional training of transformer to learn contextual relations between words and tokens (Tok1, …, TokM) in text. The transformer includes two separate mechanisms that an encoder (E1, …, EN) reads the text input and a decoder that makes prediction for the task. As opposed to single-direction models of RNN(Recurrent Neural Network) and LSTM (Long Short-Term Memory), which read the text input sequentially (left-to-right or right-to-left), the transformer encoder(T1, …, TN) reads the entire bidirectional sequence of words at once. It has caused a stir in the machine learning community by presenting state-of-the-art results in a wide variety of NLP (natural language processing) tasks and others. Figure 2 shows the multi-level structure chart of BERT model.

    Figure 2.  The multi-level structure of Bert model.

    BERT alleviates the bidirectional constraints by using “masked language model” (MLM) pre-training natural language process. The masked language model randomly masks some of the tokens from the input, and the objective is to predict the original vocabulary of the masked word based on its context. This bidirectional training model allows the BERT have deeper sense of language context and learn the context of word by all of its surrounding.

    The GRU is a specific model of recurrent neural network that uses connections through a sequence of nodes to perform machine learning tasks associated with memory and clustering, which allows GRU to carry forward information over many time periods in order to influence a following time period. GRU can be considered as a variation on the LSTM because both are designed similarly and produce equally excellent results, which gated recurrent units help to adjust neural network input weights to solve the vanishing gradient problem [33]. The structure of GRU is shown as Figure 3.

    Figure 3.  The structure of GRU model.

    As the refinement of recurrent neural network, GRU has what’s called an update gate zt and a reset gate rt. Using input vectors xt and output vectors ht, the model refines outputs by controlling the ht−1 flow of information through the model. Like other kinds of recurrent network models, GRU with gated recurrent units can retain information over a period of time t that is why the simplest way to describe these technologies as the “memory-centered” type of neural network. By contrast, other types of neural networks without gated recurrent units often don’t have the ability to retain information.

    Without attention mechanism, translation relies on reading a complete sentence and compressing all information into a fixed-length vector, which the sentence with hundreds of words represented by several words will surely lead to information loss and inadequate translation, etc. Attention partially fixes this problem. Bahdanau (2014) proposes attention mechanism to memorize the long source sentences in Neural Machine Translation (NMT) and NLP [34]. Attention probability distribution matrix of the input text sequence is obtained by attention mechanism. The weight of text feature information is calculated by the matrix, which reduces the missing and redundant information during feature extracting. It allows machine translator to look over all information that the original sentence holds, then generate the proper word according to current word it works on and the context, the structure of the attention mechanism model is shown as Figure 4.

    Figure 4.  The structure of Attention Mechanism model.

    This paper researches Bert-BiGRU-Softmax model for sentiment analysis on the e-Commerce product quality reviews. The sentiment Bert model, as input layer, makes the feature extraction at pre-processing-phase. The hidden layer of Bidirectional GRU performs dimension oriented sentiment classification by using bidirectional long short-term memory and gated recurrent units to hold the long-term dependencies, which are inherent in the text regardless of lengths and occurrences. The output layer of Softmax calculates the sentiment polarity by pooling to smaller weighted dimensions according to the attention mechanism. The Figure 5 is the structure chart of the Bert-BiGRU-Softmax model.

    Figure 5.  Information flow chart of the Bert-BiGRU-Softmax model.

    Dimensions mapping of the e-commerce product quality reviews are typically domain dependent problem. Since the dimension may appear in many product categories, we have to determine the right dimension for a given review about a certain e-Commerce product category.

    The Bert transforms the e-commerce reviews into a numerical matrix, where each column represents the identified features and each row expresses a particular review. The matrix is given as input of Bert in order to train the model by two training strategies of the MLM and next sentence prediction, with the best performance of the combined loss function [35]. When training the BERT model, sentences segmentation procedure is used for splitting a long sentence into several shot text blocks, and word sequences are regarded as nodes of text blocks, which is more efficiently to map those short sentences into corresponding dimensions. The transformer receives the pairs of sentence as input and learns to predict the second sentence in the pair on the assumption that the random sentence will be disconnected from the first sentence, which can be taken as explicit features in dimension mapping. The training process of Bert model is shown as Figure 6.

    Figure 6.  The pre-trained model and transformer coding unit of Bert model.

    Aiming at sentiment task and complex emotional features, the pre-training approaches of the sentiment Bert model focuses on learning sentiment features by context-based word sentiment prediction task, which classifies the sentiment of masked words to acquire the textual representation biased towards sentiment features.

    The sentiment Bert model can be used effectively to learn feature extraction over a variable-length sequence S by learning the distribution over the next input text vector. Given a review sentence S, we can obtain its category c and the dimension set Dc of category c directly. For each word wi (w1, w2, w3,, wm) in the review sentence S and a dimension dj, we assign a probability score which describe the probability p(S) of word wi belongs to category dj in the e-Commerce product quality reviews (as shown in Eq (1).

    p(S)=p(w1,w2,,wm)=mi=1p(wi|w1,w2,,wi1) (1)

    Transformer add sequence information by word positional embedding (PE) Eqs (2) and (3).

    PE(pos,2i)=sin(pos/100002i/dmodel) (2)
    PE(pos,2i+1)=cos(pos/100002i/dmodel) (3)

    where d is 64, the text sequence be represented as the 512 characters, 2i is even position and 2i+1 is odd position in the given sequence of input vector.

    When the Transformer make the feature extraction as x[CLS] and x[MASK]from the special words of w[CLS] and w[MASK]in the S sequence, the Bert loss function takes into consideration only the prediction of the masked values and ignores the prediction of the non-masked words, which are represented as Eqs (4, 5).

    loss=iL(xi[MASK])F(xi[MASK]) (4)
    F(xi[MASK])={k  if  xR1  if  xR (5)

    where R is the emotional words lexicon of e-Commerce product quality, L(xi[MASK]) is the loss of the masked words xi[MASK] in prediction process of input sequence sentence, F(xi[MASK]) is the weight of emotional and non-emotional words. Finally we map a review sentence S with the dimension dj that has the maximum probability score with S. The sentiment Bert model for Sentence information processing involves the following steps:

    Sentiment Bert model:
    Begin:
    In this step, each target sentence of the e-Commerce product quality review is represented by the content of the sentence. Before feeding word sequences into BERT, 15% of the words in each sequence are replaced by a [MASK] token. The model attempts to predict the original value of the masked words, based on the context provided by the other, non-masked, words in the sequence. We use this representation to learn the internal features of the sentence.
    Sentence contextual information processing:
    Here, we adopt the method that simply replace each word of the target sentence with special characters to build the meaningless string of [MASK], which is replaced by a [MASK] string to obtain the contextual information of each sentence. For example, “The mobile phone was [MASK] and the service was good”, the unmasked description of “the service was good” is the positive assessment of mobile phone, conjunction “and” express the similar sentences, so the Masked words can be deduced as the positive expression.
    Sentiment Bert integrate processing:
    This step integrates the content and contextual information of the sentence to construct the masked sentence model. It implements the integration by inputting the above two training samples together to the Bert fine-tuning procedure to train the masked sentence model for recognition task. For each dimension, we put the top_20 words with highest scores into the candidate sentiment word set. As consequence, Bert model increase the sentiment analysis accuracy of reviews by bidirectional context awareness.
    End

    Sentiment lexicons of the sentiment polarity in e-commerce reviews are widely used in sentiment analysis. However, those lexicons did not distinguish the polarities by one word on different dimensions of product. This paper uses multiple rules to enhance the performance of sentiment analysis. For example, when sentences with positive strong rule, “It is so famous brand of mobile phone that the customer would like to buy Huawei.”, which is labelled as positive without other factors. The emotion analysis has the relationship with the previous state and next state of the text sequence at the same time. GRU is one-way neural network structure of this fixed-length context vector design, which will forget the part information during the whole input context process.

    This paper proposes the bidirectional gated recurrent unit (BiGRU) to perform natural language recognition tasks associated with memory and clustering by both forward and reverse information (Grü.er-Sinopoli and Thalemann [36]). The idea of BiGRU is to split the regular GRU neurons into a forward state and a backward state. Based on the co-occurrence information with seed sentiment words, we could enlarge the original seed words iteratively. Besides, BiGRU connects two hidden layers with opposite transmission directions to the same output layer, so that the output layer obtains information both from past and future states. Therefore, the BiGRU is able to learn the information from two different directions, and make more accurate analysis of the e-commerce product quality reviews based on clause-rules recognition rules. The structure of BiGRU is shown in Figure 7.

    Figure 7.  The structure of Bidirectional GRU network.

    BiGRU has the given sequence of input vectors <x1, x2,, xt> (xt represents the concatenations of input features), setting unit computes the corresponding hidden activations <h1, h2,, ht> , and outputs a vector sequence <y1, y2,, yt> from the input data. At time t, the current hidden state of ht is determined by three parts including input vectors <x1, x2,, xt> , the forward hidden state _ht and the backward hidden state _ht at the same time. The reset gate (rt) controls the extent to which status information from previous time is ignored; the smaller the rt, the more previous status information is ignored. The update gate (zt) controls the degree to which new input information is received in the cell state. The symbol ⊗ is the element-wise multiplication. σ is the sigmoid function, and tanh is the tanh function. The hidden state ht, update gate (zt) and the reset gate (rt) of BiGRU are calculated by Eqs (6)–(10).

    zt=(wz[ht1,xt]) (6)
    rt=(wr[ht1,xt]) (7)
    _ht=GRU(xt,_ht1) (8)
    _ht=GRU(xt,_ht1) (9)
    ht=wt_ht+vt_ht+bt (10)

    wz and wr are the weight matrices vetors of update gate and reset gate. vt and wt are the weight matrices of the forward and backward hidden state. bt is the corresponding offset of hidden state at time t. the symbol * denotes the element-wise multiplication, and the symbol [] indicates that two vectors are connected to each other. The approaches of BiGRU model involve the following steps:

    BiGRU Model:
    Begin:
    Update gate: This gate calculate the update gate zt at time step t by the following steps:
    (1) The input xt is multiplied by a weight wzx;
    (2) The previous output ht-1 which hold information from previous units multiplied by weight wzh;
    (3) Both are added together and sigmoid function apply to squeeze the output between 0 and 1.
    Reset gate: This gate calculate the update gate rt at time step t by the following steps:
    (1) The input xt is multiplied by a weight wxr;
    (2) The forward output _ht1 which hold forward units information multiplied by weight whr;
    (3) Both are added together and sigmoid function [0, 1] apply to squeeze the output.
    This gate has the opposite functionality in comparison with the update gate since it is used by the model to decide how much of the past information to forget.
    Current memory content: The calculation involves the following steps:
    (1) The input xt is multiplied by a weight wx
    (2) Apply element-wise multiplication to the reset gate rt and the backward output layer _ht1; this allows to pass only the relevant past information.
    (3) Both are added together and tanh function is applied.
    Final memory at current time step: Last but not least the unit has to calculate the ht vector which holds information for the current unit and it will pass it further down to the network. Key role in this process plays the update gate zt.
    (1) Apply element-wise multiplication to the update gate zt and ht.
    (2) Apply element-wise multiplication to one minus the update gate zt and ht.
    (3) Both are added together.
    End

    The output layer of the Softmax makes the positive scores, negative scores, and neutral scores by fusing the different semantic features of the Bert-BiGRU models, the attention mechanism calculate the linear weighted sum of all the positive, negative, and neutral polarities of the sentence sequences from last hidden state level, which the attention probability and sentiment polarity weights of reviews are calculated by the Softmax layer(as shown in Figure 8).

    Figure 8.  The flow chat of Softmax with attention mechanism.

    The detailed approaches of the Softmax with attention mechanism [37] are as following:

    (1) After receiving the feature vector of attention layer, attention function calculates the weight matrix wa:

    ut=tanh(waht+ba) (11)
    αt=exp(uTtua)texp(uTtua) (12)
    sit=ni=1aithit (13)

    where tanh is the sigmoid function, the hidden state ht is fed into learning function (11) to produce probability vector ut, the vector αt is computed as average weight.

    (2) Make the normalization operation of Softmax by Eq (14):

    y=softmax(wasit+ba) (14)

    where y is the sentiment analysis result, ba is the corresponding offset of output layer.

    In order to evaluate the sentimental analysis efficiency of the Bert-BiGRU-Softmax Model, this paper conduct experiments on multi-source datasets with various domains by the Python 3.5 and Tensorflow framework at PyCharm IDE. The first dataset is IMDB, which is used and prolonged extensively as a benchmark dataset [38]. The dataset contains 25,000 tweets with the polarity of 12,500 positives, and 12,500 negatives about the movie reviews. The second dataset is the Chinese emotional corpus known as ChnSentiCorp [39], which has abundant Sentiment Corpus including ChnSentiCorpHtl, ChnSentiCorpMov and so on. Table 1 shows the experimental results of proposed emotional analysis model, the multi-Bi-LSTM model and the Weight W2V-ATT -LSTM (WWAL) model on ChnSentiCorp-Htl-ba-6000 and COAE2014-task4.

    Table 1.  The Accuracy and F-score of different models.
    Datasets Model Accuracy F1
    COAE2014-task4 CNN[40]
    Multi-Bi-LSTM[40]
    Bert-BiGRU
    0.8820
    0.8840
    0.8903
    0.8760
    0.8840
    0.8857
    ChnSentiCorp-Htl-ba-6000 WWAL(WeightW2V-ATT-LSTM) [41] 0.8727 0.8719
    Bert-BiGRU 0.8831 0.8816

     | Show Table
    DownLoad: CSV

    In first experiment of COAE2014-task4, the accuracy and F1 measure of Bert-BiGRU-Softmax is (0.8903, 0.8857), which is 1% larger than Multi-Bi-LSTM (0.8840, 0.8840). The second experiment of ChnSentiCorp-Htl-ba-6000, Bert-BiGRU-Softmax performs better accuracy (0.8831) and F1 measure (0.8816), which are slightly higher than accuracy (0.8727) and F1 measure (0.8719) of WWAL (as shown in Figure 9). When the feature classifier transfers from different topics, the machine learning techniques of sentiment classifier is severely dependent on its domains. The Multi-Bi-LSTM, Weight W2V-ATT-LSTM and Bert-BiGRU-Softmax models are suitable for sentiment analysis, which Bert-BiGRU-Softmax has better performance by Bert model as pretraining reviews feature extractor.

    Figure 9.  The Accuracy of WWAL, Multi-Bi-LSTM and Bert-BiGRU-Softmax.

    This paper crawls and analyzes a large-scale dataset with 150 predefined dimensions of 500 thousands E-commerce reviews about mobile phone products from Sunning, Taobao and other e-commerce websites, which cover almost all the aspects of different products with their polarity concerning positive and negative, handing the dimensions “quality”, “logistic” and “service”. We use the deep learning models of RNN, LSTM, GRU, BiGRU, BiLSTM, Bert-BiLSTM and Bert-BiGRU- Softmax to make sentiment analysis of e-commerce product quality reviews from dimensions of brand, ratings, price and others. Table 2 is the dataset about the mobile phone quality reviews. “trust”, “wonderful” and “good” have top-most scores. “poorly”, “frustrated” and “unsatisfied” have lowest scores. The high scores for “wonderful” and “good” could be the new brand phones. Also, the highest score for “trust” among all the emotions shows that the reviewers are conviction and trust the product.

    Table 2.  The sentiment reviews about mobile phone quality on e-Commerce websites.
    This mobile phone provided bigger screen and powerful battery, the only drawback for me is that it’s a little thick...
    This mobile phone isn’t famous brand, but price is cheaper than others and the battery life is about 2300 mAh. The cost performance is very high!
    This mobile phone has wonderful software function and appearance which could be better. My friend recommended it to me, Although charge is...
    I admit, mobile phone has good camera system with high pixel.<br/>If anything, it helps you make beautiful picture what you want<br/>.
    Poor battery efficiency and signal, I have to charge this mobile phone so many times in a day! It’s so turgid that I put it down in frustration....
    The particularly nice mobile phone is received, the audio sounds good! courier delivery it to my home, give five star rating for it...

     | Show Table
    DownLoad: CSV

    This paper compares the performances of RNN, LSTM, GRU, BiGRU, BiLSTM, Bert-BiLSTM and Bert-BiGRU-Softmax models according to the accuracy and loss value, which are based on the 400 thousands original data from the web reviews corpus, and split into training and test datasets in experiments. The hyper parameters of Bert-BiGRU-Softmax model are as following:

    Table 3.  The hyper parameters of Bert-BiGRU-Softmax model.
    Hyper parameter Value
    The nodes of hidden layer network 100
    Loss function Categorical_Crossentropy
    Optimizer Adam
    Dimensions of word vector 768
    Epoch 10

     | Show Table
    DownLoad: CSV

    The accuracy and loss value of models on the different datasets are shown in Tables 4 and 5. The accuracy and loss cure of models on the different datasets are shown as Figures 10 and 11.

    Figure 10.  The Accuracy of RNN, LSTM, GRU, BiGRU, Bert-BiLSTM and Bert-BiGRU.
    Figure 11.  The Loss of RNN, LSTM, GRU, BiGRU, Bert-BiLSTM and Bert-BiGRU.

    According to the accuracy cure, the RNN model reaches the highest accuracy value of 0.895 at the 3rd epoch, and then the accuracy value of RNN becomes volatility at the over-fitting status. The BiGRU model reaches the highest accuracy value of 0.922 at the 8th epoch, and then the accuracy value ascends to the over-fitting status. The LSTM model reaches the highest accuracy value of 0.921 at the 2th epoch. The GRU model becomes over-fitting status after getting highest accuracy of 0.925 at the 5th epoch. The BiLSTM model reaches the highest accuracy value of 0.928 at the 5th epoch. The Bert-BiLSTM model reaches the highest accuracy value of 0.931 at the 9th epoch, and then the accuracy value of LSTM descend to the over-fitting status. Comparing with the BiGRU model, the Bert-BiGRU-Softmax model reaches the highest accuracy value of 0.955 at the 8th epoch and reduces computational overhead, which is superior to the competitor algorithms.

    Table 4.  The accuracy of different deep learning models.
    Model Training Epochs
    1 2 3 4 5 6 7 8 9 10
    RNN 0.867 0.876 0.895 0.893 0.886 0.894 0.893 0.877 0.894 0.893
    LSTM 0.911 0.921 0.913 0.904 0.917 0.911 0.908 0.904 0.886 0.901
    GRU 0.904 0.872 0.918 0.911 0.925 0.921 0.915 0.901 0.919 0.917
    BiLSTM 0.919 0.901 0.905 0.924 0.928 0.912 0.917 0.926 0.923 0.926
    BiGRU 0.890 0.910 0.918 0.913 0.921 0.921 0.921 0.922 0.921 0.922
    Bert-BiLSTM 0.881 0.906 0.927 0.922 0.911 0.919 0.928 0.927 0.931 0.925
    Bert-BiGRU 0.912 0.940 0.934 0.939 0.929 0.944 0.936 0.955 0.952 0.947

     | Show Table
    DownLoad: CSV
    Table 5.  The loss value of different deep learning model.
    Model Training Epochs
    1 2 3 4 5 6 7 8 9 10
    RNN 0.342 0.329 0.303 0.300 0.346 0.359 0.378 0.427 0.410 0.428
    LSTM 0.239 0.236 0.273 0.268 0.273 0.297 0.326 0.375 0.409 0.367
    GRU 0.251 0.316 0.242 0.263 0.245 0.297 0.299 0.339 0.322 0.331
    BiLSTM 0.231 0.242 0.244 0.243 0.246 0.293 0.292 0.328 0.313 0.313
    BiGRU 0.287 0.235 0.240 0.362 0.257 0.307 0.366 0.374 0.439 0.427
    Bert-BiLSTM 0.349 0.291 0.223 0.356 0.389 0.343 0.343 0.395 0.412 0.385
    Bert-BiGRU 0.262 0.192 0.234 0.239 0.325 0.269 0.235 0.211 0.291 0.252

     | Show Table
    DownLoad: CSV

    According to the loss cure of different models, the extraction ability of RNN is the worst of all models. Comparing with LSTM model, GRU model has better performance. BiGRU model will generate large computational overhead and poor accuracy performance than BiLSTM model. Bert-BiLSTM model can reaches higher performance than that of BiGRU. The proposed Bert-BiGRU -Softmax model reaches the highest accuracy value of 0.955, which is improved by nearly 3.6% of Bert-BiLSTM at the training datasets. By comparing the models from different types of comment sets, experiments show that the Bert-BiGRU-Softmax model can change dimension flexibly and improve the accuracy of feature extraction.

    Bert-BiGRU- Softmax model is applied for the sentiment analysis of mobile phone product quality reviews, which obtains each type of polarities (positive, neutral and negative) including the attention probability distribution from input text sequences and calculates the weights of customer reviews emotional scores according to the different dimensions (camera, screen, system, audio, battery, signal, service, price and surface). The results of sentiment analysis are shown in Table 6.

    Table 6.  Sentiment analysis results of phone reviews by Bert-BiGRU-Softmax model.
    brand attitude camera screen system audio battery signal service price surface
    AP positive 0.734 0.549 0.784 0.551 0.679 0.389 0.726 0.521 0.695
    neutral 0.153 0.275 0.109 0.163 0.194 0.253 0.149 0.279 0.139
    negative 0.113 0.176 0.107 0.286 0.127 0.358 0.125 0.200 0.166
    HW positive 0.889 0.694 0.690 0.594 0.686 0.717 0.618 0.629 0.689
    neutral 0.094 0.172 0.189 0.159 0.159 0.156 0.106 0.194 0.171
    negative 0.017 0.134 0.121 0.247 0.155 0.127 0.276 0.177 0.140
    Mi positive 0.701 0.642 0.538 0.623 0.572 0.601 0.546 0.780 0.740
    neutral 0.106 0.183 0.224 0.151 0.179 0.213 0.193 0.117 0.169
    negative 0.193 0.175 0.238 0.226 0.249 0.186 0.261 0.103 0.091
    Vi positive 0.730 0.656 0.596 0.567 0.663 0.549 0.552 0.607 0.737
    neutral 0.142 0.185 0.207 0.161 0.163 0.332 0.177 0.208 0.137
    negative 0.128 0.159 0.197 0.272 0.174 0.119 0.271 0.185 0.126

     | Show Table
    DownLoad: CSV

    After the comprehensive compared, AP mobile phone gets higher satisfaction at the features of system, hardware and service than the other three brands from consumers, while gets the lowest satisfaction of signal and price problem in the four brand products. Mi mobile phone has the highest satisfaction from consumers at the feature of price, surface and audio, while the features of system, battery audio are lower than the other three brands. Customers have higher satisfaction to HW mobile phone at the features of camera, screen, battery and signal than the other three brands. Meanwhile there is no obvious low level of satisfaction at other features. The Vi mobile phone has not particularly prominent and disadvantage, which all aspects are relatively balanced in the whole. In the conclusions, HW mobile phone has the highest satisfaction degree from consumer, which AP, Mi and Vi brand are at 2nd, 3rd and 4th place according to the consumer satisfaction. The analysis results reflect the performance of four different brand mobile phones in accordance with the actual market conditions.

    E-commerce reviews reveal the customers’ opinions about the products selling at the e-commerce platforms, which are helpful for manufacturers to improve their products quality. This paper proposes the deep learning model of Bert-BiGRU-Softmax to deal with the sentiment word disambiguation and the sentiment polarity problem, which uses the Bert model to extract features selection from e-commerce reviews at input layer, the hidden layer of BiGRU model with attention mechanism to obtain semantic codes including the attention probability of input layer, and Softmax model as the output layer to classify the sentiment tendency of the e-commerce reviews. The experiment analyses the large-scale review datasets from the Sunning, Taobao websites. The experimental results show that Bert-BiGRU-Softmax model has better performance than RNN, BiGRU and Bert-BiLSTM, which improves the accuracy at least 3% on sentiment analysis of the e-commerce product quality reviews.

    This work is supported by the Zhejiang Provincial Natural Science Foundation of China (Grant No. LY18G020009), the Zhejiang Provincial Philosophy and Social Science Planning Project of China (Grant No. 20NDJC096YB), the National Key Research and Development Project of China (Grant No. 2017YFF0209605). We are thankful to Feng Qian and Hu Gu providing expertise and assistance in the research, who work in quality inspection and technical supervision institution of Hangzhou in China.

    The authors declare no conflict of interest.



    [1] P. Sasikala, L. M. I. Sheela, Sentiment analysis of online product reviews using DLMNN and future prediction of online product using IANFIS, J. Big Data, 7 (2020), 33-53. doi: 10.1186/s40537-020-00308-7
    [2] P. A. Pavlou, A. Dimoka, The nature and role of feedback text comments in online marketplaces: Implications for trust building, price premiums, and seller differentiation, Inf. Syst. Res., 17 (2006), 392-414. doi: 10.1287/isre.1060.0106
    [3] A. Abbasi, H. Chen, A. Salem, Sentiment analysis in multiple languages: feature selection for opinion classification in web forums, ACM Trans. Inf. Syst., 26 (2008), 12-21.
    [4] Z. Zhang, Y. E. Qiang, Literature review on sentiment analysis of online product reviews, J. Manage. Sci. China, 13 (2010), 84-96.
    [5] C. Chang, C. J. Lin, LIBSVM: A library for support vector machines, ACM Trans. Intel. Syst. Technol., 2 (2011), 1-27.
    [6] F. Hu, L. Li, Z. L. Zhang, et al, Emphasizing essential words for sentiment classification based on recurrent neural networks, J. Comput. Sci. Technol., 32 (2017), 785-795. doi: 10.1007/s11390-017-1759-2
    [7] Y. Mejova, P. Srinivasan, Exploring feature definition and selection for sentiment classifiers. In Proc. 5th Int. Conf. weblogs social media, Barcelona, Catalonia, Spain, (2011), 17-21.
    [8] W. Casey, G. Navendu, A. Shlomo, Using appraisal groups for sentiment analysis. In Proc. ACM SIGIR Conf. Inform. Knowl. Manag., (2005), 625-31.
    [9] L.-C. Yu, H.-L. Wu, P.-C. Chang, H.-S. Chu, Using a contextual entropy model to expand emotion words and their intensity for the sentiment classification of stock market news, Knowl. Based Syst., 41 (2013), 89-97. doi: 10.1016/j.knosys.2013.01.001
    [10] I. Jolliffe, N. A. Jolliffe, Principal Component Analysis. Springer Series in Statistics, 2nd edition, Springer-Verlag, New York, 2002.
    [11] D. Adnan, F. Song, Feature selection for sentiment analysis based on content and syntax models, Decis. Support Syst., 53 (2012), 704-11. doi: 10.1016/j.dss.2012.05.023
    [12] L. G. Thomas, S. Mark, M. B. David, B. T. Joshua, Integrating topics and syntax, Adv. Neural Inf. Process. Syst., (2005), 537-544.
    [13] A. Yadav, D. K. Vishwakarma, Sentiment analysis using deep learning architectures: a review, Artif. Intell. Rev., 53 (2020), 4335-4385. doi: 10.1007/s10462-019-09794-5
    [14] S. Kim, E. Hovy, Determining the sentiment of opinions. In Proc. Conf. Comput. Linguis., Geneva, Switzerland, (2004), 1367-1373.
    [15] P. D. Turney, M. L. Littman, Measuring praise and criticism: Inference of semantic orientation from association, ACM Trans. Inf. Syst., 21 (2003), 315-346. doi: 10.1145/944012.944013
    [16] Y. L. Zhu, J. Min, Y. Q. Zhou, Semantic orientation computing based on HowNet, J. Chin. Inf. Process., 20 (2006), 14-20.
    [17] G. Somprasertsri, P. Lalitrojwong, Mining feature-opinion in online customer reviews for opinion summarization, J. Univ. Comput. Sci., 16 (2010), 938-955.
    [18] B. Liu, Sentiment analysis and opinion mining, Synth. Lect. Hum. Lang. Technol., 5 (2012), 1-167.
    [19] B. Pang, L. Lee, S. Vaithyanathan, Thumbs up? sentiment classification using machine learning techniques. In Proc. Conf. Empir. Methods Nat. Lang. Process., Philadelphia, USA, (2002), 79-86.
    [20] A. Go, R. Bhayani, L. Huang, Twitter sentiment classification using distant supervision, CS224N Project Report, Stanford, 1 (2009), 1-12.
    [21] M. Guerini, L. Gatti, M. Turchi, Sentiment analysis: How to derive prior polarities from SentiWord Net, arXiv preprint arXiv: 1309.5843.
    [22] A. Tripathy, A. Agrawal, S. K. Rath, Classification of sentiment reviews using n-gram machine learning approach, Expert Syst. Appl., 57 (2016), 117-126. doi: 10.1016/j.eswa.2016.03.028
    [23] Y. KIM, Convolutional neural networks for sentence classification, arXiv preprint arXiv: 1408.5882.
    [24] K. Cho, B. van Merrienboer, D. Bahdanau, On the properties of neural machine translation: encoder-decoder approaches. In Proce. Eighth Workshop Syntax, Semant. Struct. Stat. Trans., Doha, Qatar, (2014), 103-111.
    [25] Z. Qu, Y. Wang, X. Wang, A Hierarchical attention network sentiment classification algorithm based on transfer learning, J. Comput. Appl., 38 (2018), 3053-3056.
    [26] J. Pennington, R. Socher, C. D. Manning, GloVe: Global vectors for word representation, Proc. Int. Conf. Empir. Methods Nat. Lang. Process., (2014), 147-168.
    [27] F. Abid, M. Alam, M. Yasir, C. Li, Sentiment analysis through recurrent variants latterly on convolutional neural network of twitter, Future Gener. Comput. Syst., 95 (2019), 292-308. doi: 10.1016/j.future.2018.12.018
    [28] J. Trofimovich, Comparison of neural network architectures for sentiment analysis of Russian tweets, In Proc. Int. Conf. Dialogue 2016, RGGU, (2016).
    [29] Q. Qian, M. Huang, J. Lei, X. Zhu, Linguistically regularized LSTMs for sentiment classification. In Proc. 55th Ann. Meet. Assoc. Comput. Ling., 1 (2016), 1679-1689.
    [30] L. Nio, K. Murakami, Japanese sentiment classification using bidirectional long short-term memory recurrent neural network. In Proc. 24th Ann. Meeti. Assoc. Nat. Lang. Process., Okayama, Japan, (2018), 1119-1122.
    [31] M.-L. Zhang, Z.-H. Zhou, A review on multi-label learning algorithms, IEEE Trans. Knowl. Data Eng., 26 (2014), 1819-1837. doi: 10.1109/TKDE.2013.39
    [32] J. Devlin, M. W. Chang, K. Lee, K. Toutanova, BERT: Pre-Training of deep bidirectional transformers for language understanding, In Proc. 2019 Conf. North Am. Chapter Assoc. Comput. Linguist. Human Lang. Tech., (2019), 4171-4186.
    [33] S. Seo, C. Kim, H. Kim, et al, Comparative study of deep learning based sentiment classification, IEEE Access, 8 (2020), 6861-6875. doi: 10.1109/ACCESS.2019.2963426
    [34] B. Dzmitry, C. Kyunghyun, B. Yoshua, Neural machine translation by jointly learning to align and translate, In Proc. Int. Conf. Learn. Representations 2015, San Diego, CA, USA, (2015), 1-15.
    [35] D. Li, F. Wei, C. Tan, Adaptive recursive neural network for target-dependent twitter sentiment classification, Baltimore: Assoc. Comput. Ling.-ACL, (2014), 49-54.
    [36] S. M. Grü.er-Sinopoli, C. N. Thalemann, Bidirectional recurrent neural networks, IEEE Trans. Signal Process., 11 (1997), 2673-2681
    [37] S. Zhang, X. Xu, Y. Pang, J. Han, Multi-layer Attention Based CNN for Target-Dependent Sentiment Classification, Neural Process. Lett., 51 (2020), 2089-2103 doi: 10.1007/s11063-019-10017-9
    [38] K. Saranya, S. Jayanthy, Onto-based sentiment classification using machine learning techniques. In Proc. 2017 Int. Conf. Innovations Inf., Embed. Commun. Syst., (2017), 1-5.
    [39] S. Tan, J. Zhang, An empirical study of sentiment analysis for Chinese documents, Expert Syst. Appl., 34 (2008), 2622-2629. doi: 10.1016/j.eswa.2007.05.028
    [40] W. Li, F. Qi, Sentiment analysis based on multi-channel bidirectional long short term memory network, J. Chin. Inf. Process., 33 (2019), 119-128.
    [41] Y. Yang, H. Yuan, Y. Wang, Sentiment analysis method for comment text, J. Nanjing Univ. Sci. Tech., 43 (2019), 280-285.
  • This article has been cited by:

    1. K. Mouthami, S. Anandamurugan, S. Ayyasamy, 2022, BERT-BiLSTM-BiGRU-CRF: Ensemble Multi Models Learning for Product Review Sentiment Analysis, 978-1-6654-8271-4, 1514, 10.1109/ICECA55336.2022.10009469
    2. Semmy Wellem Taju, Edson Yahuda Putra, Green Ferry Mandias, 2022, Sentiment Identification System for E-Commerce Mobile App Reviews Using Single Layer Neural Network, 978-1-6654-5395-0, 1, 10.1109/ICORIS56080.2022.10031580
    3. Shoucheng Shen, Jinling Fan, Emotion Analysis of Ideological and Political Education Using a GRU Deep Neural Network, 2022, 13, 1664-1078, 10.3389/fpsyg.2022.908154
    4. Hong Peng, Zhenwei Zhang, Hua Liu, Ahmed Farouk, A Sentiment Analysis Method for Teaching Evaluation Texts Using Attention Mechanism Combined with CNN-BLSTM Model, 2022, 2022, 1875-919X, 1, 10.1155/2022/8496151
    5. Yunhong Rong, Bohong Liu, 2022, Research on Opinion Mining for Sentiment Classification of Micro-blog Text Based on DeBERTa, 978-1-6654-7896-0, 5337, 10.1109/CCDC55256.2022.10033688
    6. Li Zhang, Guan Huang, Mukhtaj Khan, A Hybrid Approach for Network Rumor Detection Based on Attention Mechanism and Bidirectional GRU Model in Big Data Environment, 2022, 2022, 1875-905X, 1, 10.1155/2022/4185208
    7. Jun Liu, Baixiang Pan, Xiaheng Zhang, Doudou Li, Sang-Bing Tsai, Mobile E-Commerce Information System Based on Industry Cluster under Edge Computing, 2021, 2021, 1875-905X, 1, 10.1155/2021/7930799
    8. Mengjuan Zha, Changping Hu, Yu Shi, Sentiment lexicon construction for Chinese book reviews based on ultrashort reviews, 2022, 40, 0264-0473, 221, 10.1108/EL-07-2021-0147
    9. Shiguang Sun, Li Li, 2022, Chapter 76, 978-3-030-89507-5, 590, 10.1007/978-3-030-89508-2_76
    10. Mouthami Kuppusamy, Anandamurugan Selvaraj, A novel hybrid deep learning model for aspect based sentiment analysis, 2023, 35, 1532-0626, 10.1002/cpe.7538
    11. Kushal Devgun, 2022, Weighted Matrix Mapped CNN model for Optimizing the Sentiment Prediction, 978-1-6654-6246-4, 355, 10.1109/ICCCMLA56841.2022.9989039
    12. Siddhi Kadu, Bharti Joshi, 2022, Chapter 5, 978-3-031-10868-6, 81, 10.1007/978-3-031-10869-3_5
    13. Anju A, Barath M, Maheshwaran R, Vignesh K, Kishor Kumar S R, 2022, Sentimental Analysis for E-Commerce Website, 978-1-6654-6741-4, 1, 10.1109/ICETET-SIP-2254415.2022.9791606
    14. Mohammed Qorich, Rajae El Ouazzani, Text sentiment classification of Amazon reviews using word embeddings and convolutional neural networks, 2023, 0920-8542, 10.1007/s11227-023-05094-6
    15. Lin Wang, Zuqiang Meng, Multichannel Two-Dimensional Convolutional Neural Network Based on Interactive Features and Group Strategy for Chinese Sentiment Analysis, 2022, 22, 1424-8220, 714, 10.3390/s22030714
    16. Sudhakaran Periakaruppan, N. Shanmugapriya, Rajeswari Sivan, Self-attention generative adversarial capsule network optimized with atomic orbital search algorithm based sentiment analysis for online product recommendation, 2023, 10641246, 1, 10.3233/JIFS-222537
    17. B. Shanthini, N. Subalakshmi, 2023, Advanced Data Mining Enabled Robust Sentiment Analysis on E-Commerce Product Reviews and Recommendation Model, 978-1-6654-6216-7, 689, 10.1109/ICAIS56108.2023.10073782
    18. Huang Huang, Adeleh Asemi Zavareh, Mumtaz Begum Mustafa, Sentiment Analysis in E-Commerce Platforms: A Review of Current Techniques and Future Directions, 2023, 11, 2169-3536, 90367, 10.1109/ACCESS.2023.3307308
    19. Weibin Lin, Qian Zhang, Yenchun Jim Wu, Tsung-Chun Chen, Running a Sustainable Social Media Business: The Use of Deep Learning Methods in Online-Comment Short Texts, 2023, 15, 2071-1050, 9093, 10.3390/su15119093
    20. Ramy Riad Al-Fatlawy, Computational Intelligence-based Data Analytics for Sentiment Classification on Product Reviews, 2023, 2023, 2956-8323, 84, 10.2478/jsiot-2023-0012
    21. Neha Singh, Umesh Chandra Jaiswal, Jyoti Srivastava, 2023, Performance Evaluation of Machine Learning Classification Methods for Sentiment Analysis, 979-8-3503-3509-5, 1, 10.1109/ICCCNT56998.2023.10307844
    22. Tapasya Choudhary, Vanshika Aggarwal, Prasuk Jain, Shivani Trivedi, 2023, A MultiCriteria Approach in ECommerce Product Ranking, 9798400709418, 1, 10.1145/3647444.3647891
    23. Olamilekan Shobayo, Swethika Sasikumar, Sandhya Makkar, Obinna Okoyeigbo, Customer Sentiments in Product Reviews: A Comparative Study with GooglePaLM, 2024, 3, 2813-2203, 241, 10.3390/analytics3020014
    24. Xue Zhang, Fusen Guo, Tao Chen, Lei Pan, Gleb Beliakov, Jianzhang Wu, A Brief Survey of Machine Learning and Deep Learning Techniques for E-Commerce Research, 2023, 18, 0718-1876, 2188, 10.3390/jtaer18040110
    25. P. Vijayaragavan, Chalumuru Suresh, A Maheshwari, K Vijayalakshmi, R Narayanamoorthi, Miroslava Gono, Tomas Novak, Sustainable sentiment analysis on E-commerce platforms using a weighted parallel hybrid deep learning approach for smart cities applications, 2024, 14, 2045-2322, 10.1038/s41598-024-78318-1
    26. Joel de Goma, Rosselle Mae Mercado, Timothy Jacob Saporna, Sean Paolo Vallo, 2024, Sentiment Analysis of Tweets Towards Vaccines via Bidirectional Encoder Representations from Transformers (BERT), 9798400717055, 419, 10.1145/3669754.3669819
    27. Yanling Dong, Xiaolan Zhou, Advancements in AI-driven multilingual comprehension for social robot interactions: An extensive review, 2023, 31, 2688-1594, 6600, 10.3934/era.2023334
    28. Ramanna T, P Deepa Shenoy, Venugopal K R, 2023, Japanese Language Review Mining using Translators, Word Embedding and ML Techniques, 979-8-3503-0663-7, 171, 10.1109/ICRAIS59684.2023.10367133
    29. Min-Yuh Day, Chia-Tung Tsai, 2023, CMSI: Carbon Market Sentiment Index with AI Text Analytics, 9798400704093, 719, 10.1145/3625007.3627516
    30. Prachi Juyal, Amit Kundaliya, 2023, A Comparative Study of Hybrid Deep Sentimental Analysis Learning Techniques with CNN and SVM, 979-8-3503-1006-1, 596, 10.1109/AIC57670.2023.10263883
    31. Xiaoye Ma, Yanyan Li, Muhammad Asif, E-Commerce Review Sentiment Analysis and Purchase Intention Prediction Based on Deep Learning Technology, 2023, 36, 1546-2234, 1, 10.4018/JOEUC.335122
    32. Zhongnan Zhao, Wenjing Liu, Kun Wang, Research on sentiment analysis method of opinion mining based on multi-model fusion transfer learning, 2023, 10, 2196-1115, 10.1186/s40537-023-00837-x
    33. Ziyu Liang, Jun Chen, An Al-BERT-Bi-GRU-LDA algorithm for negative sentiment analysis on Bilibili comments, 2024, 10, 2376-5992, e2029, 10.7717/peerj-cs.2029
    34. Komal Rani Narejo, Hongying Zan, Kheem Parkash Dharmani, Lijuan Zhou, Tahani Jaser Alahmadi, Muhammad Assam, Nabila Sehito, Yazeed Yasin Ghadi, EEBERT: An Emoji-Enhanced BERT Fine-Tuning on Amazon Product Reviews for Text Sentiment Classification, 2024, 12, 2169-3536, 131954, 10.1109/ACCESS.2024.3456039
    35. Hendri Murfi, Theresia Gowandi, Gianinna Ardaneswari, Siti Nurrohmah, BERT-based combination of convolutional and recurrent neural network for indonesian sentiment analysis, 2024, 151, 15684946, 111112, 10.1016/j.asoc.2023.111112
    36. Neha Singh, Umesh Chandra Jaiswal, Sentiment classification on product reviews using machine learning and deep learning techniques, 2024, 0975-6809, 10.1007/s13198-024-02592-5
    37. K. Anuradha, M. Vamsi Krishna, Banitamani Mallik, Opinion Mining Using Normal Discriminant Piecewise Regressive (NDPR) Sentiment Classification Technique, 2023, 16, 1752-8909, 10.1142/S1752890922500131
    38. Chen Ding, Fei Qiao, Juan Liu, Dongyuan Wang, Knowledge graph modeling method for product manufacturing process based on human–cyber–physical fusion, 2023, 58, 14740346, 102183, 10.1016/j.aei.2023.102183
    39. Ashagrew Liyih, Shegaw Anagaw, Minichel Yibeyin, Yitayal Tehone, Sentiment analysis of the Hamas-Israel war on YouTube comments using deep learning, 2024, 14, 2045-2322, 10.1038/s41598-024-63367-3
    40. Alfredo Daza, Néstor Daniel González Rueda, Mirelly Sonia Aguilar Sánchez, Wilmer Filomeno Robles Espíritu, María Elena Chauca Quiñones, Sentiment Analysis on E-Commerce Product Reviews Using Machine Learning and Deep Learning Algorithms: A Bibliometric Analysis, Systematic Literature Review, Challenges and Future Works, 2024, 4, 26670968, 100267, 10.1016/j.jjimei.2024.100267
    41. J. Giftson Jeba Selva Raja, Sujitha Juliet, 2023, Deep Learning-based Sentiment Analysis of Trip Advisor Reviews, 978-1-6654-5630-2, 560, 10.1109/ICAAIC56838.2023.10140848
    42. Ganesh Kumar P, Arul Antran Vijay S, Jothi Prakash V, Anand Paul, Anand Nayyar, A context-sensitive multi-tier deep learning framework for multimodal sentiment analysis, 2023, 83, 1573-7721, 54249, 10.1007/s11042-023-17601-1
    43. N. Pughazendi, P.V. Rajaraman, Muzammil Hussain Mohammed, Graph Sample and Aggregate Attention Network optimized with Barnacles Mating Algorithm based Sentiment Analysis for Online Product Recommendation, 2023, 145, 15684946, 110532, 10.1016/j.asoc.2023.110532
    44. Yuvraj Giri, Vandana Bhagat, Ashaq Hussain Ganie, 2024, Empowering E-commerce: Leveraging Open AI and Sentiment Analysis for Smarter Recommendations, 979-8-3503-7809-2, 1, 10.1109/ICEECT61758.2024.10739003
    45. Muhammad Fairuzabadi, Kusrini Kusrini, Ema Utami, Arief Setyanto, 2024, Advancements and Challenges in Gated Recurrent Units (GRU) for Text Classification: A Systematic Literature Review, 979-8-3315-0553-0, 1, 10.1109/IC2IE63342.2024.10748229
    46. Wanying Xie, Wei Zhao, Zeshui Xu, Comparative analysis of multi-platform e-commerce online reviews based LDA and sentiment dictionary, 2024, 1756-378X, 10.1108/IJICC-08-2024-0384
    47. Esi Hairani, Ira Nurpalah, 2024, Sentiment Analysis of Muslim Pro “A1-Quran Adzan” on Google Play Using Multinomial Naïve Bayes, 979-8-3315-2879-9, 1, 10.1109/CITSM64103.2024.10775862
    48. Priya Darshini, Hardayal Singh Shekhawat, Design of a CWA-wbiLSTM Model for Aspect based Sentiment Classification for Product Reviews, 2024, 0929-6212, 10.1007/s11277-024-11690-3
    49. V. S. Raj Kumar, T. Kumaresan, P. Rajesh Kanna, S. Jagadeesan, P. Showmiya, P. Nithin, 2024, An Enhanced Product Quality Evaluation using Hybrid LSTM-FCNN Model and GWO Algorithm for Social Media Sentiment Analysis, 979-8-3503-6790-4, 972, 10.1109/ICECA63461.2024.10801023
    50. A. S. Sridhar, S. Nagasundaram, Sentiment classification for review rating prediction based on hybrid reaction diffusion support vector neural network, 2024, 14, 1869-5469, 10.1007/s13278-024-01397-5
    51. Tingting Mou, Hongbo Wang, Online comments of tourist attractions combining artificial intelligence text mining model and attention mechanism, 2025, 15, 2045-2322, 10.1038/s41598-025-85139-3
    52. Leiming Yuan, Haoyang Liu, Fangfang Fu, Yimin Liu, Xiaoyu Zuo, Limin Li, Study of Zhejiang Tangerine E-Commerce Reviews Based on Natural Language Processing, 2025, 11, 2311-7524, 151, 10.3390/horticulturae11020151
    53. Jinlan Chen, Jian Zhang, Jiajing Zhang, Shufeng Chen, 2024, Chinese Text Sentiment Analysis Model Based on BERT and BiTCN, 979-8-3315-3988-7, 127, 10.1109/ICACTE62428.2024.10871843
    54. Erfan Bakhtiari Ramezani, Sentiment analysis applications using deep learning advancements in social networks: A systematic review, 2025, 09252312, 129862, 10.1016/j.neucom.2025.129862
    55. Nilesh Anute, Rohith N R, Karthikeyan Parthasarathy, Mahesh Uday Mangaonkar, Pushpinder S. I. Dhingra, Phan Minh Duc, 2025, A Novel Clustering-Based Modeling and Evaluation of Product Differentiation for E-Commerce Businesses, 979-8-3315-1857-8, 750, 10.1109/CE2CT64011.2025.10940015
    56. Changzheng Yang, Mengmeng Guo, Muhammad Asif, Sentiment Analysis and Consumer Purchase Intention Prediction Based on BERT and DBLSTM, 2025, 37, 1546-2234, 1, 10.4018/JOEUC.372206
  • Reader Comments
  • © 2020 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(17560) PDF downloads(830) Cited by(56)

Figures and Tables

Figures(11)  /  Tables(6)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog