• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Bayesian estimation‐based sentiment word embedding model for sentiment analysis

    2022-05-28 15:16:56JingyaoTangYunXueZiwenWangShaoyangHuTaoGongYinongChenHaoliangZhaoLuweiXiao

    Jingyao Tang|Yun Xue|Ziwen Wang|Shaoyang Hu|Tao Gong|Yinong Chen|Haoliang Zhao|Luwei Xiao

    1Guangdong Provincial Key Laboratory of Quantum Engineering and Quantum Materials,School of Physics and Telecommunication Engineering,South China Normal University,Guangzhou,China

    2College of Mathematics and Informatics&College of Software Engineering,South China Agricultural University,Guangzhou,China

    3School of Foreign Languages,Zhejiang University of Finance&Economics,Hangzhou,Zhejiang,China

    4Educational Testing Service,Princeton,New Jersey,USA

    5School of Computing,Informatics and Decision Systems Engineering,Arizona State University,Tempe,USA

    Abstract Sentiment word embedding has been extensively studied and used in sentiment analysis tasks.However,most existing models have failed to differentiate high-frequency and lowfrequency words.Accordingly,the sentiment information of low-frequency words is insufficiently captured,thus resulting in inaccurate sentiment word embedding and degradation of overall performance of sentiment analysis.A Bayesian estimation-based sentiment word embedding (BESWE) model,which aims to precisely extract the sentiment information of low-frequency words,has been proposed.In the model,a Bayesian estimator is constructed based on the co-occurrence probabilities and sentiment probabilities of words,and a novel loss function is defined for sentiment word embedding learning.The experimental results based on the sentiment lexicons and Movie Review dataset show that BESWE outperforms many state-of-the-art methods,for example,C&W,CBOW,GloVe,SE-HyRank and DLJT1,in sentiment analysis tasks,which demonstrate that Bayesian estimation can effectively capture the sentiment information of low-frequency words and integrate the sentiment information into the word embedding through the loss function.In addition,replacing the embedding of low-frequency words in the state-of-the-art methods with BESWE can significantly improve the performance of those methods in sentiment analysis tasks.

    1|INTRODUCTION

    The past decade has witnessed the flourish of word embedding models together with their successful application in natural language processing research.Word embedding models have been widely used in the tasks of entity recognition [1],word sense disambiguation [2,3],dependency parsing [4],machine translation [5],sentiment analysis [6,7] and so on.Generally,the current word embedding models can be categorized into two types [8],namely the prediction-based models [9-13] and the count-based models [14-16].The former type adopts language models to predict the next word based on its context,whereas the latter uses the global word co-occurrence counts to derive word embeddings.As such,the word embedding models are able to capture the context information and map the semantically similar words into neighbouring points within the word embedding space[17].Nevertheless,considering that semantically similar words may have opposite sentiment polarities [18],the widely used word embedding models have challenges in dealing with sentiment information.According to Yu et al.[19],among the top 10 semantically similar words to each target word,about 30% of them appears to have an opposite sentiment polarity to that of the target word.For this reason,research is still ongoing to develop sentiment word embedding algorithms that can appropriately tackle both semantic and sentiment information.

    Recently,the sentiment word embedding models have been put forward to tackle sentiment information.Similar to the traditional word embedding models,the recently developed sentiment word embedding models can also be classified into the prediction-based models and count-based models.On the one hand,in the prediction-based models [18,35-37],the word sentiment is taken as the local information for model learning.That is,by using the neural networks or predicting functions,the sentiment word embedding can be learnt and the sentiment polarities can be classified.On the other hand,for the count-based models [20,21],the sentiment word embedding is derived from the word-sentiment matrices of the labelled corpora,in which case,the sentiment information is used as the global statistics.

    Despite of these new models,the application of sentiment word embedding models is still limited.Current state-of-the-arts fail to dedicatedly identify and process the low-frequency words in the texts,which result in the loss of low frequency but significant information [22].Based on the Zipf's law[23],a large number of words in any corpus are often low-frequency words[24].According to statistics,in the Stanford Sentiment Tree(SST) corpus [25],the words with the occurrence rates lower than 5 account for 72.35%of all the words,and those lower than 10 account for 84.06%.In addition,because of the small number of low-frequency word samples,the learning process is prone to fall in over-fitting of the model.Notably,the low-frequency words carry not only little sentiment information but also the noise that might affect model learning.Instead of analyzing the low-frequency words,many current sentiment word embedding models directly utilize the word information for sentiment word embedding learning.For example,in the labeled singular value decomposition(LSVD)model,the word-context matrix and the word-sentiment matrix are stitched together to obtain the sentiment word embedding through the singular value decomposition.Likewise,the LGloVe model and the DLJT1 model use both matrices as the loss function of the least squares algorithm for model learning.In this way,the sentiment word embedding models are absent of extracting the precise sentiment of lowfrequency words among the word representations.Thus,this suspending issue will result in the low accuracy of sentiment polarity classification.

    On the task of sentiment analysis,we tend to propose a novel method for sentiment word embedding learning.We design and deploy a Bayesian estimation-based sentiment word embedding(BESWE)model.Inspired by the GloVe,we introduce the sentiment probability as well as derive the loss function.Accordingly,the learnt word embedding is deemed best able to convey both the context information and the sentiment information.Seeing that the co-occurrence probability in GloVe is computed via maximum likelihood estimation,which has deficiencies in processing small samples,we thus use the Bayesian estimating method to pick up the sentiment information from the word-sentiment matrices for model learning.By introducing the prior knowledge,the Bayesian estimation can overcome the defect of insufficient sentiment information [26] and increase the robustness of the model[27].Our model is especially more accurate in capturing the sentiment information of lowfrequency words,due to its distinctive structure.On the one hand,we use a Bayesian estimator to compute the co-occurrence probabilities and the sentiment probabilities based on wordcontext matrices and word-sentiment matrices from the corpus.On the other hand,we exploit the learning of sentiment word embedding in a novel way by dedicatedly constructing a loss function.We compare our method against other state-ofthe-art models using identical experimental setups.In these configurations,the proposed model obtains a decent working performance in various natural language processing(NLP)tasks results and is considerably better than classical approaches in low-frequency word sentiment identification.

    The major contributions of this work are threefold and can be summarized as follows:

    (1) Based on the Bayesian estimation principle,our sentiment probability computation method aims to extract the sentiment information of low-frequency words from the word-sentiment matrix.

    (2) Our BESWE model is innovatively developed for the sentiment analysis of low-frequency words,which integrates sentiment information with word embeddings and achieves a higher accuracy in tasks involving low-frequency words and low-frequency sentences sentiment analysis.

    (3) The conceptual framework of BESWE can be applied to other word embedding models.By capturing the sentiment of low-frequency words,the sentiment polarity classification accuracy of the current models can be largely improved by integrating BESWE.

    The rest of this study is organized as follows.Section 2 presents the prerequisites needed for understanding the proposed model.Section 3 describes the proposed sentiment word embedding model.Section 4 presents the experimental results.A related work to our model is given in Section 5.Conclusions are drawn in Section 6.

    2|PREREQUISITE

    This section introduces the basic knowledge of GloVe model and its parameter estimation principle,with the purpose of facilitating the description of subsequent model architecture.

    2.1|GloVe

    The GloVe model is a word-embedding method that combines evidence from the local context and the global counts [28].Typically,the method involves three distinguished wordsi,jandk.Bothiandjare target words,whilekstands for the context words.Letxbe the matrix representing the word-word co-occurrence counts.We can define the elementxikas the times for wordkappearing in the contexti.Correspondingly,xi=∑kxikindicates the total occurrence counts of any word within the context of wordi.Therefore,the co-occurrence probability ofkin the context word ofiis:

    Then,Pik/Pjkdenotes the relation ofitokandjtok.As long askhas the similar relations toiandj,that is,both relevant or irrelevant,the ratioPik/Pjkwould be close to 1.The information within the ratio of co-occurrence probabilities can be formulated into:

    wherew∈Rnrefers to the target word embedding and~w∈Rnto the context embedding.

    2.2|Parameter estimation principle

    On this occasion,we tend to prove that the co-occurrence probabilityPikcan be derived from the maximum likelihood estimation.

    For every single target wordi,xitimes Bernoulli experiments are conducted to extract the context independently and randomly[29].In each experiment,there existVdifferent outcomes.The occurrence number of theVthoutcome,together with its probability,is represented byxikandPik,respectively.

    If the random variablestands for the occurrence times of all the possibilities,in whichXikis the number of occurrence for thekthone,the parameterXimust obey the multinomial distribution,which can be written as:

    To maximize the log-likelihood function in Equation (4),setting up the target function can be viewed as an equality constrained optimization problem:

    Accordingly,the corresponding Lagrangian function can be formulated as:

    together with determining the partial derivatives ofPik

    Seeing that the parameterλis with respect to

    It is clear however that the estimation ofPikis written as:

    Obviously,the co-occurrence probability in GloVe is equivalent to the maximum likelihood estimating outcome of the co-occurrence probability.Theoretically,the maximum likelihood estimation is a statistical estimating method on abundant samples.In contrast,the Bayesian estimation is more effective in small sample processing [26].

    3|METHODOLOGY

    This section introduces the architecture of our BESWE model and its working principle.

    Notation:

    · ForandVrepresenting the number of words,we definexas the word-context matrix,together withxikandPik,respectively as the occurrence counts and the occurrence probability of wordkin context of wordi,which are from the setsand

    ·stands for the total occurrence counts of any word within the context of wordi.

    · For,we definetas the word-sentiment matrix,together withti1as the number of positive texts including wordiandti0as the number of negative texts including wordi,whereti=ti0+ti1.

    · Forwe defineBias the probability of wordibeing positive and 1-Bias the probability of wordibeing negative.

    ·wistands for word embedding of wordias target.

    ·~wistands for word embedding of wordias context.

    ·sistands for bias embedding of wordi.

    3.1|Model architecture

    The architecture of BESWE model is shown in Figure 1.Our sentiment word embedding model is developed based on the foundation of GloVe.Since the feasibility of maximum likelihood estimation for parameter calculation in GloVe is demonstrated,we start with using the co-occurrence probability for word context learning and the sentiment probability for word sentiment learning.By traversing the corpus with sentiment labels,the co-occurrence countsxi→and sentiment countsti→can be obtained.Aiming at dealing with lowfrequency words,the co-occurrence number is sent to Bayesian estimator for co-occurrence probability computation.Likewise,the sentiment probability can be obtained in the same manner by using sentiment counts of the word.At this stage,the estimation of lnPikand lnBican be obtained,which arecikandei,respectively.Based on these two outcomes,the loss function is constructed for sentiment word embedding learning,which is delivered as:

    The loss function is minimized via AdaGrad optimizer[30].Along these processes,the learning ofwi,~wiandsiis conducted wherewistands for the sentiment word embedding of the BESWE model.

    More details of the Bayesian estimator and the loss function of sentiment word embedding learning are described as follows.

    3.2|Bayesian estimator

    Based on the working principle of the proposed model,the cooccurrence probability and the sentiment probability are calculated by the Bayesian estimator.Distinctively,the estimation of co-occurrence probability comes from the current D-GloVe model [16].According to the parameter distribution assumption in Section 2.2,is assumed to obey the Dirichlet prior distribution with parameteras shown in Figure 2.Thus,the Bayesian estimation of lnPikis:

    wherenkis the occurrence time of wordkin the learning samples,andλ1is the regulatory factor.

    FIGURE 1 Model architecture

    FIGURE 2 Probability distribution of parameter estimation of DGloVe model

    wheremkstands for the number of texts of the sentiment labelk,andλ2is the regulatory factor for word sentiment learning.

    FIGURE 3 Probability distribution of parameter estimation of BESWE model

    3.3|Loss function

    Considering the semantic information characterization in GloVe,we now take the ratio of sentiment probabilities to address the relationship of word sentiment.More details about the words'sentiment relation and the ratio are presented in the Appendix 7.1.For wordsiandj,the sentiment relation is expressed asBi/Bj.The objective function is established as:

    wherewjandsjstand for the word embedding and bias embedding of wordj,respectively.

    Assuming thatFis confirmed as the homomorphism between groups(R,+)and(R>0,×),we incorporate the sentiment information into the word embeddings.Based on Equations (2) and (22),we get:

    Considering the properties of group homomorphism,we transform the above equation according to the addition commutative law into:

    In line with the basic theory of GloVe,the loss function of BESWE can be constructed via the same manner:

    wherecikandeiare derived from Equations (14) and (18),respectively.

    4|EXPERIMENT

    In this section,the working performance of the proposed BESWE model is evaluated.Some state-of-the-art word embedding models,along with the learning of specific word representations,are taken for comparison.To this end,the task of word similarity is carried out.So as to deliver the sentiment embeddings,the word-and the sentence-level sentiment analysis using different models is taken into account.The sentiment analysis tasks are further divided into three subtasks,which verify the efficiency on basic sentiment classification for all words,low-frequency word sentiment identification and integration with other baseline models.The outline of the experiments is exhibited in Figure 4.

    4.1|Experiment settings

    Dataset of word embeddings:The dataset SST is used for the model training.There are five classes annotations within SST,which are originally ‘very negative’,‘negative’,‘neutral’,‘positive’and‘very positive’.To facilitate the processing,the classes‘positive’and‘very positive’are taken to represent the polarity of positive while the‘negative’and‘very negative’represent the negative.As such,both categories are applied to the experiments.

    Baseline models:We compare the effectiveness of the proposed model to other widely used models.Specifically,the models of word embeddings,such as C&W [10],CBOW [13]and GloVe [15],together with models of sentiment embeddings,including SE-HyRank [18] and DLJT1 [21],are implemented.For all the models used in this work,the word representation dimension is 50,the learning rate is set as 0.05.All the parameters associated are finetuned to get better results.

    Task 1,Word similarity measure:The capacity of word embedding models is verified via the processing in standard word similarity.Comprehensively,the average working performance on word similarity tasks is obtained on the dataset EN-WS-353-ALL,EN-WS-353-SIM and SCWS.We utilize the word embeddings to calculate the similarity score of the word pairs first.Together with the standard similarity score provided by the dataset,we shall thus compute the correlation coefficient.The evaluation metric in this task is the computed correlation coefficient,which is detailed elsewhere [31].

    Task 2,Word‐level sentiment analysis:On this occasion,the support vector machine classifier is used for training the words vectors,with each word vector representing specific sentiment.The popular sentiment polarity lexicons,NRC and MPQA [32],are taken as the ground truth of the word sentiments.The number of positive and negative words for MPQA is 2301 and 4151 while those for NRC is 2231 and 3324.TheN-fold cross validation with the valueN=5 andN=10 is performed on all datasets.The working performance refers to the classification accuracy of the models.

    The word-level sentiment analysis is carried out via the following three subtasks.

    Task 2.1,Basic word‐level sentiment analysis:The usual task for evaluating the performance is to capture the sentiment information of the words.Thus,we generate the word embeddings and apply them to the sentiment classification of the sentiment lexicon.

    Task 2.2,Low‐frequency word sentiment analysis:Since the Bayesian estimation principle is able to deal with lowfrequency words,we pick the words with the frequencies lower than 5 and 10 from the SST corpus for investigation.In this way,the low-frequency word embeddings are collected for sentiment analysis.

    Task 2.3,BESWE integration with other models:In order to improve the classification accuracy,we integrate BESWE with the baseline models.A specific word embedding set is developed,which contains low-frequency word embeddings from BESWE and nonlow-frequency word embeddings from any other baseline model.By using these word embeddings for model learning,the sentiment analysis is carried out.

    Task 3,Sentence‐level sentiment analysis:Considering the sentiment analysis for sentence,the Movie Review dataset[33]is used,which contains 10,622 samples with the proportion of each polarity 1:1.We use a convolutional neural network(CNN),namely text-CNN,with its online implementation[34].The inputs of text-CNN are word embeddings.The samples are divided into training,validating and testing,with the ratio of 6:2:2.The training episode is set as 200 epochs using the default settings.Based on the validation outcomes,we apply the optimal model to testing.The evaluation metrics are the classification accuracy and F1 score.

    Similarly,there are also three subtasks like in the word-level sentiment analysis.

    Task 3.1,Basic sentence‐level sentiment analysis:As stated in Task 2.1,we take the words from Movie Review for sentiment analysis.

    Task 3.2,Low‐frequency sentence sentiment analysis:In this task,we select the low-frequency words whose occupation is over 1 0%in the sentence to be the low-frequency sentence samples.Thereby the low-frequency sentences are taken for sentiment analysis.

    Task 3.3,BESWE integration with other models:Similar to Task 2.3,the integration of BESWE with each baseline model is devised to deal with both the low-frequency words and the nonlow-frequency words in the sentences from the Movie Review dataset.

    4.2|Experimental results

    FIGURE 4 Workflow of experiments

    Task 1,Word similarity measure:The outcomes of the word similarity task are given in Table 1.It is apparent that our BESWE method outperforms the other methods on all three datasets,verifying the capability of capturing sufficient semantic information for analysis.The sentiment word embeddings are generally more competitive than the basic word embeddings on the task of word similarity.The main reason is that,by exploiting the sentiment information,semantically more accurate word embeddings are obtained to optimize the working performance.Moreover,the Bayesian estimationprinciple is effective in tackling the large number of lowfrequency words in the corpus,based on which more accurate sentiment word embeddings are accessible.Accordingly,one can easily see a considerable gap between our model and the baseline models.

    TABLE 1 Word similarity results

    Task 2.1,Basic word‐level sentiment analysis:The wordlevel analysis is carried out on the dataset of single-word entries.For basic word-level sentiment analysis,our model is proved to be a competitive alternative to classical word embedding models(Table 2).Furthermore,the BESWE model obtains the best outcome with 10-fold cross validation on the dataset NRC.

    Task 2.2,Low‐frequency word sentiment analysis:The BESWE model shows a better accuracy in tackling lowfrequency words,as shown in Table 3.For the frequencies lower than 5 (LF-5) and 10 (LF-10),our model shows its stability in sentiment polarity classification,compared to the baseline models.As mentioned in the result of Task 1,the lowfrequency word processing ability contributes to establishing precise sentiment word embeddings.By contrast,there is also a performance gap between BESWE and other algorithms for the dataset NRC.

    Task 2.3,BESWE integration with other models:In line with the results of Task 2.2,we combine the proposed model to baseline models,not only to show its superiority,but also to optimize the word-level sentiment analysis outcomes of baseline models.As long as the BESWE algorithm is capable of processing low-frequency words,the low-frequency word embeddings (i.e.LF-5 and LF-10 from BESWE) are incorporated into the nonlow-frequency word embeddings from the baseline.The outcomes of the integration method are reported in Table 4.The classification accuracy is significantly improved for most models.The maximum performance gap of 4.44%is observed in the integration of DLJT1 and BESWE against the basic DLJT1 for the dataset MPQA.

    Task 3.1,Basic sentence‐level sentiment analysis:The effectiveness of the proposed model is further evaluated on the sentence-level sentiment analysis tasks.From Table 5,we see that the BESWE model has a better accuracy than any other baseline methods.

    Task 3.2,Low‐frequency sentence sentiment analysis:In line with the outcome on low-frequency words,the sentiment analysis on low-frequency sentence exceeds the baselines in both evaluation settings.The highest classification accuracy ofthe BESWE,based on the word frequencies lower than 5(LF5),is 81.25%,as shown in Table 6.Meanwhile,the maximum performance gap of 9.75% is observed against the DLJT1.Comparing Tables 5 and 6,the improvement in accuracy against the baseline in Table 6 outperforms that in Table 5.One possible explanation is that the low-frequency sentences in Task 3.2 contain more low-frequency words.Since the sentiment word embeddings learnt via the Bayesian principle is considerably more informative,it is reasonable to expect better performance in low-frequency words analysis,as it is the case.Besides,the accuracy on ‘LF5’ in Table 6 overperforms that of‘LF10’,which further verifies the significance of our model.

    TABLE 2 Basic word-level sentiment analysis results

    Task 3.3,BESWE integration with other models:Similar to Task 2.3,since the working performance on low-frequency sentence is highlighted,we integrate the corresponding word embeddings to improve the classification accuracy of the baseline.The outcomes of the integration methods,which combine the baseline and the proposed BESWE,are reported in Table 7.In contrast,these outcomes fail to exceed that from the direct use of the proposed model.That is,for the sentiment analysis of low-frequency words,BESWE always obtains the best and most consistent results in the identification of sentiment polarity.Whereas,the application of BESWE can narrow the performance gap.For the C&W model,a 10%improvement is presented.

    Effects of λ1and λ2:The hyperparameters in the BESWE model include the regulatory factorsλ1andλ2used to represent the semantic and sentiment information.In this experiment,to obtain the optimal settings,we vary the values ofλ1andλ2withinto learn the BESWE model.In this way,we get 64 different BESWE models.

    The results on the low-frequency sentence sentiment analysis against different hyperparameter settings are shown in Figures 5 and 6.The former indicates the accuracy on the lower-than-5-frequency words while the latter indicates that of the lower-than-10-frequency words.The variation ofλ1does not cause a significant difference,while that ofλ2has a negative correlation with the accuracy.According to Figure 5,the highest accuracy of the sentence-level sentiment analysis is 81.25% at the pointλ1andλ2=0.01.Likewise,in Figure 6,the optimal values ofλ1andλ2are both 0.01,which lead to an accuracy of 80.50%.

    TABLE 3 Low-frequency word sentiment analysis results

    TABLE 4 Word sentiment analysis results of integrating models

    TABLE 5 Basic sentence-level sentiment analysis results

    To sum up,these experimental results clarify the effectiveness of the proposed sentiment word embedding.The BESWE model outperforms other state-of-the-art models in word similarity measure.In the sentiment analysis at both word and sentence levels,our method still shows comparableoutcomes.Specifically,our model produces considerably better results than the baseline methods on sentiment information capturing of both low-frequency words and low-frequency sentences.Moreover,by integrating the lowfrequency word embeddings from BESWE to other models,the classification accuracies of the baseline models improve to a large extent.

    TABLE 6 Low-frequency sentence sentiment analysis results

    TABLE 7 Sentence sentiment analysis results of integrating models

    FIGURE 5 The sensitivity of λ1 and λ2 on the BESWE(LF5)in lowfrequency sentence sentiment analysis

    FIGURE 6 The sensitivity of λ1 and λ2 on the BESWE(LF10)in lowfrequency sentence sentiment analysis

    5|RELATED WORK

    5.1|Word embeddings

    As pointed out in the Introduction section,both predictionbased word embedding models and count-based word embedding models are applied to learn word embeddings [8].Fundamentally,Bengio et al.[9],establish neural network language model to predict target words using preceding contexts,and thus to learn the word embeddings.Following this theory,Collobert and Weston [10,11] put forward a CNN to predict the target word,based on not only preceding but also succeeding contexts.As currently edge-cutting predictionbased methods,CBOW and skip-gram models [12,13] are of simple single-layer architecture.Both of these models can efficiently compute word embeddings from large-scale datasets.Besides,researchers also focus on using the global word context co-occurrence counts in the corpus for learning word embeddings,which are taken as the basis of count-based methods.Deerwester et al.[14] novelty propose the latent semantic analysis model for exploiting word-document cooccurrence counting to learn word embedding for the first time.As a most widespread model,GloVe [15] computes the word-word co-occurrence counts via a specific weighted least squares model.A consistent and competitive result can be obtained on the tasks of sentiment analysis.

    5.2|Sentiment word embeddings

    Likewise,prediction-based sentiment word embedding models and count-based sentiment word embedding models are dedicatedly studied.In 2011,Maas et al.[35] apply a logistic regression as the predictor to learn sentiment word embeddings.On the other hand,an approach based on recursive autoencoders [36] is designed that year to learn the vector representations of phrases and full sentences,which exploits the vector representations at each node of the hierarchy and uses softmax classifier for sentiment label prediction.Tang et al.[18] propose the hybrid ranking method for learning sentiment embeddings by regulating the traditional C&W model,which encodes sentiment information in the continuous representation of words.With the development of deep learning networks,Lan et al.[37] construct a CNN to detect the semantic and sentiment information.Hereafter,two kinds of information are integrated to generate sentiment word vectors.In terms of count-based sentiment word embedding models,Li et al.[21] incorporate the sentiment count into model learning by proposing a variety of count-based models(e.g.DLJT1)on the foundation of GloVe.Furthermore,LSVD and LGloVe [20] are developed as the improvement of SVD and GloVe,respectively,In these models,the application of word-label counts facilitates the learning of sentiment word embedding with label information.

    Different from these works,our work mainly focuses on resolving the issues of low-frequency word sentiment analysis.Little attention is paid to this topic currently within the NLP domain.Our model targets at extracting the sentiment information by obtaining a more accurate sentiment word embedding.Notably,our model is orthogonal to the aforementioned models.Our model overperforms the state-of-the-arts while the adaptation of our approach into these methods also results in a better working performance.

    6|CONCLUSIONS

    In this work,a novel BESWE model is designed and deployed on the tasks of sentiment analysis.Aiming to obtain both the semantic and the sentiment information,a Bayesian estimator is developed to compute the co-occurrence probability and the sentiment probability.Furthermore,a loss function for the purpose of sentiment word embedding learning is constructed.We test our model on a variety of tasks to evaluate its working performance.Experimental results indicate that the BESWE model is a comparable alternative to the state-of-the-art methods in word similarity identification,word-and sentence-level sentiment analysis.Specifically,our model outperforms other methods on low-frequency word and lowfrequency sentence sentiment polarity classification to demonstrate its efficacy.By integrating the BESWE into the baseline,the classification accuracy can be improved considerably compared to the basic models.

    This study offers a creative and practical method for both the semantic and sentiment information capturing.Distinctively,our model shows its superiority in dealing with low-frequency words and thus results in a higher accuracy in sentiment analysis.

    ACKNOWLEDGEMENTS

    This work was supported by the National Statistical Science Research Project of China under Grant No.2016LY98,the Science and Technology Department of Guangdong Province in China under Grant Nos.2016A010101020,2016A01 0101021 and 2016A010101022,the Characteristic Innovation Projects of Guangdong Colleges and Universities (Nos.2018KTSCX049 and 2018GKTSCX069),the Science and Technology Plan Project of Guangzhou under Grant Nos.201802010033 and 201903010013,the Bidding Project of Laboratory of Language Engineering and Computing of Guangdong University of Foreign Studies (No.LEC2019 ZBKT005).

    ORCID

    Jingyao Tanghttps://orcid.org/0000-0003-1651-8480

    How to cite this article:Tang,J.,et al.:Bayesian estimation-based sentiment word embedding model for sentiment analysis.CAAI Trans.Intell.Technol.7(2),144-155 (2022).https://doi.org/10.1049/cit2.12037

    APPENDIX

    SENTIMENT RELATION OF DIFFERENT WORDS USING BAYESIAN ESTIMATION

    Specific sentiment relations calculated by Bayesian estimation are presented in Table 8.

    TABLE 8 Sentiment relation of different words using Bayesian estimation

    露出奶头的视频| 波多野结衣高清作品| 美女高潮喷水抽搐中文字幕| 中亚洲国语对白在线视频| 搡老妇女老女人老熟妇| 久9热在线精品视频| 久久6这里有精品| 国产淫片久久久久久久久| 欧洲精品卡2卡3卡4卡5卡区| 日日摸夜夜添夜夜添av毛片 | 日本a在线网址| 国产精品美女特级片免费视频播放器| 成人欧美大片| 夜夜看夜夜爽夜夜摸| 五月伊人婷婷丁香| 日日干狠狠操夜夜爽| 在线看三级毛片| 成人一区二区视频在线观看| 国内精品美女久久久久久| 国产精品一及| 久久国内精品自在自线图片| 国产色爽女视频免费观看| 伊人久久精品亚洲午夜| 久久精品国产自在天天线| 最近视频中文字幕2019在线8| 真人一进一出gif抽搐免费| 级片在线观看| 久久香蕉精品热| 国产伦精品一区二区三区视频9| 国产黄色小视频在线观看| 国内精品久久久久久久电影| 在线观看舔阴道视频| 亚洲专区中文字幕在线| 国产精品爽爽va在线观看网站| 麻豆国产av国片精品| 乱码一卡2卡4卡精品| 国产白丝娇喘喷水9色精品| 成人毛片a级毛片在线播放| 五月伊人婷婷丁香| 毛片一级片免费看久久久久 | 亚洲精品在线观看二区| 国产精品自产拍在线观看55亚洲| 51国产日韩欧美| 欧美最黄视频在线播放免费| 日韩人妻高清精品专区| 听说在线观看完整版免费高清| 成人性生交大片免费视频hd| 日韩大尺度精品在线看网址| 亚洲 国产 在线| 久久久久久久午夜电影| 亚洲人成伊人成综合网2020| 中文字幕高清在线视频| 色在线成人网| 亚洲七黄色美女视频| 国产精品伦人一区二区| 亚洲色图av天堂| 男人舔女人下体高潮全视频| 桃红色精品国产亚洲av| 日日夜夜操网爽| 免费在线观看影片大全网站| 国产国拍精品亚洲av在线观看| 精品一区二区三区视频在线| 亚洲av中文字字幕乱码综合| 亚洲 国产 在线| 国产精品爽爽va在线观看网站| 很黄的视频免费| 18+在线观看网站| 国产精品久久久久久亚洲av鲁大| 在线观看美女被高潮喷水网站| 干丝袜人妻中文字幕| 亚洲无线观看免费| 久久精品久久久久久噜噜老黄 | 三级国产精品欧美在线观看| 88av欧美| bbb黄色大片| 国产在线男女| 婷婷精品国产亚洲av在线| or卡值多少钱| 欧美日韩亚洲国产一区二区在线观看| 欧美区成人在线视频| 亚洲av美国av| 少妇人妻精品综合一区二区 | 成年免费大片在线观看| 精品福利观看| 亚洲精品色激情综合| 中亚洲国语对白在线视频| 亚洲欧美日韩高清专用| 亚洲欧美激情综合另类| 成年免费大片在线观看| 99久久精品国产国产毛片| 桃红色精品国产亚洲av| 少妇人妻精品综合一区二区 | 亚洲真实伦在线观看| www.色视频.com| 床上黄色一级片| 亚洲七黄色美女视频| 麻豆成人午夜福利视频| 亚洲精品456在线播放app | 免费高清视频大片| 麻豆国产av国片精品| 亚洲av中文字字幕乱码综合| av在线观看视频网站免费| 国内毛片毛片毛片毛片毛片| 欧美国产日韩亚洲一区| 大型黄色视频在线免费观看| 69人妻影院| 亚洲avbb在线观看| 一区二区三区激情视频| 国产黄a三级三级三级人| 狂野欧美白嫩少妇大欣赏| 国产伦精品一区二区三区四那| 亚洲专区国产一区二区| 国模一区二区三区四区视频| 亚洲精品国产成人久久av| 能在线免费观看的黄片| 啦啦啦观看免费观看视频高清| 久久久久免费精品人妻一区二区| 国产黄片美女视频| 在线观看av片永久免费下载| 国产真实伦视频高清在线观看 | 成人性生交大片免费视频hd| 国产精品久久电影中文字幕| 天堂av国产一区二区熟女人妻| 欧美高清性xxxxhd video| 色av中文字幕| 看片在线看免费视频| 伦精品一区二区三区| 成年女人看的毛片在线观看| 一级a爱片免费观看的视频| 久久久久久久久久黄片| 日本a在线网址| 精品一区二区三区视频在线| 亚洲av不卡在线观看| 午夜视频国产福利| 网址你懂的国产日韩在线| 亚洲精品一区av在线观看| 欧美黑人欧美精品刺激| 中文字幕精品亚洲无线码一区| 国产午夜精品论理片| 少妇丰满av| 禁无遮挡网站| 美女被艹到高潮喷水动态| 久久人人精品亚洲av| 中亚洲国语对白在线视频| 啦啦啦观看免费观看视频高清| 午夜免费成人在线视频| 日韩欧美精品免费久久| 欧美高清成人免费视频www| 久久国产乱子免费精品| 国产欧美日韩一区二区精品| 国产毛片a区久久久久| 久久精品国产清高在天天线| 久久久久性生活片| 老熟妇乱子伦视频在线观看| 麻豆国产97在线/欧美| 亚洲成a人片在线一区二区| 亚洲色图av天堂| 免费不卡的大黄色大毛片视频在线观看 | 一个人看视频在线观看www免费| 欧美日韩精品成人综合77777| 最后的刺客免费高清国语| 日日夜夜操网爽| 久久国产乱子免费精品| 欧美色视频一区免费| 日韩欧美在线乱码| 久久婷婷人人爽人人干人人爱| 久久久色成人| 日日摸夜夜添夜夜添av毛片 | 可以在线观看毛片的网站| 最好的美女福利视频网| 久久国内精品自在自线图片| 午夜日韩欧美国产| 国产精品一及| 国产中年淑女户外野战色| 婷婷亚洲欧美| 国产精品一区二区三区四区久久| 极品教师在线视频| 国产高清视频在线播放一区| 久久6这里有精品| 天堂网av新在线| 亚洲成人精品中文字幕电影| 高清毛片免费观看视频网站| 欧美激情国产日韩精品一区| 久久人人爽人人爽人人片va| 99久国产av精品| 国产精品久久久久久久电影| 成人午夜高清在线视频| 亚洲av中文字字幕乱码综合| 91麻豆精品激情在线观看国产| 可以在线观看的亚洲视频| 国内精品一区二区在线观看| 五月玫瑰六月丁香| 成熟少妇高潮喷水视频| 此物有八面人人有两片| 亚洲精品影视一区二区三区av| 国产探花在线观看一区二区| 色综合亚洲欧美另类图片| 无人区码免费观看不卡| 国产熟女欧美一区二区| 成人午夜高清在线视频| 在线播放无遮挡| 久久久成人免费电影| 无人区码免费观看不卡| 精品不卡国产一区二区三区| 国产伦人伦偷精品视频| 国产av麻豆久久久久久久| 91av网一区二区| 亚洲欧美日韩高清专用| 麻豆成人av在线观看| 亚洲av日韩精品久久久久久密| 亚洲美女搞黄在线观看 | 有码 亚洲区| 欧美成人免费av一区二区三区| 亚洲成人久久性| av在线天堂中文字幕| 很黄的视频免费| 一区二区三区四区激情视频 | 又爽又黄无遮挡网站| 成人美女网站在线观看视频| 变态另类成人亚洲欧美熟女| 美女免费视频网站| 九九热线精品视视频播放| 午夜爱爱视频在线播放| 国产成人a区在线观看| 九九在线视频观看精品| 国产精品免费一区二区三区在线| 91久久精品国产一区二区成人| aaaaa片日本免费| 色综合婷婷激情| 2021天堂中文幕一二区在线观| 免费看美女性在线毛片视频| 女人被狂操c到高潮| 看十八女毛片水多多多| 国产一区二区亚洲精品在线观看| 国产aⅴ精品一区二区三区波| 亚洲成人精品中文字幕电影| 美女 人体艺术 gogo| 国产精品1区2区在线观看.| 国产精品自产拍在线观看55亚洲| 国产av麻豆久久久久久久| 久久婷婷人人爽人人干人人爱| 一级黄片播放器| 欧美日韩综合久久久久久 | 美女被艹到高潮喷水动态| 少妇丰满av| 可以在线观看的亚洲视频| 变态另类成人亚洲欧美熟女| 欧美zozozo另类| 午夜免费男女啪啪视频观看 | 好男人在线观看高清免费视频| 欧美激情久久久久久爽电影| 观看美女的网站| 可以在线观看毛片的网站| 久久久成人免费电影| 麻豆精品久久久久久蜜桃| 亚洲精品一卡2卡三卡4卡5卡| 中文字幕熟女人妻在线| 老女人水多毛片| 少妇人妻一区二区三区视频| 在线免费观看不下载黄p国产 | 美女 人体艺术 gogo| 全区人妻精品视频| 日本 欧美在线| 成人永久免费在线观看视频| 亚洲无线在线观看| 免费一级毛片在线播放高清视频| 88av欧美| 国产久久久一区二区三区| 亚洲美女黄片视频| 51国产日韩欧美| 亚洲电影在线观看av| 免费人成视频x8x8入口观看| av在线亚洲专区| 国产精品一区二区三区四区久久| 真实男女啪啪啪动态图| 日韩欧美一区二区三区在线观看| 欧美高清性xxxxhd video| 村上凉子中文字幕在线| 美女 人体艺术 gogo| 不卡一级毛片| 毛片一级片免费看久久久久 | 综合色av麻豆| 欧美国产日韩亚洲一区| 3wmmmm亚洲av在线观看| 国产单亲对白刺激| 色在线成人网| 狂野欧美白嫩少妇大欣赏| 免费观看在线日韩| 久久久久九九精品影院| 在线观看66精品国产| 欧美性猛交╳xxx乱大交人| 99精品在免费线老司机午夜| 国产高清激情床上av| 色视频www国产| 国内少妇人妻偷人精品xxx网站| 一本久久中文字幕| 国产亚洲精品久久久久久毛片| 久久久久久久精品吃奶| 黄色一级大片看看| 99国产精品一区二区蜜桃av| 亚洲va日本ⅴa欧美va伊人久久| 男女边吃奶边做爰视频| 免费人成在线观看视频色| 色在线成人网| 一个人免费在线观看电影| 久久精品国产清高在天天线| 免费大片18禁| 亚洲国产日韩欧美精品在线观看| 欧美+亚洲+日韩+国产| www.www免费av| 真人做人爱边吃奶动态| 国内少妇人妻偷人精品xxx网站| 久久久久久久午夜电影| 欧美色视频一区免费| 中文字幕av在线有码专区| 久久精品久久久久久噜噜老黄 | 五月玫瑰六月丁香| 性欧美人与动物交配| 国产精品98久久久久久宅男小说| 久久久精品欧美日韩精品| 精品久久久久久久久久久久久| 国产精品一区二区性色av| 亚洲最大成人av| 午夜福利成人在线免费观看| 久久久久久九九精品二区国产| 舔av片在线| 国产av一区在线观看免费| 熟女电影av网| 88av欧美| 久久久久九九精品影院| 国产高清三级在线| 亚洲 国产 在线| 亚洲国产精品成人综合色| 男女那种视频在线观看| 国产av在哪里看| 桃红色精品国产亚洲av| 亚洲成av人片在线播放无| 91av网一区二区| 91在线精品国自产拍蜜月| 亚洲第一电影网av| 嫩草影视91久久| 最近中文字幕高清免费大全6 | 天天一区二区日本电影三级| 97碰自拍视频| 天美传媒精品一区二区| 久久久久精品国产欧美久久久| 精品人妻偷拍中文字幕| 天美传媒精品一区二区| 九色国产91popny在线| 精品久久久久久久久亚洲 | 成人av一区二区三区在线看| 天美传媒精品一区二区| 中文字幕久久专区| 国产一区二区三区av在线 | www日本黄色视频网| 亚洲av免费高清在线观看| 欧美绝顶高潮抽搐喷水| 色吧在线观看| 美女高潮的动态| 亚洲中文字幕日韩| 不卡一级毛片| 色哟哟哟哟哟哟| 国产一区二区在线av高清观看| 18禁黄网站禁片午夜丰满| 日本熟妇午夜| 久久人人爽人人爽人人片va| 女的被弄到高潮叫床怎么办 | 免费观看人在逋| 老女人水多毛片| 色在线成人网| 精品欧美国产一区二区三| 国产精品亚洲美女久久久| 免费不卡的大黄色大毛片视频在线观看 | 精品人妻1区二区| 亚洲精品久久国产高清桃花| 看片在线看免费视频| 亚洲国产欧洲综合997久久,| 日韩一区二区视频免费看| 女生性感内裤真人,穿戴方法视频| 美女被艹到高潮喷水动态| 国国产精品蜜臀av免费| 日韩在线高清观看一区二区三区 | av专区在线播放| 久久久久久国产a免费观看| 中文字幕高清在线视频| 在线观看美女被高潮喷水网站| 欧美绝顶高潮抽搐喷水| 亚洲av熟女| 亚洲国产精品合色在线| 在线观看午夜福利视频| av在线蜜桃| 国产淫片久久久久久久久| 91久久精品电影网| 国产日本99.免费观看| 2021天堂中文幕一二区在线观| 欧美丝袜亚洲另类 | 内地一区二区视频在线| 免费av毛片视频| 男女那种视频在线观看| 国国产精品蜜臀av免费| 国产成人一区二区在线| 免费搜索国产男女视频| 久久精品91蜜桃| 好男人在线观看高清免费视频| 一区二区三区高清视频在线| 麻豆精品久久久久久蜜桃| 日韩欧美精品v在线| 91精品国产九色| 久久精品夜夜夜夜夜久久蜜豆| 欧美高清性xxxxhd video| 他把我摸到了高潮在线观看| 日韩大尺度精品在线看网址| 嫩草影院精品99| 不卡视频在线观看欧美| 毛片一级片免费看久久久久 | 亚洲国产精品成人综合色| 国产精品1区2区在线观看.| 久久久久九九精品影院| 22中文网久久字幕| 又爽又黄无遮挡网站| 男人舔女人下体高潮全视频| 久久人妻av系列| 天堂√8在线中文| 中文字幕高清在线视频| 亚洲七黄色美女视频| 波野结衣二区三区在线| 亚洲四区av| 精品人妻偷拍中文字幕| 免费看日本二区| 搡女人真爽免费视频火全软件 | 久久久久免费精品人妻一区二区| 69av精品久久久久久| 国产老妇女一区| 亚洲欧美日韩高清专用| 亚洲真实伦在线观看| 欧美成人免费av一区二区三区| 国产蜜桃级精品一区二区三区| 国产大屁股一区二区在线视频| 深夜精品福利| 午夜视频国产福利| 岛国在线免费视频观看| 亚洲中文字幕日韩| 中文字幕精品亚洲无线码一区| 亚洲国产精品sss在线观看| 亚洲精品乱码久久久v下载方式| 我的女老师完整版在线观看| 一级黄片播放器| av天堂在线播放| 欧美成人免费av一区二区三区| 久久久成人免费电影| 久久久精品大字幕| 狂野欧美激情性xxxx在线观看| 国产伦在线观看视频一区| 99久久无色码亚洲精品果冻| 精品一区二区三区人妻视频| 国产精品三级大全| 国产伦在线观看视频一区| 夜夜爽天天搞| 1000部很黄的大片| 干丝袜人妻中文字幕| 在线观看一区二区三区| av在线老鸭窝| 最近视频中文字幕2019在线8| 不卡视频在线观看欧美| 中国美女看黄片| 日韩欧美 国产精品| 欧美丝袜亚洲另类 | 联通29元200g的流量卡| 最近视频中文字幕2019在线8| 国产精品一区二区三区四区免费观看 | 午夜免费激情av| 亚洲专区中文字幕在线| 色av中文字幕| 十八禁网站免费在线| 韩国av一区二区三区四区| 色播亚洲综合网| 身体一侧抽搐| 成人二区视频| 日韩欧美三级三区| 中国美白少妇内射xxxbb| 少妇猛男粗大的猛烈进出视频 | 国产精品亚洲一级av第二区| 成人av在线播放网站| 午夜影院日韩av| 国内久久婷婷六月综合欲色啪| 亚洲内射少妇av| 国产伦在线观看视频一区| 久久午夜福利片| 国产免费一级a男人的天堂| 国产69精品久久久久777片| 久久久成人免费电影| 国内精品久久久久精免费| 中文字幕熟女人妻在线| 免费av不卡在线播放| 在现免费观看毛片| 村上凉子中文字幕在线| 国产aⅴ精品一区二区三区波| a级毛片免费高清观看在线播放| 久久久久久久久久成人| 春色校园在线视频观看| 人人妻人人澡欧美一区二区| 中文字幕人妻熟人妻熟丝袜美| 有码 亚洲区| 老司机深夜福利视频在线观看| 国模一区二区三区四区视频| 国产三级中文精品| 色尼玛亚洲综合影院| 99在线人妻在线中文字幕| 亚洲,欧美,日韩| 97热精品久久久久久| 日韩精品青青久久久久久| 亚洲av电影不卡..在线观看| 波多野结衣高清作品| 国产精品99久久久久久久久| 99riav亚洲国产免费| 欧美+亚洲+日韩+国产| 蜜桃久久精品国产亚洲av| 国产成人福利小说| 少妇高潮的动态图| 熟女人妻精品中文字幕| 久久精品夜夜夜夜夜久久蜜豆| 日日夜夜操网爽| 久久6这里有精品| 亚洲成人久久性| 极品教师在线视频| 少妇的逼好多水| 两性午夜刺激爽爽歪歪视频在线观看| 日韩欧美精品免费久久| av天堂在线播放| 日韩欧美国产在线观看| 俄罗斯特黄特色一大片| 夜夜看夜夜爽夜夜摸| 啦啦啦啦在线视频资源| 久久中文看片网| 国产伦在线观看视频一区| 国产精华一区二区三区| 美女大奶头视频| 国产色婷婷99| 99国产精品一区二区蜜桃av| 22中文网久久字幕| 日本在线视频免费播放| 亚洲av成人av| 免费av观看视频| 极品教师在线免费播放| 尾随美女入室| 日本爱情动作片www.在线观看 | 看免费成人av毛片| 99国产精品一区二区蜜桃av| 亚洲色图av天堂| 久久久久久久精品吃奶| 色综合色国产| 国产高清有码在线观看视频| 91在线观看av| 狂野欧美激情性xxxx在线观看| netflix在线观看网站| 香蕉av资源在线| 亚洲图色成人| 久久精品国产亚洲av香蕉五月| 久久精品国产亚洲av天美| 黄色女人牲交| 欧美日韩亚洲国产一区二区在线观看| 一a级毛片在线观看| 色av中文字幕| 男女做爰动态图高潮gif福利片| 内地一区二区视频在线| 日韩欧美一区二区三区在线观看| 国产精品永久免费网站| 丰满人妻一区二区三区视频av| 黄色欧美视频在线观看| 亚洲精品国产成人久久av| 色综合色国产| av女优亚洲男人天堂| 日本与韩国留学比较| 蜜桃久久精品国产亚洲av| 综合色av麻豆| 免费观看在线日韩| 亚洲av电影不卡..在线观看| 91久久精品国产一区二区三区| 久久香蕉精品热| 老司机午夜福利在线观看视频| 国产成人福利小说| 欧美日韩综合久久久久久 | 亚洲 国产 在线| 日韩欧美国产在线观看| 一夜夜www| 一区二区三区高清视频在线| 亚洲第一电影网av| 免费观看人在逋| 日韩欧美 国产精品| 国产精品永久免费网站| 我要搜黄色片| ponron亚洲| 亚洲天堂国产精品一区在线| 一夜夜www| 日韩一本色道免费dvd| 国产又黄又爽又无遮挡在线| 日日干狠狠操夜夜爽| 97碰自拍视频| 99热这里只有是精品50| 久久6这里有精品| 神马国产精品三级电影在线观看| 亚洲人成网站高清观看| 免费搜索国产男女视频| 2021天堂中文幕一二区在线观| 中文字幕久久专区| 天天躁日日操中文字幕| 亚洲成人精品中文字幕电影| 国产欧美日韩精品一区二区| 亚洲精品久久国产高清桃花| 精华霜和精华液先用哪个| 午夜免费男女啪啪视频观看 | 欧美潮喷喷水| 午夜老司机福利剧场| 五月伊人婷婷丁香| 亚洲午夜理论影院|