• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Research on Micro-blog New Word Recognition Based on SVM

    2017-10-10 11:31:05ChaotingXiaoJianhouGanBinWenWeiZhangandXiaochunCao
    關鍵詞:伐木葛優(yōu)新詞

    Chaoting Xiao, Jianhou Gan, Bin Wen, Wei Zhang, and Xiaochun Cao

    ResearchonMicro-blogNewWordRecognitionBasedonSVM

    Chaoting Xiao, Jianhou Gan*, Bin Wen, Wei Zhang, and Xiaochun Cao

    New word discovery possesses a significant role in the field of Natural Language Processing (NLP). As the effect of mutual information on multi-string is not good, we improve the traditional mutual information and adjacency entropy method respectively and put forward enhancement of mutual information and relative adjacency entropy. As multi-feature massive data brings the problem of slow speed, we use the MapReduce parallel computing model to extract some features, such as, enhancement of mutual information, relative adjacency entropy and background document frequency. With the extracted eight features, the feature vectors of the candidate words are formed, and the SVM model can be trained by the labelled corpus. The experiments show that the proposed method accelerates the computing speed and shortens the time required by the whole recognition process. In addition, comparing with the existing methods, we can see that theFvalue reaches 86.98%.

    new words recognition; Natural Language Processing(NLP); enhanced mutual information; relative adjacency entropy; mapReduce; SVM

    1 Introduction

    In the process of Chinese word segmentation, new word recognition is quite difficult. Sproat and others pointed out that 60% errors of Chinese word segmentation are caused by new words[1]. Now, many new words are spreading via micro-blog. New words such as, ‘伐木累’, ‘葛優(yōu)癱’ and ‘北京癱’, etc, have been created. Micro-blog text contains a considerable proportion of new words, the linguists have concluded according to the statistics that the average annual production of new words is more than 800[2]. In the field of new word recognition, there is no definition for ‘new word’. Based on existing research, people think that new words should have the following properties. From the perspective of word itself, it should be an independent word. From the perspective of appearing frequency, the new word should be widely adopted. Even in corpus, the new word has a high frequency of appearance in many documents and is used by numerous people. From the perspective of time, the word has just appeared within a certain period of time, or it has a new meaning which is ‘the new use of old word’[3].

    At present, new word recognition methods are mainly divided based on the rule, statistics and combination of both rule and statistics. The method based on rule needs to build a rule bank to match via template. The precision is high but manual rule is difficult to write and the cost is high. In addition, the rule is highly related to the field. Meanwhile, the advantages of this kind of method based on statistics include flexiblility, good adaptability and portability. However, it needs a large corpus to calculate the statistics and thus consumes much time. For example, Sui and others[4]extracted the words with close relationship through computing the static union rate among the words after word segmentation of corpus. Then they used the grammar rule and field features to get the field terms with high confidence. The rule only has the features of field, so it is not suitable for other corpora. Sornlertlamvanich and others used decision-tree model to train the new word recognition model, with a precision result of 85%[5]. Unfortunately, it is not suitable for large-scale corpus. Peng[6]and others adopted the statistical method to do unified consideration to do segmentation and new word discovery, by using the CRF model of combining lexical features and field knowledge to extract the new words. At the same time, they added the discovered new words into the dictionary to enhance the recognition effect of the model. The method improves the accuracy of word segmentation but costs a long time. Liu[7]and others applied the left & right information entropy and likelihood rate (LLR) to determine the word boundary to extract the candidate new words. The extracted features of the method are less and the precision rate is not high. Lin and others[8]counted and extracted new words based on word’s internal model and combined with mutual information, IWP and position-word probability. The proposed mutual information is not suitable for multi-strings and there are limitations. Li and other scholars[9]employed word frequency, word probability, etc., to train a SVM model and consider the new recognition from the perspective of classification. The limitation of the method is that it cannot recognize the low-frequency new words, thus it will produce a lot of garbage strings. Zhao and others[10]iteratively used the mutual information, left (right) entropy, left (right) adjacency right (left) average entropy, etc., to obtain the candidate list of new words. Then they used a Chinese collocation library to filter the list to get new words. The limitation of the method is that when obtaining the mutual information, the multi-strings will be divided into two substrings in the calculation. This will affect the results of recognizing the new words. Wang and others[11]explored the new words from the internet based on time series information and used the combination between dynamic feature’s new method and commonly employed statistical method. The method compares the curve’s change trend of each candidate word within a period of time, and states that each part of the new word should have the identical change trend. However, the accuracy of the method is not very high. Shuai and others[12]proposed a filter method to stop word with redefinition and the filter method by using iterative context entropy algorithm to recognize new words and introduce lexical features. The rule dependency of this method is strong. Su and others[13]proposed to improve the adjacency entropy with a weighted adjacency entropy to optimize and improve the performance, and achieved good performance. Li and others[14]exploited the internal word probability, mutual information, word frequency and word probability rule as features to train the SVM model, but the precision rate was only 61.78%.

    Due to the problems of speed caused by the above methods when processing large-scale corpus, as well as the low precision rate caused by the statistical method when recognizing new words, this paper analyzes the micro-blog corpus. We firstly reduces the noise in the micro-blog corpus. Then we use N-Gram statistical method to extract new word candidates based on a word segmentation. We propose a new filtering algorithm and combine it with the stop word list launched by Harbin Institute of Technology to filter the candidates. Then, a SVM classifier will be trained by multiple eigenvalues obtained through improving enhancement of mutual information, relative adjacency entropy and background document frequency method. At last, with the trained SVM model for recognizing the new words of micro-blog of test set. The method improves the speed of new word training recognition model caused by the multi-featured massive data of large-scale corpus statistics, meanwhile, the proposed method can also improve the precision rate of new word recognition. The detailed process is as shown in Fig.1.

    2 New Word Discovery Method Based on Micro-blog Content

    2.1Preprocessingofthecorpus

    Micro-blog has a strong randomness in the word use and grammar, and causes a large number of noisy data. It will affect the feature extraction and increase the model training time. According to the features of the micro-blog corpus, this paper reduces the negative influence of noisy data in the process of training. Through statistics and analysis, we find that many contents are accompanied with topic and expression labels, etc. The specific labels are shown in Table 1.

    Fig.1 Overall flow chart of micro-blog new words extraction.

    LabelDescription#words#Hottopic,‘words’meanthekeywordsofthetopic@nameRemindauser,‘name’meanstheus?ernametobereminded【sentence】Micro?blogtheme,‘sentence’isthesummaryortitleofthemainbodyofmicro?blog[word]Expressionword,‘word’meanthemindandemotion,etc,oftheauthororpaper

    Based on the analysis of micro-blog corpus, we can find that, ‘@’, [expression] and URL links exist in most of the micro-blog content. ‘@’ is usually followed by a user name, where many usernames are random, so there is zero possibility to appear a new word. [Expression] and URL link labels will also have a zero possibility of new words. These noisy data have great influences on the generation of candidate words, so it is necessary to eliminate these noises. This paper eliminates the above three kinds of labels through the method of building regular expression of the micro-blog data.

    2.2 Filtering algorithm

    We introduce N-gram algorithm for preprocessed micro-blog corpus data, the basic idea of the method is to carry out N-size sliding window operation to text content. Based on the word segmentation of the micro-blog data, we use the N-gram to recognize the candidate words. If the corpus is ‘中文新詞識別’, the result after word segmentation will be ‘中文新/新詞/詞識別’. WhenNis 2, the candidate word will be ‘中文新/新詞/詞識別’. WhenNis 3, the candidate word will be ‘中文新詞/新詞識別’. We extract all the candidate new words from micro-blog corpus under the conditions ofNis 2, 3 and 4. These candidate new words contain many garbage strings, so they need to be filtered. After analyzing the news corpus provided by Fudan University, we conclude that the content of corpus are news before 2002, where the language was formal and without oral language and the new words from the internet were rare. Therefore, this paper proposes a filtering algorithm of combining news corpus and stop words. The pseudo algorithm is shown in Table 2, whereNis the news corpus,Wstands for the candidate new word set of micro-blog,Tis the stop word list andNLmeans the candidate new word set after filtering.

    3 Feature Selection of Candidate New Words

    After the filtering algorithm, the candidate words still have some noises, such as: ‘富美喜’, ‘女票不’ and ‘逼格真’. Therefore, we use the statistical method to quantify the features of these candidate new words. The employed statistical method includes mutual information[10]which can measure internal coagulation. Information entropy[10]and background document frequency can measure the external freedom degree.

    Table 2 Filtering algorithm.

    3.1 Internal coagulation

    Mutual information[10]measures the correlation between two events. Traditional mutual information formula[10]only gives the calculation formula of two character strings, which can only be applied to the two character new words. For the multi-character string, if the micro-blog candidate new wordsS={s1s2…sn}, the common method nowadays is to take the two longest substringsSleft={s1s2…sn-1} andSright={s2s3…sn}. A high correlation between the two longest substrings of a candidate new word S shows that the combination of the two will be more closely, and S is more likely to be a word. On the contrary, a lower correlation between them indicates that they are less dependent on each other. However, the mutual information value obtained under this condition is not very accurate. Under current circumstances, this paper aims to improve the traditional mutual information formula[10]and proposes enhancement mutual information which is suitable for the multi-word strings, the definition is as follows.

    (1)

    whereWis the total words of micro-blog corpus andPis the frequency of string in the corpus. The greater the value of the enhanced mutual information, the higher the possibility of the current string is a new word.

    3.2 External freedom degree

    Information entropy can reflect the average information content brought by an event’s results. Intuitively, the meaningful new words are not only repeated in the text, but also appeared in different context, which reflects the string’s independent ability and the freedom degree of usage. However, more appearances of a string in the corpus, it is more probable to have a larger value. Therefore, the adjacency entropy is not conducive to the low-frequency strings. This paper proposes a relative adjacency entropy. In our opinion, the string with a higher word probability than its substring will be regarded as a new word. For stringW= {w1w2…wn} and its longest substringWleft= {w1w2…wn-1} andWright= {w2w3…wn}. We subtract the adjacency entropy after taking the weight and the substring’s adjacency entropy, and take the minimum of relative adjacency entropy.

    Respectively, we defineα={α1,α2,…,αn} andβ={β1,β2,…,βn} to be the context sets of the candidate repeated stringsωin the corpusX. The entropy ofωin the left, right and context of the corpusXis defined as follows.

    (2)

    (3)

    The adjacency entropy after taking the weight and relative adjacency entropy is defined as follows.

    Cr(ω)=λCr(ω)-(1-λ)Cr(ωleft)

    (4)

    CL(ω)=λCL(ω)-(1-λ)CL(ωright)

    (5)

    The minimum formula of relative adjacency entropy is defined as:

    C(ω)=min{CL(ω),Cr(ω)}

    (6)

    From the above definitions we can conclude that for the character strings that are only used in the fixed context, the relative adjacency entropy is small. On the contrary, the relative adjacency entropy is big for character strings that are used in many different contexts.

    3.3 Background document frequency

    Considering from the perspective of human memory, we think that new words have never appeared in previous memories. We use a large-scale background corpus to simulate the human memory to compare the frequency of string in the background corpus and the string in the corpus of the extracted string (foreground corpus). If the frequency of the string in foreground corpus is much larger than that in the large-scale background corpus, the string is likely to be a new word. This method is also useful for high-frequency word adhesion, i.e., the repeated strings and garbage filtering, such as: ‘也不’ and ‘了一’, etc. The frequencies of these strings in the foreground corpus and the background corpus are similar. The formula of the relevant frequency ratio of stringωinXandYof two corpora is defined as follows.

    (7)

    wheref(ω,X) andf(ω,Y) are the corresponding frequencies ofωinXandYof corpora,Xis the foreground corpus andYis the background corpus.

    3.4 Dice

    The Dice of candidateWis estimated by Equation 8. In Equation 8,xidenotes the characters in candidatew. For example,w=x1,x2, …,xn.

    (8)

    3.5 SCP

    The SCP of candidatewis estimated by Equation 9. In Equation 9,xidenotes the characters in candidatew.

    (9)

    4 Parallel Implementation of New Word Feature Quantization Algorithm

    4.1MapReduceparallelcomputingmodel

    MapReduce is a programming model, which is divided into Map and Reduce two stages. The input and output of each stage is based on key-value pairs. In Map stage, Map function changes each line of input to the key-value pairs (K1,V1) form. After the processing of the Map function, it outputs many new key-value pairs List (K2,V2). In Reduce stage, all output in Map stage will be divided according to the key (K2, List(V2)). This process is called shuffle. Each group (K2, List (V2)) is the input of the Reduce function. After the processing of Reduce function, it outputs the final results (K3,V3). Overall new word recognition speed will be affected due to the large scale of the micro-blog corpus and the large amount of time to perform the new word feature quantization algorithm. Thus we parallel the calculations of the background document frequency, the relative adjacency entropy and the enhanced mutual information algorithm with the MapReduce model.

    4.2Parallelimplementationofbackgrounddocumentfrequencyalgorithm

    Background document frequency mainly refers to the ratio between the candidate wordsw’s frequencies inXof the foreground corpus and inYof the background corpus. In order to improve the coupling efficiency of the multi-features, we calculate the word frequencies inXandYcorpus respectively and calculate the frequency ratio in the multi-features coupling Reduce. The calculation process of the background document frequency of the candidate word inXcorpus is as follows. First, we segmentXcorpus via Split and transfer each segmentation to map function. Then we input the candidate word setWto the Map function via configuration method. In Map function, we calculate the frequencykof each candidate word in the corpus fragments, and use output key as the candidate wordw. The specific pseudo code of the algorithm is shown in Table 3, where it outputs Value as frequencyk. In Reduce function, the same key values are accumulated to get the key value, which is the frequency of the candidate wordwinXcorpus. Then it is divided byXcorpus size to obtain the frequency of the candidate word which is the output. The specific algorithm is shown in Table 4. The overall system diagram is shown in Fig.2.

    Table3Mapoperationonparallelizationofbackgrounddocumentfrequency.

    Inwhich,theimplementationprocessofstep(3)is:1.InputW[],〈k1,v1〉2.Forifrom0tolengthofW3.While(index!=?1)4.Index

    Table4Reduceoperationonparallelizationofbackgrounddocumentfrequency.

    Theimplementationprocessofstep(4)is:1.Input〈W[i],List〈k〉〉2.Forxfrom0tolengthofList〈k〉3.sum

    Fig.2 Parallelization of string frequency

    4.3Parallelizationofrelativeadjacencyentropy

    In order to obtain relative adjacency entropy, the context entropy should be obtained first. The context entropy can be classified as the left and right entropy respectively, where their algorithm is similar. Here we introduce the algorithm processed by taking the left entropy as an example. After theXcorpus is segmented via split, each segmentation is inputted to Map. We employ the Configuration method to input the candidate word setWto Map function. Then we find the candidate wordw’s adjacent left character in Map. We setupwfor Key. Value is setup to be the adjacent left characteraas the output. The specific algorithm is shown in Table 5. In Reduce, we count the number of the same left character of candidate word, then obtain the left entropy value of the candidate word and output the left entropy value of each candidate word. The specific algorithm implementation is shown in Table 6. The overall process is shown in Fig.3.

    Table5Mapoperationonparallelizationofrelativeadjacencyentropy

    Inwhich,theprocessofstep(3)is:1.InputW[],1then6.a(chǎn)

    4.4Multi-featuredatacouplingandSVM

    After obtaining the feature data of the candidate new word (such as: the enhanced mutual information, the relative adjacency entropy and the background document frequency), we need to couple the multiple feature data which can form feature vector. First, the data is from different files such as the relative adjacency entropy, the enhanced mutual information and the background document frequency. Each mapper already knows the file name of data stream processed by it. Here it is the key wordwand is marked by the file name. After sealing each input function, map will implement division, shuffle and sort operation indicated by Mapreduce. The Reduce function receives the input data and carries out the complete cross product to the value. The Reduce function generates all consolidated results of these values and limits each value to be marked at most once in each consolidation. The overall process is shown in Fig.4.

    Table6Reduceoperationonparallelizationofrelativeadjacencyentropy.

    Theprocessofstep(4)is:1.Input〈W[i],List〈a〉〉2.Forxfrom0tolengthofList〈a〉3.IfList〈a〉.xnotinA〈k,v〉then//IfthereisnoList〈a〉.xinA〈k,v〉4.A〈k,v〉

    Compared to other classifiers, SVM has better classification results and is a popular statistical machine learning method. We map the sample points from low-dimension to high-dimension feature space and find a super plane, such that the distance between each class of data and hyper plane is the maximum (namely, the optimal hyper plane). The final results will be different if the kernel function is different. For the feature vector composed of the multi-feature data of the candidate new word obtained through the above methods, this paper employs 70% of the multi-feature data as the training set and 30% of the multi-feature data as the testing set. Labeled in a manual manner, we train the SVM micro-blog new word recognition model and carry out 10-fold cross-validation.

    Fig.3 Parallelization of Relative Adjacency Entropy

    Fig.4 Overall parallelization process

    5 Experimental Results and Analysis

    In the experiment, we employ a 5.2 million micro-blog corpus, which includes 591 micro-blog of 2009, 60795 micro-blog of 2010, 763027 micro-blog of 2011, 1699484 micro-blog of 2012, 17882335 micro-blog of 2013, 681449 micro-blog of 2014 and 198925 micro-blog of 2015. We make the micro-blog from 2009 to 2013 as the background corpus to simulate human memory. We let the micro-blog of 2014 and 2015 be the foreground corpus. The news corpus which is used by the filtering algorithm is from Fudan University, which includes 9804 articles. The results of the candidate new word will be obtained by the filtering algorithm of combining the news corpus with stop word. The preliminary results have a total of 13273 words. In relative adjacency entropy algorithm, we set the weight to be 0.62. Through the experiment we find that the proposed method can avoid the new word to be deleted accidentally via the reduction of weight of substring. The hardware of the parallelization improvement is one set of Lenovo computer, with CPU of Celerondual-core T30001.80Ghz and 16G memory. The distributed environment is simulated through running VirtualBox virtual machine on the computer. In total, there are six computers with memory of 512M that are virtual, and we installed the CentOS operation system for each computer. The experimental platform is Eclipse and developed by java language. Kernel function of SVM uses the RBF kernel function.

    5.1 Methods and standards

    This method employs the precision rate (P), recall rate (R) andF-measure to measure the experimental results. The specific definitions of the measurements are as follows.

    (10)

    (11)

    (12)

    In formula (10),Rmeans the recall rate of the recognized new word. In formula (11),Pmeans the precision rate of the recognized new word. In formula (12),β=1 while theFvalue is the harmonic mean of the precision rate and the recall rate. These measurements can comprehensively reflect the overall performance of the new word recognition.

    5.2 Experimental results and analysis

    Comparing the advantages and disadvantages between the proposed method and previous methods, the specific experimental results are shown in Table 7. In Table 7, BF represents the Background document frequency.MI*represents the enhanced mutual information.E*means the relative adjacency entropy. F represents the frequency. We can conclude from the experiment that both the precision rate and recall rate of the enhanced mutual information are improved when comparing (MI+F+Dice+SCP+LCE+RCE) with (MI*+F+Dice+SCP+LCE+RCE). Similarly, both the precision rate and recall rate of (MI*+F+Dice+SCP+LCE+RCE) are higher than that of (MI*+F+Dice+SCP+LCE+RCE+E*). Via the comparison between (MI*+F+Dice+SCP+LCE+RCE+E*) and (MI*+F+Dice+SCP+LCE+RCE+E*+BF), we can observe that the new word recognition accuracy without the BF is lower. According to the results, the enhanced mutual information and the relative adjacency entropy have positive effects. The combination of some features improves the precision rate, recall rate and F value of micro-blog new word recognition when comparing with traditional method (MI+F+Dice+SCP+LCE+RCE). The finalFvalue is 86.98%. Experimental results of new word recognition are shown in Table 8 and experimental results of non-new word recognition are shown in Table 9.

    Table 7 Experimental results of micro-blog new word recognition. %

    Table 8 Experimental results of new word recognition.

    Table 9 Experimental results of non-new word recognition.

    The parallelization algorithm sets the processing speed under the node of one, three and six sets. We obtain the operation situation of micro-blog corpora with different sizes to count the time spent by the system when recognizing the new words. The results are shown in Fig.5. Experimental results indicate that when we increase the number of the node machines, and the recognition speed of the micro-blog new word also increases.

    Fig.5Operationspeedchartonmicro-blognewwordrecognition.

    6 Conclusion

    We improve the traditional mutual information and adjacency entropy method respectively and put forward the enhancement of mutual information and relative adjacency entropy. After experimental verification, the parallelization shortens the overall time of new word recognition. The precision rate and recall rate of micro-blog new word recognition are improved by a trained SVM classification model with the features generated from the above methods. The proposed method can achieve very good classification and recognition performance. In the future, we will try to fully explore the effective new word detection features and adopt them into the model to further improve the performance of the micro-blog new word recognition task.

    Acknowledgment

    The research is supported by National Key Research and Development Plan (No. 2016YFB0800603), National Natural Science Foundation of China (No. 61562093, 61422213, 61650202), Key Project of Applied Basic Research Program of Yunnan Province No.2016FA024, Key Program of the Chinese Academy of Sciences (No. QYZDB-SSW- JSC003).

    [1]R.Sproat and T.Emerson, The first international Chinese word segmentation bake off, inProceedingsofthesecondSIGHANworkshoponChineselanguageprocessing, Sapporo, Japan, vol.17,pp.133-143,2003.

    [2]Zhang Dexin.‘There will be no fish if the water is too clear ’-My Normative View on New Words,JournalofPekingUniversity:PhilosophyandSocialScience, vol.37, no.5, pp.106-119, 2000.

    [3]X.Huang and R.F.Li, Discovery Method of New Words in Blog Contents,ModernElectronicsTechnique, vol.36,no.2, pp.144-146, 2013.

    [4]Z.F.Sui, Y.R.Chen, and Y.R.Wu, etal.The Research on the Automatic Term Extraction in the Domain of Information Science and Technology.http://icl.pku.edu.cn/icl_tr/ papers_2000-2003 /2002/E026-szf-The Research on the Automatic Term Extraction in the Domain of Information Science and Technology.pdf

    [5]V.Sornlertlamvanich, T.Potipiti, and T.Charoenporn, Automatic Corpus-Based Thai Word Extraction with the C4.5 Learning Algorithm,inProceedingsofInternationalConferenceonComputationalLinguistics, Germany,vol.2, pp.802-807, 2000.

    [6]F.Peng, F.Feng, and A.McCallum.Chinese segmentation and new word detection using conditional random fields, inProceedingsofthe20thinternationalconferenceonComputationalLinguistics, Switzerland,pp.562,2004.

    [7]T.Liu, B.Q.Liu, Z.M.Xu, and X.L.Wang.Automatic domain-specific term extraction and its application in text classification,ActaElectronicaSinica, vol.35, no.2, pp.328, 2007.

    [8]Z.F.Lin and X.F.Jiang. New Word Recognition Based on Internal Model of Word,ComputerandModernization, no.11,pp.56-58, 2010.

    [9]H.Li, C.N.Huang, J.Gao, and X Fan, The Use of SVM for Chinese New Word Identification, inProceedingsofFirstInternationalJointConferenceonNaturalLanguageProcessing, Sanya,China, pp.723-732,2004.

    [10] X.B.Zhao and H.P.Zhang, New Word Recognition Based on Iterative Algorithm,ComputerEngineering, vol.40, no.7, pp.154-158, 2014.

    [11] M.Wang, L.Lin, and F.Wang.New word identification in social network text based on time series information, inProceedingsofIEEEInternationalConferenceonComputerSupportedCooperativeWorkingDesign, IEEE Press, pp.552-557,2014.

    [12] C.Xiao, J.Gan,and B.Wen, et al.New Word Recognition Based on Micro-blog Contents,PatternRecognitionandArtificialIntelligence, vol.27,no.2, pp.141-145, 2014.

    [13] Q L Su and B Q Liu.Chinese new word extraction from MicroBlog data, inProceedingsofInternationalConferenceonMachineLearningandCybernetics, Tianjin, China: IEEE Press,pp.1874-1879,2013.

    [14] C Li and Y Xu.Based on Support Vector and Word Features New Word Discovery Research,TrustworthyComputingandServices.Berlin, Germany: Springer Press, pp.698-701, 2012.

    JianhouGanreceived Ph.D. degree in Metallurgical Physical Chemistry from Kunming University of Science and Technology, China, in 2016. In 1998, he was a faculty member at Yunnan Normal University, China. Currently, he is professor in Yunnan Normal University, China. He has published over 40 refereed journal and conference papers.His research interest covers education informalization for nationalities, semantic Web, database, intelligent information processing.

    BinWenreceived Ph.D. degree in computer application technology from China University of Mining & Technology, Beijing, China, in 2013. In 2005,he was a faculty member at Yunnan Normal University, China.Currently, he is associate professor in Yunnan Normal University,China. Now, his research interest covers intelligent information processing and emergency management.WeiZhangis now an Assistant Professor in Institute of Information Engineering, Chinese Academy of Sciences, Beijing 100093, China. He received his Ph.D degree from Department of Computer Science in City University of Hong Kong, Hong Kong, China, in 2015. Before joining Chinese Academy of Sciences, he was a visiting scholar in DVMM group of Columbia University, New York, NY, USA, in 2014. His research interests include large-scale visual instance search and mining, multimedia and digital forensic analysis. He has won the second place in TRECVID Instance Search task in 2012, the Best Demo Award in ACM-HK openday 2013.

    XiaochunCaoreceived the B.E. and M.E. degrees in computer science from Beihang University, Beijing, China, and the Ph.D. degree in computer science from the University of Central Florida, Orlando, FL, USA. He has been a Professor with the Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China, since 2012. He spent about three years with ObjectVideo Inc., as a Research Scientist. From 2008 to 2012, he was a Professor with Tianjin University, Tianjin, China. He has authored and co-authored over 120 journal and conference papers. He is a fellow of the IET. He is on the Editorial Board of the IEEE Transactions of Image Processing. His dissertation was nominated for the University of Central Floridas university-level Outstanding Dissertation Award. In 2004 and 2010, he was a recipient of the Piero Zamperoni Best Student Paper Award at the International Conference on Pattern Recognition.

    2016-12-20; accepted:2017-01-20

    M.S. degree in computer application technology from Yunnan Normal University, China, in 2017. Now, His research interest covers Pattern recognition and natural language processing.

    ?Chaoting Xiao and Bin Wen are with School of Information Science and Technology, Yunnan Normal University, Kunming, Yunnan, China. E-mail: xiaochaoting@gmail.com; wenbin@ynnu.edu.cn.

    ?Jianhou Gan is with Key Laboratory of Educational Informatization for Nationalities(Yunnan Normal University),Ministry of Education, Kunming,Yunnan, China. E-mail: ganjh@ynnu.edu.cn.

    ?Chaoting Xiao, Wei Zhang and Xiaochun Cao are with Institute of Information Engineering,Chinese Academy of Sciences, Beijing, China. E-mail: wzhang.cu @ gmail. com, caoxiaochun@iie.ac.cn.

    *To whom correspondence should be addressed. Manuscript

    猜你喜歡
    伐木葛優(yōu)新詞
    葛優(yōu)一句話懟空談者
    做人與處世(2020年1期)2020-09-22 01:27:55
    《微群新詞》選刊之十四
    WE ARE伐木累 伐木累戰(zhàn)隊
    焚詩記
    詩潮(2018年3期)2018-03-26 12:29:30
    跟蹤導練(三)等
    伐木壘——戶外墻體涂鴉 Family=家庭
    童話世界(2016年26期)2016-08-22 12:17:48
    『葛優(yōu)躺』其實很傷身
    葛優(yōu)兩口子
    海峽姐妹(2015年3期)2015-02-27 15:09:58
    小議網(wǎng)絡新詞“周邊”
    語文知識(2014年12期)2014-02-28 22:01:18
    外教新詞堂
    一二三四社区在线视频社区8| 在线观看日韩欧美| 精品久久久久久久久久免费视频| 欧美激情久久久久久爽电影| 亚洲成人中文字幕在线播放| 欧美激情在线99| 热99re8久久精品国产| 亚洲天堂国产精品一区在线| 全区人妻精品视频| 久久中文看片网| 亚洲五月天丁香| 91麻豆av在线| 少妇高潮的动态图| 一区福利在线观看| 久久天躁狠狠躁夜夜2o2o| 婷婷丁香在线五月| 国产精品一区二区三区四区免费观看 | 内地一区二区视频在线| 国产激情欧美一区二区| 在线a可以看的网站| 日韩成人在线观看一区二区三区| 深爱激情五月婷婷| 无人区码免费观看不卡| 脱女人内裤的视频| 日韩欧美 国产精品| 欧美黄色片欧美黄色片| 国产精品嫩草影院av在线观看 | 中国美女看黄片| 特大巨黑吊av在线直播| 黄色成人免费大全| 美女黄网站色视频| 一级作爱视频免费观看| 亚洲中文日韩欧美视频| 久久人人精品亚洲av| 亚洲狠狠婷婷综合久久图片| 欧美日韩一级在线毛片| 日韩欧美在线乱码| 午夜影院日韩av| 欧美一区二区国产精品久久精品| 很黄的视频免费| 天堂av国产一区二区熟女人妻| 两个人的视频大全免费| 亚洲av日韩精品久久久久久密| 成人av在线播放网站| 亚洲在线观看片| 亚洲中文字幕日韩| 高清日韩中文字幕在线| 日韩欧美精品免费久久 | 婷婷精品国产亚洲av| 亚洲美女黄片视频| 国产亚洲欧美在线一区二区| 亚洲黑人精品在线| 成年人黄色毛片网站| 国产成人av教育| 国产视频内射| 欧美另类亚洲清纯唯美| 国产久久久一区二区三区| 天堂影院成人在线观看| 欧洲精品卡2卡3卡4卡5卡区| 两个人看的免费小视频| 欧美高清成人免费视频www| 手机成人av网站| 国产三级在线视频| 久久久成人免费电影| 丰满乱子伦码专区| 亚洲黑人精品在线| 日韩有码中文字幕| 亚洲第一电影网av| 国产伦在线观看视频一区| 国产极品精品免费视频能看的| 久久久久久九九精品二区国产| 亚洲精品日韩av片在线观看 | 久久久久性生活片| 在线播放无遮挡| av天堂在线播放| 噜噜噜噜噜久久久久久91| 热99re8久久精品国产| 午夜免费男女啪啪视频观看 | 蜜桃亚洲精品一区二区三区| www日本在线高清视频| 18禁黄网站禁片免费观看直播| 性色av乱码一区二区三区2| 成人鲁丝片一二三区免费| 99国产精品一区二区三区| 亚洲精品日韩av片在线观看 | 99久久99久久久精品蜜桃| 99国产极品粉嫩在线观看| 99国产综合亚洲精品| 性欧美人与动物交配| 一个人看视频在线观看www免费 | 久久久久久久久中文| 神马国产精品三级电影在线观看| 亚洲性夜色夜夜综合| 欧美3d第一页| 在线a可以看的网站| 精品乱码久久久久久99久播| 国产成人av教育| 国产亚洲精品av在线| 久久人人精品亚洲av| 91在线精品国自产拍蜜月 | 亚洲精品美女久久久久99蜜臀| 久久久国产成人免费| 国产精品野战在线观看| 国产男靠女视频免费网站| 我要搜黄色片| 国产精品 欧美亚洲| 熟女电影av网| 国产一区二区在线av高清观看| 热99re8久久精品国产| 亚洲无线观看免费| 少妇丰满av| 午夜福利在线观看吧| 国产高清三级在线| av中文乱码字幕在线| 欧美日韩乱码在线| 真实男女啪啪啪动态图| 国产精品久久久久久亚洲av鲁大| 日本免费a在线| 欧美日韩亚洲国产一区二区在线观看| 国产午夜福利久久久久久| 国产一区二区在线av高清观看| 免费在线观看亚洲国产| 老熟妇乱子伦视频在线观看| 老汉色av国产亚洲站长工具| 国产69精品久久久久777片| 免费看光身美女| 丰满的人妻完整版| 国产中年淑女户外野战色| 久久精品亚洲精品国产色婷小说| 日韩 欧美 亚洲 中文字幕| 中亚洲国语对白在线视频| 国产色爽女视频免费观看| 亚洲电影在线观看av| 国产一区二区激情短视频| 亚洲人成伊人成综合网2020| 一个人看的www免费观看视频| 香蕉av资源在线| 好男人在线观看高清免费视频| a级一级毛片免费在线观看| 男女午夜视频在线观看| 亚洲黑人精品在线| a级一级毛片免费在线观看| 欧美日本亚洲视频在线播放| 一边摸一边抽搐一进一小说| 久久精品国产综合久久久| 国产成人系列免费观看| 青草久久国产| 日韩国内少妇激情av| 男女下面进入的视频免费午夜| 一区二区三区免费毛片| 在线播放无遮挡| 欧美bdsm另类| 精品熟女少妇八av免费久了| 国产精品1区2区在线观看.| 日韩欧美三级三区| 欧美中文日本在线观看视频| 亚洲电影在线观看av| 欧美性猛交黑人性爽| 老司机午夜福利在线观看视频| 国产精品99久久99久久久不卡| 黄片小视频在线播放| 午夜激情欧美在线| 欧美激情久久久久久爽电影| 国产一区二区激情短视频| 又爽又黄无遮挡网站| 又紧又爽又黄一区二区| 99国产综合亚洲精品| 日本三级黄在线观看| 午夜福利在线观看免费完整高清在 | 噜噜噜噜噜久久久久久91| 成人亚洲精品av一区二区| 精品一区二区三区视频在线观看免费| 亚洲精品国产精品久久久不卡| 国产精华一区二区三区| 成年女人毛片免费观看观看9| 99热这里只有精品一区| 淫妇啪啪啪对白视频| 免费av观看视频| 两个人的视频大全免费| 精品乱码久久久久久99久播| 欧美成人一区二区免费高清观看| 欧美中文日本在线观看视频| netflix在线观看网站| 人妻夜夜爽99麻豆av| 亚洲,欧美精品.| 一区二区三区免费毛片| 法律面前人人平等表现在哪些方面| 成人高潮视频无遮挡免费网站| 午夜福利在线在线| 99riav亚洲国产免费| 在线国产一区二区在线| 一a级毛片在线观看| 久久午夜亚洲精品久久| 日韩精品中文字幕看吧| 97碰自拍视频| 欧美zozozo另类| 啦啦啦韩国在线观看视频| 精品久久久久久,| 在线观看午夜福利视频| 伊人久久精品亚洲午夜| 国产私拍福利视频在线观看| 少妇的丰满在线观看| 在线天堂最新版资源| 欧美性猛交黑人性爽| 国产精品 欧美亚洲| 在线观看av片永久免费下载| 国产精华一区二区三区| 亚洲色图av天堂| 琪琪午夜伦伦电影理论片6080| 久久精品国产99精品国产亚洲性色| 国产精品亚洲一级av第二区| 性色av乱码一区二区三区2| 精品久久久久久久久久免费视频| 精品国产三级普通话版| xxxwww97欧美| 久久久久久九九精品二区国产| 久久久久久久精品吃奶| 欧美高清成人免费视频www| 精品国产美女av久久久久小说| 真人一进一出gif抽搐免费| 特级一级黄色大片| 麻豆国产97在线/欧美| 亚洲男人的天堂狠狠| 日韩欧美 国产精品| 女人十人毛片免费观看3o分钟| 天堂网av新在线| 午夜福利18| 国产高清有码在线观看视频| 波多野结衣高清作品| 亚洲美女视频黄频| 两个人视频免费观看高清| tocl精华| 色吧在线观看| 俄罗斯特黄特色一大片| 亚洲精品456在线播放app | 深爱激情五月婷婷| 欧美另类亚洲清纯唯美| 亚洲av五月六月丁香网| 久久婷婷人人爽人人干人人爱| 母亲3免费完整高清在线观看| 精品日产1卡2卡| 日本黄色视频三级网站网址| 国产精品日韩av在线免费观看| 欧美3d第一页| 久久久久国产精品人妻aⅴ院| 精品人妻偷拍中文字幕| 亚洲av成人av| 国产精品美女特级片免费视频播放器| netflix在线观看网站| 精品久久久久久,| 午夜福利在线观看吧| 日韩大尺度精品在线看网址| 亚洲精品一区av在线观看| 久久久久久大精品| 麻豆成人午夜福利视频| 国产成人a区在线观看| 全区人妻精品视频| 宅男免费午夜| 亚洲欧美激情综合另类| 欧美激情久久久久久爽电影| 亚洲av免费在线观看| 非洲黑人性xxxx精品又粗又长| 亚洲va日本ⅴa欧美va伊人久久| 91麻豆精品激情在线观看国产| 最近在线观看免费完整版| 国产熟女xx| 他把我摸到了高潮在线观看| 久99久视频精品免费| 看黄色毛片网站| 国产乱人视频| 淫秽高清视频在线观看| 99热这里只有精品一区| 国产精品综合久久久久久久免费| 床上黄色一级片| 日本黄色片子视频| 波野结衣二区三区在线 | 高清毛片免费观看视频网站| 国产精品亚洲av一区麻豆| 亚洲va日本ⅴa欧美va伊人久久| 亚洲美女黄片视频| 亚洲精品影视一区二区三区av| 一级毛片女人18水好多| 欧美日本亚洲视频在线播放| 免费看十八禁软件| 看片在线看免费视频| 国产爱豆传媒在线观看| 欧美午夜高清在线| 1024手机看黄色片| 国产又黄又爽又无遮挡在线| 老司机在亚洲福利影院| 中文亚洲av片在线观看爽| 午夜福利18| 九色成人免费人妻av| 久久亚洲真实| 国产色婷婷99| ponron亚洲| 国内毛片毛片毛片毛片毛片| 色综合欧美亚洲国产小说| 动漫黄色视频在线观看| 久久久久久久午夜电影| 狂野欧美激情性xxxx| 亚洲av二区三区四区| 国产高清视频在线观看网站| 非洲黑人性xxxx精品又粗又长| 99国产精品一区二区三区| 国产91精品成人一区二区三区| 在线观看日韩欧美| 内射极品少妇av片p| 精品国产美女av久久久久小说| 别揉我奶头~嗯~啊~动态视频| 国产伦一二天堂av在线观看| 久久婷婷人人爽人人干人人爱| 别揉我奶头~嗯~啊~动态视频| 久久久久免费精品人妻一区二区| 噜噜噜噜噜久久久久久91| 精品国内亚洲2022精品成人| 天堂√8在线中文| 国产色爽女视频免费观看| 婷婷六月久久综合丁香| 久久伊人香网站| 国产一区在线观看成人免费| av片东京热男人的天堂| 亚洲精品在线观看二区| 国产视频内射| 久久欧美精品欧美久久欧美| 中文字幕精品亚洲无线码一区| 两个人的视频大全免费| 成人性生交大片免费视频hd| 久久久久久九九精品二区国产| 欧美成狂野欧美在线观看| 国产视频一区二区在线看| 老鸭窝网址在线观看| 天天一区二区日本电影三级| 国产黄色小视频在线观看| 欧美日本视频| 欧美性猛交╳xxx乱大交人| 国产av麻豆久久久久久久| 国产高清视频在线观看网站| 国产色婷婷99| 有码 亚洲区| 淫妇啪啪啪对白视频| 美女黄网站色视频| 男女视频在线观看网站免费| 欧美精品啪啪一区二区三区| 搡老妇女老女人老熟妇| www日本在线高清视频| 一个人免费在线观看电影| 国产午夜精品论理片| 噜噜噜噜噜久久久久久91| 麻豆一二三区av精品| 亚洲精品国产精品久久久不卡| 熟女少妇亚洲综合色aaa.| 色哟哟哟哟哟哟| 免费av毛片视频| 日本免费一区二区三区高清不卡| 亚洲欧美一区二区三区黑人| 国产成人av激情在线播放| 精品久久久久久,| 亚洲成人久久爱视频| 在线天堂最新版资源| 嫩草影院精品99| svipshipincom国产片| 亚洲美女黄片视频| 久9热在线精品视频| 国产精品1区2区在线观看.| 亚洲av一区综合| 天堂网av新在线| 18禁黄网站禁片免费观看直播| 日韩欧美精品免费久久 | 桃红色精品国产亚洲av| 看免费av毛片| 丁香欧美五月| 久久久久国内视频| 午夜免费男女啪啪视频观看 | 亚洲中文字幕一区二区三区有码在线看| 成人特级av手机在线观看| 国产真实乱freesex| 三级男女做爰猛烈吃奶摸视频| 久久亚洲精品不卡| 99精品久久久久人妻精品| 99国产精品一区二区三区| 色综合亚洲欧美另类图片| 亚洲最大成人手机在线| 一区二区三区国产精品乱码| 五月伊人婷婷丁香| 狂野欧美白嫩少妇大欣赏| 看片在线看免费视频| 国产真实伦视频高清在线观看 | 亚洲人成网站在线播| 欧美日韩黄片免| 最新中文字幕久久久久| 欧美成狂野欧美在线观看| 精品久久久久久久人妻蜜臀av| 最新中文字幕久久久久| 免费在线观看亚洲国产| netflix在线观看网站| www.色视频.com| svipshipincom国产片| 国产午夜福利久久久久久| 亚洲aⅴ乱码一区二区在线播放| 中文字幕久久专区| 国产精品,欧美在线| 精品午夜福利视频在线观看一区| 精品国产亚洲在线| 日本a在线网址| 国产伦精品一区二区三区四那| 丰满的人妻完整版| 日本撒尿小便嘘嘘汇集6| 日韩中文字幕欧美一区二区| 亚洲av第一区精品v没综合| 99热精品在线国产| 99久久精品热视频| 夜夜夜夜夜久久久久| 我的老师免费观看完整版| 免费搜索国产男女视频| 亚洲无线观看免费| 麻豆国产97在线/欧美| 久久久久久久亚洲中文字幕 | 女人十人毛片免费观看3o分钟| 18禁美女被吸乳视频| 人人妻,人人澡人人爽秒播| 亚洲国产精品999在线| 噜噜噜噜噜久久久久久91| 欧美性猛交黑人性爽| 午夜精品在线福利| 麻豆一二三区av精品| 国产精品av视频在线免费观看| 欧美在线一区亚洲| 99久久综合精品五月天人人| 成人精品一区二区免费| 欧美三级亚洲精品| 久久久久久久午夜电影| 日韩欧美在线二视频| 精品一区二区三区视频在线观看免费| 国产黄a三级三级三级人| 欧美中文日本在线观看视频| 国产精品国产高清国产av| 日韩 欧美 亚洲 中文字幕| 男女视频在线观看网站免费| 婷婷丁香在线五月| 免费看光身美女| 美女大奶头视频| 激情在线观看视频在线高清| 悠悠久久av| 国内毛片毛片毛片毛片毛片| 男女那种视频在线观看| 亚洲国产精品999在线| 亚洲av免费高清在线观看| 欧美乱码精品一区二区三区| 少妇丰满av| 精品欧美国产一区二区三| 人人妻人人看人人澡| 国产三级在线视频| 免费搜索国产男女视频| xxx96com| 日本免费一区二区三区高清不卡| 欧美一区二区国产精品久久精品| 中文字幕精品亚洲无线码一区| 成人永久免费在线观看视频| 村上凉子中文字幕在线| 久久久久九九精品影院| 99精品在免费线老司机午夜| 国产不卡一卡二| 欧美日韩一级在线毛片| 日韩精品中文字幕看吧| 亚洲精品国产精品久久久不卡| 欧洲精品卡2卡3卡4卡5卡区| 久久久久亚洲av毛片大全| 午夜影院日韩av| 精品乱码久久久久久99久播| 操出白浆在线播放| 91九色精品人成在线观看| 国产主播在线观看一区二区| 高清毛片免费观看视频网站| 一级毛片高清免费大全| avwww免费| 国内精品久久久久久久电影| 午夜两性在线视频| 男女之事视频高清在线观看| 欧美最黄视频在线播放免费| 男女做爰动态图高潮gif福利片| x7x7x7水蜜桃| 成年女人看的毛片在线观看| 亚洲aⅴ乱码一区二区在线播放| 少妇的逼水好多| 亚洲av免费在线观看| 成人亚洲精品av一区二区| 人人妻,人人澡人人爽秒播| 好男人在线观看高清免费视频| 婷婷精品国产亚洲av| 亚洲狠狠婷婷综合久久图片| xxx96com| 国产亚洲欧美在线一区二区| 日本熟妇午夜| 久久午夜亚洲精品久久| 亚洲真实伦在线观看| 欧美日韩福利视频一区二区| av中文乱码字幕在线| 熟女少妇亚洲综合色aaa.| 深夜精品福利| 欧美日本视频| 亚洲精华国产精华精| 三级国产精品欧美在线观看| 成年免费大片在线观看| 国产精品久久久久久精品电影| 99热精品在线国产| 中文字幕熟女人妻在线| 变态另类成人亚洲欧美熟女| 国产麻豆成人av免费视频| 少妇人妻一区二区三区视频| 最近最新中文字幕大全电影3| 色综合站精品国产| 婷婷亚洲欧美| 1024手机看黄色片| 亚洲自拍偷在线| 丰满人妻熟妇乱又伦精品不卡| 亚洲熟妇熟女久久| 成人高潮视频无遮挡免费网站| 桃红色精品国产亚洲av| 欧美又色又爽又黄视频| www.熟女人妻精品国产| a级毛片a级免费在线| 成人国产综合亚洲| 精品一区二区三区视频在线 | 99久久九九国产精品国产免费| 亚洲五月婷婷丁香| 午夜免费成人在线视频| 国内久久婷婷六月综合欲色啪| 亚洲精品国产精品久久久不卡| 757午夜福利合集在线观看| 精品欧美国产一区二区三| 成人三级黄色视频| 男女那种视频在线观看| 国产欧美日韩精品亚洲av| 黄色丝袜av网址大全| 18禁黄网站禁片免费观看直播| 欧美一区二区国产精品久久精品| 亚洲精品久久国产高清桃花| 国产亚洲精品一区二区www| 天堂动漫精品| 国内久久婷婷六月综合欲色啪| 搡老岳熟女国产| 国产69精品久久久久777片| 97碰自拍视频| 麻豆成人午夜福利视频| 午夜两性在线视频| 一个人观看的视频www高清免费观看| 成人精品一区二区免费| 老汉色av国产亚洲站长工具| 真实男女啪啪啪动态图| 成人三级黄色视频| 国产亚洲欧美98| 亚洲精品亚洲一区二区| 午夜精品久久久久久毛片777| 亚洲成人免费电影在线观看| 久久久久久久精品吃奶| 人人妻人人看人人澡| 久久久久久久久大av| 国产久久久一区二区三区| 国产三级在线视频| 免费在线观看亚洲国产| 全区人妻精品视频| 欧美在线一区亚洲| 成人高潮视频无遮挡免费网站| 一个人免费在线观看的高清视频| 日本一本二区三区精品| 91麻豆精品激情在线观看国产| 国产精品久久久久久久久免 | 国产97色在线日韩免费| 此物有八面人人有两片| 我要搜黄色片| 69av精品久久久久久| 中文资源天堂在线| 淫秽高清视频在线观看| 国产 一区 欧美 日韩| 国产真人三级小视频在线观看| 最近在线观看免费完整版| 女同久久另类99精品国产91| 18美女黄网站色大片免费观看| 婷婷精品国产亚洲av| 亚洲精品影视一区二区三区av| 免费看日本二区| 日韩人妻高清精品专区| bbb黄色大片| 亚洲电影在线观看av| 嫩草影院精品99| 免费av观看视频| 深爱激情五月婷婷| 亚洲最大成人中文| 色播亚洲综合网| 欧美最新免费一区二区三区 | 少妇人妻一区二区三区视频| 成人鲁丝片一二三区免费| 搡老岳熟女国产| 一卡2卡三卡四卡精品乱码亚洲| 91久久精品电影网| 午夜福利视频1000在线观看| 999久久久精品免费观看国产| 一个人观看的视频www高清免费观看| 午夜日韩欧美国产| 女生性感内裤真人,穿戴方法视频| 亚洲一区二区三区不卡视频| 给我免费播放毛片高清在线观看| 精华霜和精华液先用哪个| 免费一级毛片在线播放高清视频| a在线观看视频网站| 国产真实伦视频高清在线观看 | av黄色大香蕉| 精品免费久久久久久久清纯| 大型黄色视频在线免费观看| 国产 一区 欧美 日韩| 欧美zozozo另类| 又粗又爽又猛毛片免费看| 国产 一区 欧美 日韩| 亚洲aⅴ乱码一区二区在线播放|