• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Survey of Knowledge Graph Construction Using Machine Learning

    2024-01-20 13:00:36ZhigangZhaoXiongLuoMaojianChenandLingMa

    Zhigang Zhao ,Xiong Luo,2,3,★ ,Maojian Chen,2,3 and Ling Ma

    1School of Computer and Communication Engineering,University of Science and Technology Beijing,Beijing,100083,China

    2Shunde Innovation School,University of Science and Technology Beijing,Foshan,528399,China

    3Beijing Key Laboratory of Knowledge Engineering for Materials Science,Beijing,100083,China

    ABSTRACT Knowledge graph(KG)serves as a specialized semantic network that encapsulates intricate relationships among real-world entities within a structured framework.This framework facilitates a transformation in information retrieval,transitioning it from mere string matching to far more sophisticated entity matching.In this transformative process,the advancement of artificial intelligence and intelligent information services is invigorated.Meanwhile,the role of machine learning method in the construction of KG is important,and these techniques have already achieved initial success.This article embarks on a comprehensive journey through the last strides in the field of KG via machine learning.With a profound amalgamation of cutting-edge research in machine learning,this article undertakes a systematical exploration of KG construction methods in three distinct phases: entity learning,ontology learning,and knowledge reasoning.Especially,a meticulous dissection of machine learningdriven algorithms is conducted,spotlighting their contributions to critical facets such as entity extraction,relation extraction,entity linking,and link prediction.Moreover,this article also provides an analysis of the unresolved challenges and emerging trajectories that beckon within the expansive application of machine learning-fueled,large-scale KG construction.

    KEYWORDS Knowledge graph(KG);semantic network;relation extraction;entity linking;knowledge reasoning

    1 Introduction

    The continuous development of information technologies brings significant convenience to human life,paralleled by an exponential surge in information proliferation.Massive amounts of data are collected and studied in many fields such as social network,biomedical engineering,security science,and many others.Under this background,search engine has become an indispensable instrument,facilitating people’s quest for knowledge and information online.Traditionally,a search engine has a user input a query term,whereupon it furnishes hyperlinks directing to the most relevant web pages corresponding to the provided keyword[1].

    In May 2012,the emergence of knowledge graph(KG)brought a novel paradigm for enhancing search engines.Within this framework,user search results transcend the realm of single web page links,encompassing instead a tapestry of structured entity information closely related to the search query.This transformative approach even delves into the realm of potential hidden knowledge within the KG.The intelligent optimization of search answers through KG can effectively improve the functions of future search engines in three aspects: refining responses,nurturing interactive dialogues,and bolstering predictive capabilities[2].This multifaceted augmentation leads into an era of heightened search engine functionality.Furthermore,the scope of KG’s influence extends considerably into domains beyond search,encompassing intelligent question answering,knowledge engineering,data mining,and digital library.

    Recent years have witnessed a remarkable surge of interest from various disciplines engineering and science.As depicted in Fig.1 and supported by data from the Web of Science,the number of published papers with “Knowledge Graph”in their title has exhibited a steady rise up until the end of 2022.Additionally,Figs.2-4 provide insight into the volume of published papers across diverse research areas,publication resources,and institutions.In the field of academic research,Computer Science has emerged as a primary hub for the propagation of KG,closely followed by Mathematics and Engineering.This observation highlights the substantial contribution and enthusiasm originating from the Computer Science towards the exploration,development,and advancement of KG-related topics.Notably,the significant research output in Mathematics and Engineering emphasizes the interdisciplinary essence of KGs,signifying their influence across domain that extend beyond the realm of computer.Shifting the focus to the publication platforms,it is evident thatLectureNotes inComputerSciencehas emerged as the leading platform for KG-centric research,closely followed byLectureNotesinArtificialIntelligenceandIEEEAccess.This indicates thatLectureNotesinComputer Sciencehas been the preferred choice for researchers to share their findings and advancements in the field of KGs.Lastly,the Chinese Academy of Sciences,University of Chinese Academy of Science,and Rluk Research Libraries UK have emerged as the leading contributors in terms of publishing papers related to KG.These institutions have demonstrated a robust presence and active participation in KG research,highlighting their expertise and dedicated contributions to the advancement of this field.Their significant published works attest to the valuable role played by these institutions in the exploration and evolution of KG-related topics.

    Figure 1:Number of published papers with“Knowledge Graph”in the title in recent years

    Figure 2:The top 10 research areas ranked by the number of publication related to KG

    Figure 3:The top 10 publication resources ranked by the number of publication in KG

    Figure 4:The top 10 institutions ranked by the number of published papers related to KG

    Generally speaking,the establishment and utilization of large-scale KG necessitate the synergy of diverse intelligent information processing technologies.In recent years,machine learning methods have as a pivotal force within the KG construction.Therefore,this article conducts a survey in relation to this field.We first introduce the development history,fundamental principles and technical framework of KG.Second,we summarize the machine learning-based key technologies integral to the actualization,dissected across three pivotal dimensions.Finally,we expound upon the current challenges and future forthcoming trends poised to guide the construction of large-scale KG.The main contributions are as follows:

    · This article offers an extensive and up-to-date review of existing research and literature in the field of KG construction,specifically focusing on methodologies driven by machine learning techniques.Meanwhile,it provides a structured categorization and classification of diverse machine learning-driven approaches utilized for constructing KGs.This categorization could help readers understand the landscape and taxonomy of methods in this field.

    · This article delves into the various machine learning methodologies employed in the construction of KGs.Meanwhile,we conduct the comparative evaluation and analysis of different machine learning techniques,showcasing their respective performance,scalability,and suitability under various conditions.

    · This article identifies and discusses challenges and open research questions within the domain of KG construction using machine learning,highlighting potential avenues for further exploration and innovation.

    The organization of this article is arranged as follows.In Section 2,we present some fundamental concepts and traditional technical architecture of KG.Section 3 focuses on a comprehensive exploration of KG design propelled by the prowess of machine learning methods,meticulously partitioned into three parts,i.e.,entity learning,ontology learning,and knowledge reasoning.In Section 4,we discuss the prospective research directions and challenges of large-scale KG construction technologies.Through machine learning methods,we dissect this discussion into three distinct segments,relation extraction,link prediction,and construction of industrial KG.Section 5 provides a conclusion and reflections.

    2 Knowledge Graph

    In this section,we simply introduce the essence and the foundational framework of KG.

    2.1 The Development of Knowledge Graph

    With the development of the Internet,Web technology has gone through the “Web 1.0” era characterized by the web of documents and the “Web 2.0” era characterized by the web of data.Today,the trajectory points towards the“Web 3.0”era characterized by the web of knowledge[3]and even anticipates the“Web 4.0”era defined by the Metaverse paradigm[4].Driven by the continuous growth of user-generated content and open-linked data on the Internet,the quest for knowledge interconnected aligning with the ever-evolving network information resources becomes imperative.This quest takes a fresh perspective in accordance with the principles of knowledge organization in the big data environment,aiming to reveal deeper cognitive insights [5].In the midst of this dynamic context,Google introduced KG in May 2012.Its goal is to enhance search outcomes,describe the various entities and concepts inherent to the real world,and illuminate their relationships.By these merits,KG emerges as a substantial stride forward from prevailing semantic web technologies.Illustrated in Fig.5 are pivotal milestones making the history of KG across different years.For instance,conception of the semantic network as a vehicle for knowledge representation was proposed in 1960.Furthermore,the philosophical concept of“ontology”was integrated into the KG in 1980,facilitating the structured and formalized description of knowledge.

    Figure 5:Milestones in the development of KG

    The origin of designing KG comes from a series of practical applications,spanning fields such as semantic search,machine question answering,information retrieval,online learning and others.With the exploration of KG advances,various structured KGs have been developed by both academic researchers and industry practitioners.Currently,a tableau of prominent and expansive large-scale open knowledge bases associated with KG exists globally,as enumerated in Table 1.Here,large-scale knowledge bases like Freebase [6],DBPedia [7],and Wikidata [8] take center stage using Wikipedia as a foundational source.Notably,Freebase differentiates itself by its user-generated content,open accessibility,and structured data,which supports all its entries.

    Table 1: The sizes of some prominent large-scale knowledge bases

    In recent years,an array of research results on the Chinese KG has grown vigorously.For example,Sogou established“Knowledge Cube”,marking the inception of the foremost knowledge base search product in the domestic search engine industry.Through effectively integrating fragmented Internet knowledge,Baidu founded“Baidu Zhixin”and brought forth a next-generation search engine product.Contributions extend to academia as well,Tsinghua University built“XLore”,a pioneering large-scale Chinese-English cross-language KG.The Institute of Computing Technology of Chinese Academy of Sciences established a prototype system termed “People Cube,Work Cube,Knowledge Cube”based on an open knowledge network OpenKN.Shanghai Jiao Tong University designed‘Zhishi.me’,a dedicated research platform for Chinese KG.Additionally,the GDM Lab at Fudan University launched the Chinese KG project.Generally,these products and projects have given rise to expansive knowledge bases spanning diverse fields,providing users with intelligent search and question-andanswer services.

    2.2 The Definition of Knowledge Graph

    KG is a special semantic network composed of nodes and directed edges,and it is also known as a heterogeneous information network or semantic knowledge base.In the KG,each node represents an entity in the real world,while directed edges interlink these nodes to denote the intricate relationships between those entities.Facts are generally represented in the form of triples(subject,predicate,object)(SPO),where subject and object signify entities,and predicate represents the relation between them[10].For example,the textual data “Chao Deng is an actor who played the character Tailang Xu in the comedy movie Duckweed”can be expressed via the following set of SPO triples exemplified in Table 2.The transformed version of Table 2 into a KG is described in Fig.6.This KG encapsulates the interrelationships between various entities,allowing for a structured representation of the given information.

    Table 2: An example of SPO triples extracted from text data

    Figure 6:The updated version of Table 2 after translating it into a KG

    2.3 The Technical Architecture of Knowledge Graph

    The structure of KG includes two aspects:a logical structure and a technical structure.The former includes the data layer and the pattern layer,while the latter refers to the technological process involved in KG construction.This involves a sequence of stages,including data acquisition,entity learning,ontology learning,knowledge reasoning,and knowledge update.

    The data layer,alternatively known as the entity layer,functions as a repository for knowledge housing information in the form of facts.These facts are succinctly conveyed through triples as the basic expression of facts,where a graph database is chosen as a storage medium.

    Above the data layer,the pattern layer,commonly referred to as the ontology layer,represents and stores refined concepts and knowledge.The pattern layer leverages ontology constructs,effectively serving as an embodiment of the KG.It plays an important role in defining and organizing entities,properties,classes,and relationships within a KG.Typically,ontology is represented using formal languages like the Web Ontology Language(OWL)[11],which is grounded in description logic.OWL’s expressive capabilities empower ontologies to accurately define semantic relationships among entities,properties,classes,and relationships.Moreover,ontology-based reasoning facilitates the identification of missing information in the KG,the discovery of hidden relationships between entities,and the ability to address intricate semantic queries.Furthermore,ontology enables seamless interaction and knowledge sharing among different KGs.This fosters the construction of larger,more comprehensive KG,and promotes the reuse of knowledge,thereby augmenting the overall value of knowledge graph.Generally,the axioms,rules,and constraints in the ontology base are used to standardize the entities,the types and the attributes of entities,and the relationship between the entities,so that the KG has a strong structure and less redundancy[12].

    The technical architecture of KG construction can be classified into two principal paradigms:top-down and bottom-up.In the former,the pattern layer is first defined,and the construction begins from the top-level concept.It then proceeds to progressively refine,layer,and generate instances downward.The process is depicted in Fig.7.Conversely,the latter starts from the underlying entities,extracts entities,and gradually abstracts them upwards to form upper-level concepts and knowledge.This architectural construct is illustrated in Fig.8.It starts from the original semi-structured data or unstructured data and adopts a series of technologies to extract knowledge.Then,it integrates with the structural data,with the ontology layer contributing to the enrichment of upper-level concepts.Finally,a complete KG is generated.Furthermore,the comprehensive KG is continuously updated and augmented.New knowledge is continuously extracted to promote the KG refinement.Within this framework,each iteration generally includes data acquisition,entity learning,ontology learning,knowledge reasoning,and knowledge updating[13].

    Figure 7:The technical architecture of the top-down KG

    Figure 8:The technical architecture of bottom-up KG

    3 Knowledge Graph Construction Using Machine Learning

    A typical machine learning approach is usually operated on a structured data matrix,where each row in the matrix corresponds to an object characterized by an attribute eigenvector.The main task of machine learning is to achieve the mapping from these eigenvectors to various forms of output through learning.Additionally,unsupervised learning can facilitate clustering and factor analysis.Then,according to the bottom-up KG construction process described in Fig.8,we summarize the realization of KG driven by machine learning from the following three parts,each bearing its unique set of implementation techniques and challenges.

    3.1 Entity Learning

    Entity learning refers to the intricate construction process of entity layer within the KG.From bottom to top,it includes three modules:entity extraction,relationship extraction,and entity linking.

    3.1.1EntityExtraction

    Entity extraction stands as the initial and pivotal step in the knowledge extraction,involving the automatic identification of named entities from the original corpus.This foundational process relies on the automatic detection and categorization of named entities within a given corpus.Fig.9 illustrates various categories to which these named entities can be ascribed,such as Person,Country,City,and more.This process enables the identification of important entities,laying the foundation for subsequent knowledge extraction and analysis tasks.

    Generally,there are three typical methods in this field,and they are the rule-based,the traditional machine learning-based,and deep learning-based extraction methods.Fig.10 displays the classification results obtained from these methods,showcasing the effectiveness and performance of each strategy in identifying and categorizing named entities.Here,rule extraction is an early pattern implemented by manually designing rules primarily toward proper nouns within a specific domain’s text.It is based on painstakingly handcrafted patterns,necessitating a lot of human efforts,culminating in limited extraction capacity and constrained scalability.Meanwhile,the related extraction methods based on machine learning models have witnessed remarkable progress.These methods integrate machine learning algorithms into entity extraction to achieve automatic or semi-automatic entity identification.Finally,with the development of artificial neural networks,some deep learning-based methods have been proposed to attain heightened proficiency in entity extraction tasks with reduced human intervention.This progression signifies a remarkable leap forward in the field.

    Figure 9:Examples of different types of entities

    Figure 10:Classification results of entity extraction methods

    For the traditional machine learning-based entity extraction,the inception involved the utilization of supervised machine learning algorithms.For example,the decision trees and the conditional random field (CRF) model [14] were employed to realize entity recognition for Telugu-English code-mixed social media data.In addition,Sykes et al.[15]conducted a comprehensive comparison between rulebased and machine learning-based extraction methods,effectively emphasizing the latter’s superior adaptability and versatility.

    Furthermore,with the progressive evolution of web technologies,the combination of openlinked data with machine learning algorithms frequently yields more compelling results.Within this framework,the basic idea revolves around employing machine learning to extract entities with similar contextual features from the web page,subsequently achieving entity classification or clustering.For example,Whitelaw et al.[16] proposed an iterative approach for expanding the entity corpus in a network environment.This approach depended on the construction of feature models grounded in known entities,enabling the processing of massive datasets.With this mechanism,they effectively modeled new entities to achieve continuous iterative expansion of the entity.Furthermore,employing unsupervised learning algorithms,Jain and Pennacchiotti[17]successfully extracted newly emerging named entities from the server logs of search engines.Then,this method found practical application in search engine technology,allowing for automatic information completion based on user-input keywords.Additionally,while constructing KG,an application of intelligent corpus annotation for entity extraction was presented[18].

    Deep learning-based methods offer the advantage of automated text feature selection for entity extraction,which reduces incompleteness and manual work.They have shown promising results in this area.One such method employs convolutional neural networks(CNN)to automatically learn features from the input text data.For example,a deep learning-based method[19-21]was proposed for entity extraction that utilized a CNN to learn contextual features from the input text data.Those models were trained on a large dataset of annotated text datasets and exhibited satisfactory performance on several benchmark datasets.Similarly,Cho et al.[22] presented a deep learning-based strategy for entity recognition in biomedical texts,combining a CNN with a long short-term memory (LSTM)network.Their model acquired complex features from the input text data,attaining remarkable accuracy on a challenging biomedical entity recognition task.Overall,deep learning-based methods have the potential to advance entity extraction,thereby constructing more precise and comprehensive KG.Hence,this enhancement leads to more powerful applications in many fields,such as natural language processing (NLP) and information retrieval.Notably,Table 3 showcases the top 10 entity extraction models from 2019 to 2022,ranked by F1-score on various open-source datasets(CoNLL 2003[23],ACE 2005[24],and Ontonotes 2005[25])This information is sourced from the Papers with Code website(https://paperswithcode.com).

    Table 3: The top 10 entity extraction models from 2019 to 2022

    Notes:For more detailed and up-to-date information about the models than what is presented in the article,readers can refer to the following links: https://paperswithcode.com/sota/named-entity-recognition-ner-on-conll-2003;https://paperswithcode.com/sota/named-entityrecognition-on-ace-2005;https://paperswithcode.com/sota/named-entity-recognition-ner-on-ontonotes-v5.

    For the CoNLL 2003 dataset,several remarkable models have garnered high F1-score,highlighting their prowess in entity extraction.The Automated Concatenation of Embeddings (ACE)combined with document-context [26] stands out at an impressive 94.60%.Close on its heels,the Co-regularized Language Understanding with Knowledge-based Embeddings(LUKE)[27]achieves a commendable 94.22%,while the Autoregressive Structured Prediction (ASP) fused with Text-to-Text Transfer Transformer(T5)-3B[28]achieves a noteworthy 94.10%(the“3B”represents 3 billion parameters).These models have demonstrated outstanding performance in extracting entities from the CoNLL 2003 dataset.

    In the ACE 2005 dataset,the Princeton University Relation Extraction system (PURE) [35]emerges triumphant with an F1-score of 90.90%.Not far behind,PromptNER [36] has a second position with 88.26%,followed by Parallel Instance Query Network(PIQN)[37]with 87.42%.These models have shown strong performance in entity extraction from the ACE 2005 dataset.

    For the Ontonotes 2005 dataset,the Bidirectional Encoder Representations from Transformers(BERT)-Machine Reading Comprehension (MRC)+dice coefficient (DSC) [44] attains the pinnacle with the highest F1-score of 92.07%.Packed Levitated(PL)-Marker[30]follows closely behind with 91.90%,closely trailed by Baseline+Boundary Smoothing(BS)[45]achieving 91.74%.These models have demonstrated their efficacy in entity extraction from the Ontonotes 2005 dataset.

    These state-of-the-art models serve as prime examples of the evolutions made in entity extraction techniques,leveraging various approaches including deep learning,Prompt,and co-regularization methods.With remarkable F1-score on their respective datasets,these models indicate their effectiveness in extracting entities from text.

    3.1.2RelationExtraction

    Relation extraction is an important subtask of knowledge extraction.In the past,this task entailed manual rule construction,followed by pattern-matching techniques to extract corresponding relation instances from text.However,the advent of machine learning has revolutionized relation extraction methods.It leverages lexical and syntactic attributes for model training,effectively transmuting relation extraction challenges into classification or clustering problems.According to the extent of human involvement and dependence on labeled corpus,machine learning-based relation extraction approaches can be divided into the supervised learning-based relation extraction,the semi-supervised learning-based relation extraction,and the unsupervised learning-based relation extraction.

    (1)Relation Extraction with Supervised Learning Method

    Supervised learning-based relation extraction is an automatic mode on the basis of meticulously labeled training data.Through ongoing learning from these training samples,the classification and predictions are performed on datasets.Within this paradigm,binary relation extraction is treated as a classification problem.As shown in the following definition,the triple(ei,rk,ej)indicates that there is a semantic relationshiprkbetween the head entityeiand the tail entityej,and functionf(·)represents the relationship classifier employed in the context:

    f(ei,rk,ej)=

    Supervised learning employs two primary categories of relationship classification methods:feature vector-based and kernel function-based.In the first category,features are extracted from the training samples and represented as sequence feature vectors,using the results of part-of-speech tagging and syntax parsing.Here,the prominent methods are support vector machine(SVM)[49]and maximum entropy(ME)[50].For example,a classifier system employing SVM integrated lexical features from polarity lexicons and lists of offensive/profane words to identify and classify offensive language in social media[49].The second category of methods effectively avoids the challenges of dimensionality caused by nonlinear transformations.Recent years have witnessed widespread utilization of kernel function-based approaches in many fields[51].In relation extraction using kernel functions,such as convolution kernel[52],tree kernel[53],subsequence kernel[54]and some improved kernel methods[55]play a pivotal role.The key of these methods involves projecting the implicit feature vector of a sentence into the feature space using the kernel function and calculating the inner product between these projections,so as to assess the similarity of the relationship between entities.For example,an innovative tree kernel,termed feature-enriched tree kernel(FTK)was proposed[53],while achieving a 5.4%enhancement in F-measure over the traditional convolution tree kernel.

    Supervised learning-based relation extraction methods yield excellent experimental results,but their effectiveness heavily depends on the classification features provided by part-of-speech tagging and syntactic parsing.To address this problem,the use of supervised relation extraction has witnessed a surge of interest driven by the deep learning model[56].A recurrent neural network(RNN)-based relation extraction model was proposed by Socher et al.[57].This approach involved vectorizing each node of the syntactic tree through syntactic analysis.Guided by the syntactic structure,it iterated continuously from the lowest word vector of the tree.Finally,the vector representation of the sentence was attained and employed as the foundation for relationship classification.This method effectively utilized syntactic structure information,but overlooked the position information of words.Then,convolutional neural network (CNN) took center stage.Here,the word vector was treated as an initialization parameter,engaging in convolution training with dynamic optimization during the learning process,culminating in classification[58].To complement the local dependencies captured by piecewise CNNs,a self-attention mechanism was proposed to capture rich contextual dependencies[59].The experiments were performed on the NYT dataset and the experimental results demonstrated that the model provided a new benchmark in the area under curve(AUC)metric.Expanding on the CNN-based methodology presented[57],a refined iteration emerged[60].This advancement involved inputting both word vectors and word positional vectors,with the sentence representation being obtained through the learning of convolutional,pooling,and nonlinear layers.This method fully considered the entity location information and other related lexical features,which achieved good relation extraction results.On the standard SemEval-2018 Task 7 dataset,the CNN method achieved superior performance when compared to alternative relation extraction methods [61].Furthermore,except for deep learning models,deep reinforcement learning has also been used in the relation extraction [62].This approach casts the relation extraction as a two-step decision-making game,employing the Q-Learning algorithm with value function approximation to learn control policy.The experiments were conducted on the ACE 2005 corpus,and they showed that the deep reinforcement learning model achieved a state-of-the-art performance in relation extraction tasks.

    (2)Relation Extraction with Semi-supervised Learning Method

    Semi-supervised learning-based relation extraction aims to realize the binary relation classification with limited training samples,thus circumventing the constraints imposed by manual annotation of extensive training data.It mainly adopts the bootstrapping method and some other methods.

    The idea of bootstrapping is to artificially construct a small set of initial relation instances as a seed.This seed set serves as the foundation for model training,and through iterative expansion,it gradually augments to encompass a more extensive collection of relation instances,and finally completes the relation extraction task.Here,the entity alignment technique was improved to reduce the data noise [63].However,this method operates under the assumption that a single entity pair corresponds to just one relationship.To address this limitation,a multi-instance multi-label(MIML)method was proposed to model the relationship extraction.This methodology describes the situation where an entity pair may have multiple relationships [64].Moreover,the integration of a Bayesian network with MIML was explored for relation extraction[65],further expanding and capabilities of the approach.

    Although the bootstrapping method is intuitive and effective,it may introduce a large number of noisy instances during seed expansion,leading to semantic drift.To address it,a deep co-learning was proposed [66],and it was a semi-supervised end-to-end deep learning method for evaluating the credibility of Arabic Blogs.A coupled semi-supervised learning method was used to establish constraints between different categories of extraction templates [67].This strategic implementation effectively curbed the generation of false templates,thereby bolstering the precision of relation extraction.Meanwhile,a method was proposed by combining matrix-vector recursive neural network(MV-RNN) with bootstrapping [68].Here,through the tree structure in MV-RNN,the semantic information of the entire sentence could be extracted as relation classifier features,and it greatly improved the accuracy of the results avoiding the problem of requiring a large amount of corpus in MV-RNN.More recently,the integration of transfer learning,a popular machine learning strategy,into the semi-supervised learning,yielded a novel framework [69].Applied within the context of low-resource entity and relation extraction in the scientific domain,this framework demonstrated satisfactory performance,underscoring its potential and versatility.

    (3)Relation Extraction with Unsupervised Learning Method

    Unsupervised learning-based relation extraction assumes that pairs of entities with the same semantic relation have similar contexts,and it transforms the relation extraction task into a clustering problem.Hence,it does not require manual corpus annotation,but the accuracy rate is relatively low.

    Generally,it is implemented using various clustering algorithms.For example,the large pretrained language model was used for adaptive clustering on contextualized relational features to improve computational performance in relation classification [70].After taking the entity set in the Wikipedia entry as the object,and using the dependency features and shallow grammar templates,all semantic relationship instances corresponding to entities were extracted in a large-scale corpus by pattern clustering [71].Meanwhile,the templates were extracted and aggregated from search engine summaries,and they were clustered to discover implicit semantic relationships represented by entity pairs[72].To further elevate the efficacy of relational templates’clustering,a co-clustering algorithm was used,leveraging the dual nature of the dual of relational instances and relational templates.Moreover,the integration of a logistic regression model played a pivotal role in filtering representative extraction templates from the clustering results of relational templates[73].

    For the above three basic types of machine learning,Table 4 offers a comprehensive comparison and analysis of the algorithms used in relation extraction.The table highlights the distinctions between classical algorithms,extraction ideas,human intervention levels,and extraction performance across different methods.Specifically,supervised and semi-supervised methods extract relationships by classifying learning sample data labels,while unsupervised-based methods cluster data and group related entities together to achieve relationship extraction.A noteworthy observation is that the performance of the extraction model improves with an increased corpus size and manual intervention.Taking into account the trade-off between performance and computational efforts,the semi-supervised learning methods emerge as favorable choices for relation extraction.It not only ensures better extraction performance but also avoids the limitation of requiring an extensive amount of manually annotated corpus.Moreover,Table 5 provides a comprehensive overview of the performance exhibited by various excellent relation extraction models on two famous datasets(NYT[74]and WebNLG[75])in recent years.The F1-score achieved by these models indicate their effectiveness in relation extraction.

    Table 4: Comparison of three kinds of machine learning-based relation extraction methods

    Table 5: The top 10 relation extraction models from 2020 to 2022

    Notes:For more detailed and up-to-date information about the models than what is presented in the article,readers can refer to the following links: https://paperswithcode.com/sota/relation-extraction-on-nyt;https://paperswithcode.com/sota/relationextraction-on-webnlg.

    Focusing on the NYT dataset,UniRel[76]achieved the highest F1-score of 93.7%,followed by Relation Extraction By End-to-end Language generation(REBEL)[77]with 93.4%,and Djacency lIst oRiented rElational faCT(DIRECT)[78]at 92.5%.These models have exhibited robust competence in extracting relations from the NYT dataset.

    For the WebNLG dataset,UniRel[78]maintains its lead,achieving the highest F1-score of 94.7%,followed by Partition Filter Network (PFN) [79] with 93.6%,and Set Prediction Networks (SPN)[80] with 93.4%.These models have exhibited remarkable performance in relation extraction from the WebNLG dataset.Furthermore,other models such as Translating Decoding Schema for Joint Extraction of Entities and Relations (TDEER) [81] and Representation Iterative Fusion based on Heterogeneous Graph Neural Network for Joint Entity and Relation Extraction (RIFRE) [82] also achieved comparable performance on both datasets,further highlighting the consistent excellence of these methodologies.

    These excellent models utilize various techniques,including RNNs,partition-based methods,and joint learning frameworks.The impressive F1-score attained by these models underscores their ability to effectively extract relations from text across different datasets.This analysis showcases the advancements in relation extraction models and their potential applications in numerous NLP tasks.

    3.1.3EntityLinking

    Due to the diversity of information expression,entity ambiguity remains a frequent and formidable hurdle in natural language understanding.For example,“apple”,can signify either a fruit or the renowned technology company.However,when contextual information such as “Steve Jobs”is provided,it becomes evident that in the given text,the entity“apple”refers to the company,as illustrated in Fig.11.The presence of contextual cues helps disambiguate the intended meaning of entities,aiding in enhancing the precision and comprehension of text.

    Figure 11:An example of entity linking

    Entity linking is an effective disambiguation method.By linking entity mentions to the corresponding entities in the knowledge bases,this method conducts precise entity annotation within documents.This enables computers to attain a more profound grasp of the semantic information of the text,effectively addressing the problems of synonym and polysemy.Commonly,entity linking leverages well-established knowledge bases such as Wikipedia,DBPedia,and Freebase.Specifically,open knowledge bases like Baidu Encyclopedia and Interactive Encyclopedia also find prominent applications in Chinese entity linking.Generally speaking,entity linking includes two subtasks:entity recognition and entity disambiguation.By skillfully addressing both aspects,this technique contributes to a more comprehensive and nuanced understanding of textual content.

    (1)Entity Recognition

    Entity recognition aims to identify fragments of text,and it may link to specific entries in the knowledge bases,including a specific word or phrase.Typical entity types include place names,person names,institution names,times,dates,percentages,and amounts.With the development of Internet information,novel entity categories have surfaced in recent years,including movie titles and product names,reflecting the expanding landscape of entity recognition[89,90].

    Currently,entity recognition technology is mainly based on statistical machine learning methods,treating the task as a sequence labeling problem.There are three main solutions: hidden Markov model (HMM),maximum entropy Markov model (MEMM),and CRF model.An overview of these models,their key evaluation criteria,and comparisons are delineated in Table 6.Over recent years,numerous studies have utilized these three models.For instance,a novel generative model was proposed[91],linking it to HMM while proving its generically identifiable nature without any observed training labels.However,HMM exhibits the tag bias problem by assuming that the current tag solely depends on the previous one.Additionally,its use of local normalization to compute the probability of observation series makes calculating the function complex,leading to reduced efficiency.MEMM addresses the tag bias problem but introduces the tag inconsistency issue.Furthermore,it often relies on manually designed features,which are pivotal for the model’s performance and generalization ability,making feature selection a challenging task.CRF model is different from HMM and MEMM.By employing global modeling,CRF simultaneously considers the observation and labeling of the entire sequence when calculating the probability.This enables the avoidance of label bias and local normalization problems found in HMM and MEMM.Moreover,the CRF model resolves the label inconsistency problem present in MEMM by modeling dependencies between observations and label sequences concurrently.Additionally,CRF utilizes complex and rich feature representations that capture dependencies between observed and labeled sequences,thereby enhancing model performance and generalization.Unlike MEMM,CRF model is less dependent on manually selected features and can learn feature weights that suit the task,thereby streamlining feature engineering.By overcoming some of the main shortcomings of HMM and MEMM models,CRF achieves superior performance in sequence labeling and has been widely used in the entity recognition field[92].

    Table 6: Comparative analysis of three named entity recognition methods

    (2)Entity Disambiguation

    In the context of a specific entity mentioned in the text,entity disambiguation is mainly used to analyze the semantic information and select the corresponding entity from the candidate entities.This process depends on the candidate entity and the contextual information.Usually,given the ambiguity in natural language,there are numerous candidate entities vying for consideration.Methods of entity disambiguation are mainly based on supervised learning and unsupervised learning.

    Most works adopt supervised learning methods for disambiguation and use training data to automatically design ranking models.The requisite training data for entity linking comprises an ordered list of all candidate entities associated with the target mentioned in a given context.In this list,the first entity is usually the one that the mention refers to in this context.By linearly combining features like entity popularity,semantic similarity,and connection between entities,a maximum margin-based data were employed to train feature weights,while achieving entity disambiguation through a ranking model[93].Moreover,two machine learning sorting-based methods using listwise and pairwise were presented to implement entity disambiguation,outperforming traditional disambiguation methods[94].The listwise method is from the LiNet algorithm,which uses the ordered list as a training instance to obtain a sorting model.The essence of the pairwise method is to transform the sorting problem into a classification problem.It combines the items in the ordered list into pairs and constructs training instances according to the relative positional relationship between the items to develop a sorting perception.Furthermore,recent developments have seen the combination of supervised learning with graph theory to address the entity disambiguation task [95],showcasing the innovative fusion of established techniques to enhance disambiguation accuracy.

    On the other side,unsupervised learning algorithms have been used in entity disambiguation.For example,a clustering-based personal name disambiguation system was proposed to extract personal attributes and social relations between entities from text,subsequently mapping them onto an undirected weighted graph [96].Clustering algorithms were then used to cluster these graphs,each cluster contained all web pages that directed a person.Significant models have been shown in Table 7 in the field of entity disambiguation,where datasets are ACE2004 [97] and AIDA-CoNLL[98].These models focus on disambiguating entities,which is crucial for accurately identifying the intended meaning of ambiguous terms.

    For the ACE 2004 dataset,Entity Disambiguation by Reasoning over a Knowledge Base(KBED)[99] stands out with the highest F1-score of 93.4%,followed by LUKE[confidence-order] [100] with 91.9%,and Representation and Fine-grained typing for Entity Disambiguation (ReFinED) [101] at 91.6%.These models have shown strong performance in entity disambiguation on the ACE 2004 dataset.

    Table 7: The top 5 entity disambiguation models

    Regarding the AIDA-CoNLL dataset,LUKE[confidence-order][100] maintains its prominence with the highest F1-score of 95.0%,closely followed by Dynamic Context Augmentation (DCA)-Supervised Learning (SL)+Triples [104] and DeepType [105] with F1-score of 94.9%.These models have demonstrated excellent performance in disambiguating entities in the AIDA-CoNLL dataset.It is worth mentioning that LUKE[confidence-order] [100] is the only model present in both datasets,indicating its robustness and effectiveness across different evaluation scenarios.

    The top-performing models employ diverse techniques such as knowledge-based methods,confidence ordering,and deep learning approaches.Their high F1-score emphasize their proficiency in accurately disambiguating entities and determining their intended meanings in different contexts.

    3.2 Ontology Learning

    An ontology is a formalized specification of the shared conceptual model,and it defines the pattern layer of KG.The composition of an ontology asO,entails the components(C,root,R).Here,Cis the set of upper-level concepts,therootis the root identifier,andRis the binary relationship onC,including the synonymy relation and the hyponymy relation as Fig.12 [108].The purpose of an ontology lies in establishing an organized framework,thereby facilitating the organization and categorization of concepts within the KG,enabling efficient retrieval and navigation of knowledge stored within.Generally speaking,ontology construction has three ways: manual construction,automatic construction,and semi-automatic construction.Manual construction method necessitates the participation of domain experts and entails the utilization of dedicated ontology editing tools.However,this approach tends to demand significant human and material resources,leading to scalability issues.Hence,it cannot keep up with the rapid development and update of Internet data.Ontology automatic construction or ontology learning includes extraction of concept,synonymy relation,and hyponymy relation from bottom to top.This process is predominantly automated,often relying on data-driven or cross-language knowledge linkage techniques that are grounded in machine learning principles.For instance,in order to address the ontology automation issues in the semantic web,ontology learning was achieved through the presentation of automatic or semi-automatic,aimed at either generating new ontology resources or repurposing existing ones[109].

    Figure 12:The flowchart of ontology modeling

    Concept is the most basic unit of human comprehension.The common methods of concept extraction include linguistic methods[110],statistical methods[111],and machine learning methods.Within the domain of machine learning-based concept extraction,prevalent methods center around SVM[49],HMM[91],bootstrapping[112],and clustering.These techniques entail the extraction of pertinent categorical attributes from the dataset.For example,bootstrapping was used to automatically extract domain vocabulary from large-scale,unlabeled real corpus[112].A bootstrapping-based seed expansion mechanism was developed to realize the automatic extraction of domain seed words[113].Using the concept of clusters,the multi-topic extraction algorithm was introduced to acquire multiple topics by clustering concepts[114].

    Relationship extraction between concepts mainly refers to the synonymy and the hyponymy relation.Synonymy relation extraction is to examine the degree of probability that any two entities belong to the same conceptual level.For example,entities like“Beijing”and“Shanghai”,both denoting city names,exhibit a synonymy relation.Hyponymy relation extraction gauges the probability that any two entities establish a hierarchical relation,where one serves as a subtype of the other.For example,“Beijing” is the hyponymy of “city”.For the supervised learning method,a novel approach was proposed for synonym identification using the principle of distributional similarity[115].Compared to the traditional similarity models,the experimental results showed that a satisfactory performance was achieved while increasing by over 120%on theF1metric.A syntactic parser was used to construct a syntax tree,and the contextual features of concepts were used as concept attributes to generate concept lattices[116].This led to the establishment of a partial order relationship of concept lattices,which subsequently formed the conceptual hierarchy of ontology.For the unsupervised learning method,a method was proposed to learn taxonomy from a collection of text documents,each dedicated to describing a distinct concept [117].Specifically,with the continuous development of online encyclopedias,the machine learning method using the linked data has gradually become an efficient strategy for hyponymy relation extraction.Machine learning techniques were used to explicitly represent the meaning of any text as a weighted vector of Wikipedia-based concepts.The cosine of the angle between these vectors was then calculated to measure the similarity between concepts or texts,effectively fostering the extraction of hyponymy relations[118].

    3.3 Knowledge Reasoning

    Building upon the foundation of existing entities and relationships,knowledge reasoning amis to mine implicit connections between entities through a sophisticated reasoning mechanism.The ultimate goal is to enhance and amplify the original KG by unveiling implicit relationships.For instance,as illustrated in Fig.13,if the KG contains the information “Lion is-a animal” and “Animal can run”,it becomes possible to infer the knowledge that “Lion can run” through logical reasoning.This process stands as a testament to the potential of knowledge reasoning,allowing the enrichment and augmentation of the KG through the generation of novel insights from preexisting information.This iterative process culminates in heightened completeness and a more profound grasp of semantic understanding.

    Figure 13:An example of knowledge reasoning

    Traditional knowledge reasoning is mainly based on logical reasoning methods,including predicate logic,description logic,rule-based reasoning,and others.The predicate logic method is generally designed for simple entity relations.This method takes propositions as the fundamental units of reasoning,where atomic propositions are generally decomposed into two parts:individual words and predicates.Description logic can be used for complex relationships between entities.The typical path ranking algorithm plays a pivotal role in establishing rule-based reasoning.Distinctive relationship paths were used as one-dimensional features and the classification feature vector of the relationship was constructed by counting a large number of relationship paths in the graph [119].However,it is important to acknowledge that logical reasoning necessitates the formulation of rules,a task that often proves computationally onerous and encounters challenges posed by data sparsity.

    Link prediction is a new type of knowledge reasoning method under the statistical machine learning framework.It is to predict the possibility of the linked relationship between two unlinked nodes through the known nodes and link information in the KG,while discovering the implicit relationship between entities.The link prediction in the KG is generally realized by the representation learning method on the basis of the triples (ei,rk,ej) used to constitute the knowledge.For the representation learning method,the semantic information of an entity is represented as dense lowdimensional real-valued vectors.Within this space,gauging the semantic similarity between objects is facilitated by mathematical methods such as cosine distance and Euclidean distance[120].

    Recently,a large number of works on the representation learning-based link prediction have been proposed.One illustrative instance is the structured embedding (SE) method.This method projects the entity vectorseiandejthrough the two relation matrices of the inter-entity relationrkto the corresponding space ofrk[121].Then,the distance between two projection vectors on this space was calculated to gauge the confidence ofrk.The single layer model (SLM) was proposed through the application of a single-layer neural network[122].This network employs nonlinear operation to define a scoring function for each triple.fostering the synergistic representation of the semantic connection between entities and relationships to reason about unknown relationships.While the accuracy of results exhibited significant enhancement over traditional methods,the adoption of nonlinear operations inevitably led to heightened computational complexity[122].Another innovative contribution is the semantic matching energy(SME)model,which relies on low-dimensional vectors to represent entities and relationships[123].Multiple projection matrices are employed to represent the connections between entities and relationships.The latent factor model(LFM),delves into a relationship-centered bilinear transformation.This transformation encapsulates the semantic connection between entities and relationships [124].The RESACL model was presented as a typical knowledge representation method through matrix decomposition [125].In this study,all triples (ei,rk,ej) were represented as a large tensor,wherein the presence or absence of a triple dictates the value at the corresponding tensor position.Through the tensor decomposition algorithm,the tensor value corresponding to each triple in the tensor could be decomposed into entity and relation representations.Based on the characteristic of translation invariance of the word vector space,TransE model presents a pioneering perspective [126].As depicted in Fig.14,in this model,the semantic relationrkbetween entitieseiandejwas regarded as some kind of translation vector,and it actually was the translation from the head entityeito the tail entityej.Demonstrating substantial improvements in establishing intricate semantic connections within vast,sparse KGs,TransE has solidified its status as a pivotal model in this domain.Recent developments have exhibited a growing interest in the TransE model,exemplified by the increased attention it has garnered in studies[127,128].

    Figure 14:A brief description of the TransE model

    4 The Future Challenges of Constructing a Large-Scale Knowledge Graph

    Amidst the era of big data,the KG provides a new learning paradigm to efficiently organize,manage and understand massive amounts of information,while presenting this data in a manner closely aligned with human cognition.Hence,it promotes rapid advancements across many fields such as information retrieval,knowledge recommendation,and others.Although significant improvements have been achieved in the research of KG,the recently developed machine learning technologies and methods in big data analysis still cannot effectively match the demands of exploiting and using KG due to the complexity of real world application scenarios.In effect,a myriad of technical challenges persist,necessitating adept resolution to release the full potential of large-scale KG as a powerful methodology for intelligent information service.Here,we summarize the future research trends and challenges of machine learning-driven large-scale KG construction from three aspects.

    4.1 Relation Extraction

    On one hand,the current focus of relation extraction predominantly revolves around the monolingual text.However,in the real-world,factual knowledge finds its repository in diverse sources,such as multilingual texts,pictures,audio,and video.Expanding the horizons of relation extraction to encompass these various sources stands as a promising avenue for future exploration,heralding the potential to broaden the spectrum of extracted relations and extend the scope of knowledge coverage[56].On the other hand,for the deep learning-based relation extraction methods,the integration of syntactic trees via neural network models yields the effective amalgamation of syntactic information.However,it also leads to the introduction of a large amount of noise,which poses an impact on the accuracy of the model [57,60].Constructing multiple possible syntactic trees of sentences and fusing them for relation extraction may be a development prospect.Furthermore,the open fieldbased relationship extraction is constantly updated and iterative.How to introduce deep learning models to achieve rapid learning of new relationships and knowledge is also a problem that needs to be explored[129].

    4.2 Link Prediction

    Link prediction plays an important role in the design of KG,serving as a critical component in inferring absent relationships.For example,as shown in Fig.15,where known relationships between entities A and B,B and C,and C and D,the link prediction method enables the inference of the relationship type between entities A and D.By leveraging the existing relationships within the KG,this technique enables the identification and prediction of unobserved connections,thereby enhancing the overall comprehension and knowledge extraction from the graph.Generally,the implementation of link prediction is mainly based on triples used to constitute knowledge.However,the types of knowledge are rich and diverse,and some complex knowledge cannot be directly represented by triples.Hence,different knowledge representation methods need to be set up for different scenarios.For instance,considering the temporal dynamism of factual values,a compound value type structure has been introduced,involving auxiliary nodes to representn-order relations and temporal attributes for facts [130].Multiplen-order relations can be represented by a single (n+1)-order tensor,which is solved by higher-order tensor decomposition using the RESACL model[131].The representation of learning-based link prediction is still in the initial stage of exploration[132].It achieves unsatisfactory performance on large-scale KG with strong sparsity and the representation of low-frequency entities and relationships.It is urgent to design a more efficient online learning scheme for KG.Concurrently,the domain of network embedding-based algorithms,including those grounded in graph neural networks(GNN),has showcased compelling computational prowess in task completion.Specifically,the attention mechanism-based heterogeneous GNN is conducive to capturing information of various semantics in KGs [133].Additionally,the introduction of the multi-scale dynamic convolutional network(M-DCN)has provided a framework for representing KG embeddings[134].Therefore,how to creatively investigate those algorithms in the achievement of link prediction for KG is also an interesting direction[135,136].

    Figure 15:Node roles for link prediction

    4.3 Construction of Industrial Knowledge Graph

    In a broader context,the construction of knowledge graphs(KGs)has spanned general domains,and the corresponding theoretical investigations have followed suit.However,the construction of KG for a specific field,especially in industry,has attracted less attention in recent years.General KGs emphasize the entity layer and it is difficult to generate a global ontology pattern.Notably,there are many differences between industrial KG and general KG [137].The industrial KG has a clear industry background,while the entities have rich data patterns.Furthermore,the industrial KGs need comprehensive consideration of personnel at various hierarchical levels.This has led to new diverse challenges for researchers focusing on the design of a large-scale industrial KG.Thus,it needs to be explored in this direction,while presenting some new machine learning-driven methods in these fields[138,139].

    Over the last year,significant advancements have been made in NLP tasks with the emergence of large language models(LLMs)like ChatGPT[140],Dolly[141],and LLaMA[142].Despite their achievements,LLMs are black-box models,lacking transparency in tracking their search process.On the other hand,KG offers a high level of professional feasibility,providing more accurate answers and better interpretability compared to LLMs.Leveraging the respective advantages of LLMs and KG,we can utilize LLMs for data annotation and data enhancement in the future,facilitating the rapid application of industrial KGs[143].Meanwhile,industrial KG can serve as an external knowledge base,enabling the introduction of specified constraints to LLMs[144].This allows for controlled content generation and enhances the adaptability of LLMs within specific industrial fields.

    5 Conclusion

    The centrality of KGs within the realm of next-generation search engines has established them as a pivotal focal point within the sphere of intelligent information processing,coinciding with the advent of the big data era.With the ever-increasing demand for processing performance in consideration of complex application scenarios,there are many exploratory works to be explored in the KG construction.In this article,we comprehensively review the key technologies in KG construction from the perspective of machine learning-related implementation methods.We are concerned with the core construction algorithms of KG in three aspects,including entity learning,ontology learning,and knowledge reasoning.Especially,the machine learning-driven algorithms for entity extraction,relation extraction,entity linking,and link prediction are deeply analyzed.In addition,considering the current development level of machine learning methods,we summarize some key problems and possible research trends in the construction of large-scale KG to serve as an impetus for researchers to work in the future.

    Acknowledgement:None.

    Funding Statement:This work was supported in part by the Beijing Natural Science Foundation under Grants L211020 and M21032,in part by the National Natural Science Foundation of China under Grants U1836106 and 62271045,and in part by the Scientific and Technological Innovation Foundation of Foshan under Grants BK21BF001 and BK20BF010.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design:Z.Zhao,X.Luo;data collection:Z.Zhao,M.Chen,L.Ma;analysis and interpretation of results:Z.Zhao,X.Luo,M.Chen;draft manuscript preparation:Z.Zhao,X.Luo,L.Ma.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The presented data are open-sourced and can be accessed through the“Web of Science”and“Papers with Code website”.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    韩国av一区二区三区四区| 美女免费视频网站| 床上黄色一级片| 岛国在线免费视频观看| 18禁黄网站禁片午夜丰满| 亚洲人成网站在线播| 麻豆av噜噜一区二区三区| 国产精品久久久久久久久免| 精品午夜福利视频在线观看一区| 欧美最黄视频在线播放免费| 在线天堂最新版资源| 18禁黄网站禁片免费观看直播| 日韩欧美一区二区三区在线观看| 搡老妇女老女人老熟妇| 非洲黑人性xxxx精品又粗又长| 国产大屁股一区二区在线视频| 日本熟妇午夜| 亚洲真实伦在线观看| 亚洲av美国av| 97热精品久久久久久| 色视频www国产| 99热精品在线国产| 亚洲专区国产一区二区| 精品国产三级普通话版| 日韩av在线大香蕉| 国产精品永久免费网站| 精品不卡国产一区二区三区| 国产精品免费一区二区三区在线| 中亚洲国语对白在线视频| 如何舔出高潮| 亚洲av五月六月丁香网| 久久精品国产亚洲av香蕉五月| 日本a在线网址| 99热这里只有是精品在线观看| 免费一级毛片在线播放高清视频| 久久欧美精品欧美久久欧美| 国产黄色小视频在线观看| 韩国av在线不卡| 波多野结衣巨乳人妻| 欧美一区二区精品小视频在线| 国产精品人妻久久久影院| 日韩高清综合在线| 在线播放国产精品三级| 亚洲av二区三区四区| 最新中文字幕久久久久| 最后的刺客免费高清国语| 国产精品久久电影中文字幕| 国产老妇女一区| 草草在线视频免费看| 69av精品久久久久久| 久久久精品欧美日韩精品| 欧美色视频一区免费| 真实男女啪啪啪动态图| 国产精品综合久久久久久久免费| 琪琪午夜伦伦电影理论片6080| 能在线免费观看的黄片| 亚洲精华国产精华液的使用体验 | 别揉我奶头 嗯啊视频| 欧美日韩瑟瑟在线播放| 久久精品国产亚洲av香蕉五月| 夜夜爽天天搞| 国产高清三级在线| 国产乱人视频| 国内毛片毛片毛片毛片毛片| 色视频www国产| 欧美人与善性xxx| av在线观看视频网站免费| 91久久精品国产一区二区三区| 18禁黄网站禁片午夜丰满| 日韩一区二区视频免费看| 12—13女人毛片做爰片一| АⅤ资源中文在线天堂| 非洲黑人性xxxx精品又粗又长| 在线国产一区二区在线| 精品人妻1区二区| 欧美性猛交╳xxx乱大交人| xxxwww97欧美| 尾随美女入室| 一a级毛片在线观看| 欧美zozozo另类| 亚洲av电影不卡..在线观看| 全区人妻精品视频| 男女边吃奶边做爰视频| 干丝袜人妻中文字幕| 成人永久免费在线观看视频| 日韩精品有码人妻一区| 真人一进一出gif抽搐免费| 久久久精品大字幕| 人人妻人人澡欧美一区二区| 俄罗斯特黄特色一大片| 亚洲男人的天堂狠狠| 婷婷亚洲欧美| 国产亚洲91精品色在线| 麻豆精品久久久久久蜜桃| 欧美日本视频| 少妇被粗大猛烈的视频| 日韩高清综合在线| 色综合婷婷激情| 波多野结衣高清无吗| 狂野欧美激情性xxxx在线观看| 一级黄片播放器| 亚洲精华国产精华液的使用体验 | av在线蜜桃| 麻豆国产97在线/欧美| 欧美xxxx性猛交bbbb| 久久久久久久久中文| 人人妻人人澡欧美一区二区| 久久天躁狠狠躁夜夜2o2o| 乱码一卡2卡4卡精品| 午夜福利在线观看免费完整高清在 | 欧美一级a爱片免费观看看| 在线观看免费视频日本深夜| 国产精品日韩av在线免费观看| 午夜激情欧美在线| 免费无遮挡裸体视频| 少妇丰满av| 日韩av在线大香蕉| 国产精品电影一区二区三区| 亚洲成人中文字幕在线播放| 高清毛片免费观看视频网站| 国产激情偷乱视频一区二区| 国产三级在线视频| 午夜老司机福利剧场| 日本五十路高清| 噜噜噜噜噜久久久久久91| 最新在线观看一区二区三区| 久久精品国产亚洲av涩爱 | 久久欧美精品欧美久久欧美| 欧美日韩乱码在线| 国产成人a区在线观看| 12—13女人毛片做爰片一| 嫩草影院精品99| 久久国内精品自在自线图片| 国产高清三级在线| 成人国产一区最新在线观看| 免费高清视频大片| 久久精品国产亚洲网站| 国产毛片a区久久久久| 丰满的人妻完整版| 真人一进一出gif抽搐免费| 久久久久久久久中文| 深夜a级毛片| 性插视频无遮挡在线免费观看| 最近最新中文字幕大全电影3| 悠悠久久av| 国产精品自产拍在线观看55亚洲| 精品午夜福利视频在线观看一区| 12—13女人毛片做爰片一| 亚洲性夜色夜夜综合| 女的被弄到高潮叫床怎么办 | 能在线免费观看的黄片| h日本视频在线播放| 美女黄网站色视频| 少妇人妻精品综合一区二区 | 欧美另类亚洲清纯唯美| 欧美一区二区精品小视频在线| 深爱激情五月婷婷| 久久精品国产亚洲av天美| 欧美三级亚洲精品| a级毛片a级免费在线| 国产v大片淫在线免费观看| 变态另类成人亚洲欧美熟女| 亚洲av成人av| 久久99热6这里只有精品| 成年版毛片免费区| 久久久久久久午夜电影| 日韩欧美在线二视频| 日韩中文字幕欧美一区二区| 欧美性猛交黑人性爽| 男人狂女人下面高潮的视频| 国产一区二区三区视频了| 亚洲狠狠婷婷综合久久图片| 观看免费一级毛片| 亚洲乱码一区二区免费版| 国产精品久久久久久av不卡| 欧美一级a爱片免费观看看| 日本五十路高清| 日本 欧美在线| 国产色婷婷99| 亚洲精品成人久久久久久| 国内精品宾馆在线| 国内精品一区二区在线观看| 久久人人爽人人爽人人片va| 成人鲁丝片一二三区免费| 一个人看的www免费观看视频| 长腿黑丝高跟| 精品99又大又爽又粗少妇毛片 | 亚洲成a人片在线一区二区| 白带黄色成豆腐渣| 亚洲av日韩精品久久久久久密| 如何舔出高潮| av在线蜜桃| 国产黄色小视频在线观看| 国产又黄又爽又无遮挡在线| 国产 一区精品| 男插女下体视频免费在线播放| 在线播放无遮挡| 亚洲美女视频黄频| 97超级碰碰碰精品色视频在线观看| 少妇高潮的动态图| 成人特级av手机在线观看| 老司机福利观看| 人妻夜夜爽99麻豆av| 国产一区二区亚洲精品在线观看| 亚洲久久久久久中文字幕| 不卡一级毛片| 亚洲欧美日韩卡通动漫| 成人精品一区二区免费| 亚洲av美国av| 欧美高清性xxxxhd video| 老熟妇仑乱视频hdxx| 最近中文字幕高清免费大全6 | 成人国产麻豆网| 能在线免费观看的黄片| 成人特级黄色片久久久久久久| 舔av片在线| 亚洲第一电影网av| 国产精品无大码| 国产亚洲av嫩草精品影院| 国产伦一二天堂av在线观看| 麻豆国产97在线/欧美| 一卡2卡三卡四卡精品乱码亚洲| 国产免费一级a男人的天堂| 两个人视频免费观看高清| 亚洲欧美日韩高清在线视频| 亚洲美女视频黄频| 亚洲精品456在线播放app | 成年免费大片在线观看| 12—13女人毛片做爰片一| 久久6这里有精品| 成人欧美大片| 久久精品国产清高在天天线| 老司机福利观看| 久久久久久九九精品二区国产| 一个人观看的视频www高清免费观看| 国产精品野战在线观看| 亚洲一级一片aⅴ在线观看| 亚洲欧美激情综合另类| 一进一出抽搐gif免费好疼| 欧美zozozo另类| 国产精品野战在线观看| 美女高潮喷水抽搐中文字幕| 成人二区视频| 大型黄色视频在线免费观看| 嫁个100分男人电影在线观看| 国产精品亚洲美女久久久| 波多野结衣高清作品| 国产v大片淫在线免费观看| ponron亚洲| 色精品久久人妻99蜜桃| 伊人久久精品亚洲午夜| 欧美性猛交╳xxx乱大交人| 淫妇啪啪啪对白视频| 18禁黄网站禁片免费观看直播| 能在线免费观看的黄片| 国产一级毛片七仙女欲春2| 日本爱情动作片www.在线观看 | 日韩人妻高清精品专区| 一本一本综合久久| 欧美日韩综合久久久久久 | 看黄色毛片网站| 香蕉av资源在线| 性欧美人与动物交配| 国产一级毛片七仙女欲春2| 欧美在线一区亚洲| 国产免费男女视频| 91久久精品国产一区二区三区| 老女人水多毛片| 少妇人妻一区二区三区视频| 久久精品国产亚洲网站| 久久精品人妻少妇| 亚洲va在线va天堂va国产| 日本与韩国留学比较| 亚洲精华国产精华精| 免费av观看视频| 免费不卡的大黄色大毛片视频在线观看 | 少妇人妻精品综合一区二区 | 中文字幕久久专区| 黄色女人牲交| 亚洲av不卡在线观看| 男女啪啪激烈高潮av片| 久久久国产成人免费| 亚洲av中文字字幕乱码综合| 日韩中字成人| 麻豆av噜噜一区二区三区| 色综合色国产| 88av欧美| 国产综合懂色| 国产精品久久久久久精品电影| 成人鲁丝片一二三区免费| 日本免费一区二区三区高清不卡| 桃红色精品国产亚洲av| 久久久久国产精品人妻aⅴ院| 舔av片在线| 99热精品在线国产| 亚洲自偷自拍三级| 高清在线国产一区| 午夜a级毛片| 成年免费大片在线观看| 亚洲真实伦在线观看| 亚洲av免费高清在线观看| 亚洲精品一卡2卡三卡4卡5卡| 亚洲最大成人av| 成人美女网站在线观看视频| 在线免费观看的www视频| 如何舔出高潮| 亚洲国产精品sss在线观看| 欧美xxxx性猛交bbbb| 色哟哟哟哟哟哟| 久久午夜福利片| 日韩精品中文字幕看吧| 亚洲欧美日韩卡通动漫| 黄色一级大片看看| 欧美成人性av电影在线观看| 日日摸夜夜添夜夜添av毛片 | 日韩国内少妇激情av| 九九热线精品视视频播放| 亚洲在线观看片| 亚洲av.av天堂| 国产久久久一区二区三区| 国产一区二区三区av在线 | 日本欧美国产在线视频| 一区二区三区激情视频| 一本久久中文字幕| 免费av不卡在线播放| 99国产精品一区二区蜜桃av| 少妇高潮的动态图| 欧美黑人巨大hd| 亚洲va日本ⅴa欧美va伊人久久| 日韩强制内射视频| 中文字幕高清在线视频| 噜噜噜噜噜久久久久久91| 99久久精品一区二区三区| 亚洲自偷自拍三级| 国产探花在线观看一区二区| 波多野结衣高清作品| 久久人妻av系列| 婷婷六月久久综合丁香| 99热这里只有精品一区| 国产精品无大码| 国产视频内射| 午夜免费成人在线视频| 色视频www国产| 亚洲真实伦在线观看| 少妇人妻精品综合一区二区 | 免费av观看视频| 琪琪午夜伦伦电影理论片6080| 亚洲三级黄色毛片| 女生性感内裤真人,穿戴方法视频| 精品久久久久久久久久免费视频| 97超视频在线观看视频| 国产色婷婷99| 成年女人永久免费观看视频| 国产伦精品一区二区三区四那| 亚洲av成人av| 天堂av国产一区二区熟女人妻| 久久精品影院6| 国内精品久久久久精免费| 国产91精品成人一区二区三区| 日韩欧美在线二视频| 成人国产麻豆网| 国产精品久久电影中文字幕| 精品一区二区免费观看| 男女那种视频在线观看| 国产av不卡久久| 国产又黄又爽又无遮挡在线| 波多野结衣高清作品| 亚洲在线观看片| 国产精品一区二区三区四区久久| 国内精品久久久久精免费| 久久久久久久久久黄片| 啦啦啦观看免费观看视频高清| 亚洲国产欧洲综合997久久,| 99精品在免费线老司机午夜| 国产精品日韩av在线免费观看| 免费一级毛片在线播放高清视频| 久久精品综合一区二区三区| 中国美白少妇内射xxxbb| 极品教师在线免费播放| 国产免费男女视频| 亚洲黑人精品在线| 亚洲人成网站在线播放欧美日韩| 日本欧美国产在线视频| 精品人妻偷拍中文字幕| 国产成人av教育| 大型黄色视频在线免费观看| 桃色一区二区三区在线观看| 日韩欧美精品v在线| 91精品国产九色| 国产一区二区在线观看日韩| 黄色丝袜av网址大全| 九九久久精品国产亚洲av麻豆| 久久精品国产鲁丝片午夜精品 | 免费一级毛片在线播放高清视频| a级毛片a级免费在线| 日韩在线高清观看一区二区三区 | 大又大粗又爽又黄少妇毛片口| 亚洲国产欧洲综合997久久,| 中文字幕高清在线视频| 听说在线观看完整版免费高清| 在线观看舔阴道视频| 亚洲av一区综合| 亚洲,欧美,日韩| 岛国在线免费视频观看| 精品不卡国产一区二区三区| 精品欧美国产一区二区三| 两个人的视频大全免费| 亚洲欧美激情综合另类| 国产精品无大码| 特大巨黑吊av在线直播| 国产蜜桃级精品一区二区三区| 乱码一卡2卡4卡精品| 亚洲自偷自拍三级| 人人妻人人看人人澡| 狂野欧美激情性xxxx在线观看| 尾随美女入室| 啪啪无遮挡十八禁网站| 国产熟女欧美一区二区| av女优亚洲男人天堂| 国产精品综合久久久久久久免费| 亚洲成a人片在线一区二区| 亚洲国产精品sss在线观看| 中文字幕免费在线视频6| 九九久久精品国产亚洲av麻豆| 中文在线观看免费www的网站| 国产成年人精品一区二区| 国产色婷婷99| 亚洲性夜色夜夜综合| 两性午夜刺激爽爽歪歪视频在线观看| 91久久精品国产一区二区成人| 国产精品1区2区在线观看.| 男插女下体视频免费在线播放| 三级国产精品欧美在线观看| 午夜激情福利司机影院| 久久国产乱子免费精品| 又爽又黄a免费视频| 国产男人的电影天堂91| netflix在线观看网站| 免费看美女性在线毛片视频| 欧美激情久久久久久爽电影| 男插女下体视频免费在线播放| 国产蜜桃级精品一区二区三区| 亚洲天堂国产精品一区在线| 国产一区二区在线av高清观看| 韩国av一区二区三区四区| 岛国在线免费视频观看| 久久香蕉精品热| 国产69精品久久久久777片| 国产一区二区三区av在线 | 99热这里只有是精品在线观看| 美女免费视频网站| 国产大屁股一区二区在线视频| 久久人人爽人人爽人人片va| 国产午夜精品论理片| 亚洲国产精品成人综合色| 精品不卡国产一区二区三区| 午夜爱爱视频在线播放| 成人三级黄色视频| 国产又黄又爽又无遮挡在线| 91久久精品国产一区二区三区| 两性午夜刺激爽爽歪歪视频在线观看| videossex国产| 三级国产精品欧美在线观看| 日本黄色片子视频| 国产精品久久久久久精品电影| 人妻丰满熟妇av一区二区三区| 国国产精品蜜臀av免费| 搞女人的毛片| 成人无遮挡网站| 国产一区二区三区视频了| 韩国av在线不卡| 一a级毛片在线观看| 男女那种视频在线观看| 我要看日韩黄色一级片| 又黄又爽又免费观看的视频| 欧美成人一区二区免费高清观看| 不卡一级毛片| 欧美3d第一页| 国产精品一及| 欧美高清成人免费视频www| 最近中文字幕高清免费大全6 | 丝袜美腿在线中文| 身体一侧抽搐| 日本撒尿小便嘘嘘汇集6| 一区福利在线观看| 成人一区二区视频在线观看| 精品免费久久久久久久清纯| 久久精品91蜜桃| 一级黄片播放器| 最新中文字幕久久久久| 99精品在免费线老司机午夜| 亚洲av第一区精品v没综合| av福利片在线观看| 欧美xxxx性猛交bbbb| 老熟妇乱子伦视频在线观看| 禁无遮挡网站| 自拍偷自拍亚洲精品老妇| 九九爱精品视频在线观看| АⅤ资源中文在线天堂| 亚洲va日本ⅴa欧美va伊人久久| 高清毛片免费观看视频网站| 神马国产精品三级电影在线观看| 中文字幕免费在线视频6| 欧美精品国产亚洲| 欧美人与善性xxx| 俺也久久电影网| 久久欧美精品欧美久久欧美| 欧美国产日韩亚洲一区| 欧美黑人欧美精品刺激| 在线观看舔阴道视频| 久久天躁狠狠躁夜夜2o2o| 22中文网久久字幕| 如何舔出高潮| 久久精品国产清高在天天线| 亚洲av成人精品一区久久| 我要搜黄色片| 免费观看在线日韩| 国产精品美女特级片免费视频播放器| 精品久久久久久久久亚洲 | 长腿黑丝高跟| 日韩国内少妇激情av| 99久久精品国产国产毛片| 国产精品久久久久久亚洲av鲁大| 亚洲图色成人| 女生性感内裤真人,穿戴方法视频| 少妇熟女aⅴ在线视频| 嫁个100分男人电影在线观看| 精品人妻1区二区| 久久久久精品国产欧美久久久| 色综合站精品国产| .国产精品久久| 少妇人妻精品综合一区二区 | 麻豆成人av在线观看| 中文字幕精品亚洲无线码一区| 久久精品国产亚洲av香蕉五月| 国产在线男女| 午夜福利在线观看吧| 久久精品人妻少妇| 高清在线国产一区| 亚洲性夜色夜夜综合| 九九在线视频观看精品| 麻豆精品久久久久久蜜桃| 国产精品一及| 国产午夜精品论理片| 禁无遮挡网站| 成人三级黄色视频| 成人综合一区亚洲| 日本a在线网址| 1000部很黄的大片| 91麻豆av在线| 精品久久国产蜜桃| 免费av不卡在线播放| 很黄的视频免费| 精品一区二区免费观看| 国产亚洲精品av在线| 高清毛片免费观看视频网站| 无遮挡黄片免费观看| 国产 一区 欧美 日韩| 18禁在线播放成人免费| 亚洲电影在线观看av| 国产单亲对白刺激| 欧美精品国产亚洲| 白带黄色成豆腐渣| 一级黄片播放器| 色吧在线观看| 日韩国内少妇激情av| 麻豆成人午夜福利视频| eeuss影院久久| 亚洲av中文字字幕乱码综合| 国模一区二区三区四区视频| 波多野结衣高清无吗| 国产aⅴ精品一区二区三区波| 美女xxoo啪啪120秒动态图| 日本-黄色视频高清免费观看| 干丝袜人妻中文字幕| 长腿黑丝高跟| 国产黄片美女视频| xxxwww97欧美| 欧美黑人欧美精品刺激| 黄色一级大片看看| 少妇人妻精品综合一区二区 | 日本黄大片高清| 国产精品久久久久久久久免| 有码 亚洲区| 一级黄色大片毛片| 啦啦啦啦在线视频资源| 国产精品人妻久久久影院| 欧美激情久久久久久爽电影| 热99re8久久精品国产| 久久精品国产清高在天天线| 两性午夜刺激爽爽歪歪视频在线观看| 简卡轻食公司| 久久久久久久久大av| 两性午夜刺激爽爽歪歪视频在线观看| 99精品久久久久人妻精品| 国产蜜桃级精品一区二区三区| 亚洲性久久影院| 国产欧美日韩一区二区精品| 婷婷六月久久综合丁香| 国产午夜精品论理片| 亚洲国产色片| 国产精品久久久久久精品电影| 亚洲,欧美,日韩| 99精品久久久久人妻精品| 丰满的人妻完整版| 免费看光身美女| 18禁黄网站禁片午夜丰满| 免费电影在线观看免费观看| 国内精品一区二区在线观看| 一级黄片播放器| 国产精品一区二区三区四区久久| 国产黄片美女视频| 啦啦啦啦在线视频资源| 麻豆国产97在线/欧美| 九九在线视频观看精品|