• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Exploiting Structural Similarities to Classify Citations

    2021-12-15 12:45:54MuhammadSaboorAhmedandMuhammadTanvirAfzal
    Computers Materials&Continua 2021年2期

    Muhammad Saboor Ahmed and Muhammad Tanvir Afzal

    Capital University of Science and Technology, Islamabad, Pakistan

    Abstract:Citations play an important role in the scientific community by assisting in measuring multifarious policies like the impact of journals,researchers,institutions, and countries.Authors cite papers for different reasons, such as extending previous work, comparing their study with the state-of-the-art, providing background of the field, etc.In recent years, researchers have tried to conceptualize all citations into two broad categories, important and incidental.Such a categorization is very important to enhance scientific output in multiple ways, for instance, (1) Helping a researcher in identifying meaningful citations from a list of 100 to 1000 citations (2) Enhancing the impact factor calculation mechanism by more strongly weighting important citations, and (3) Improving researcher,institutional, and university rankings by only considering important citations.All of these uses depend upon correctly identifying the important citations from the list of all citations in a paper.To date,researchers have utilized many features to classify citations into these broad categories:cue phrases, in-text citation counts,and metadata features,etc.However,contemporary approaches are based on identification of in-text citation counts, mapping sections onto the Introduction, Methods, Results, and Discussion (IMRAD) structure, identifying cue phrases,etc.Identifying such features accurately is a challenging task and is normally conducted manually, with the accuracy of citation classification demonstrated in terms of these manually extracted features.This research proposes to examine the content of the cited and citing pair to identify important citing papers for each cited paper.This content similarity approach was adopted from research paper recommendation approaches.Furthermore, a novel section-based content similarity approach is also proposed.The results show that solely using the abstract of the cited and citing papers can achieve similar accuracy as the stateof-the-art approaches.This makes the proposed approach a viable technique that does not depend on manual identification of complex features.

    Keywords:Section-wise similarity; citation classification; content similarity;important citation

    1 Introduction

    Researchers in all disciplines build upon the foundations laid by former researchers.This notion is succinctly summed up in Ziman’s statement that “a scientific paper does not stand alone; it is embedded in the literature of a subject”[1].Research in the same field is interlinked,which means that existing research must always be brought in relation to former researches.New findings must be written up in the form of a scientific research paper.This research paper is then shared with other researchers so that the research process can be validated and can be continued.Therefore,while writing research findings,scholars acknowledge the scientific support they have received from former work.These acknowledgements are found in the reference section and termed as‘citations’.Ziman[1]and Narin[2]highlighted the true strength of analyzing citations can aid in producing and authenticating different research studies.They argue that the popularity and significance of a scientific work is expressecd through the frequency with which it is cited.The citations are considered an important tool for assessing the academic and scientific strength of institutions and individuals.They can also be used to investigate authors’ or institutions’ reputations within the overall scientific community[1,2].

    The utility of citation-based measures is multifaceted.They are used to decide award nominees such as the Nobel prize[3]as well as research funding[4].They can also be used to evaluate peer judgments[5]rank researchers [6-9] and countries [10].In the late 1960s, Garfield, the founder of Thomson ISI, defined a number of reasons for citations [11,12].This definition offers numerous opportunities to critically investigate citation behaviour[13,14].

    Although citations are included to achieve specific objectives,citation count approaches[15]have never tried to distinguish between these objectives.Consequently, such approaches fail to maintain a balance between the act of citation itself and the purpose for which a citation is made.Instead, they blindly consider all citations equal.This discrepancy has led to achieve research in this area [16,17].A detailed examination of citation counts was carried out by Benedictus et al.[18].They concluded that citation count-based measures have inherited problems for example,they shift focus from quality to quantity.

    Researchers have developed recommendations for improving the quality and reducing the emphasis on quantity in citation counts [19,20].Generally, researchers believe that the reasons for citations must be critically considered in order to acknowledge the quality of different scholars’ work [21].Is it possible to differentiate between various reasons for citations? Existing citation annotation approaches proceed manually.The manual approaches rely upon interviewing the citer.Usually, authors are interviewed to share the reason for citing a particular piece of work on two different occasions:after the publication process is over and while writing the article[22,23].

    Finney[24]argues that the citation classification process can be automated.Confirming this argument,later researchers took steps to classify citations into various categories [21,25].However, while this idea made a significant contribution, it also brought a discouraging element to the fore, as citations were classified based on several ambiguous reasons.As a result of this ambiguity, the major limitation of a simple citation count approach was not effectively addressed.Presently, two major types of citations have been identified:important and un-important classes [16,17,26].

    What do we mean by important and un-important classes? Generally, during the process of writing a paper, only a few citations in the reference list have a significant impact on the citing study.This impact needs to be precisely described.Zhu et al.[17] has provided a solution by arguing that an influential research study convinces the research community to adopt or extend the presented idea [17].To establish a clear distinction between important and un-important citations, we need to examine contemporary citation classification mechanisms.Garzone et al.[25] extends the work of Finney [24] by implementing her suggestion of “associating cue words with citation function and using citation location in the classification algorithm” [25].Both [16] and [17] argue that the citation relations discussed by Finney[24] and Garzone et al.[25] are important.In contrast, Garzone et al.[25] also cite several other studies as background information, such as the citation categories introduced by Garfield [12].Based on the aforementioned discussion, the studies [16,17] classify citations into two major categories.The first category of citations aims to provide background knowledge, which forms the foundation of the proposed study.Researchers such as Zhu et al.[17] have termed this category as “non-influential and incidental”,whereas Valenzuela et al.[16] have termed it as non-important and incidental.We use the term “nonimportant” for this category.The second category of citations seeks to extend or apply the cited work.This category is termed as “influential” by Zhu et al.[17] and “important” by Valenzuela et al.[16].We use the term “important”for this category.

    Researchers have recently proposed different features and strategies to identify the important categories.For example,Valenzuela et al.[16]evaluated 12 features and concluded that in-text citation count was the most accurate feature,with a precision of 0.65.However,identifying citation tags from research papers is a challenge[27].The Valenzuela’s approach was further extended recently by Nazir et al.[28]wherein in-text citation counts within different logical sections of the paper(Introduction,Related Work,Methodology,and Results)were examined.This approach has achieved a precision of 0.84.However,there are also two major issues with this approach:(1)Accurately identifiying logical sections and mapping section headings onto the logical sections,and(2)Accurately identifying in-text citations[27].The best-known approach for mapping section headings onto section categories has an accuracy of 78%[29].Another binary citation classification approach presented by Qayyum et al.[26]has achieved a precision of 0.72 by examining metadata and cue phrases.However,this approach again involves the construction of cue phrases and identification of in-text citation frequencies.All of these recent approaches have certain limitations resulting from their reliance on the accurate identification of the following parameters:in-text citation counts,an updated dictionary of cue phrases, in-text citation extraction from sentences, and mapping section headings to logical sections.The extraction accuracy of each of the above parameters is around 70% [27,29].However, the above approaches extract these parameters in a semi-automatic way, which has been demonstrated to be accurate when the parameters are readily and accurately available.

    This critical discussion highlights the need for an approach that does not involve such a complex extraction of parameters, which is often inaccurate.Examinination of the relevant related literature shows that a content-based approaches were successfully employed by 55% of around 200 papers applied in the research paper recommendation domain in the last 16 years [30].This motivated us to evaluate the suitability of the content-based approach for identifying important citations.Moreover, in addition to evaluating the existing content-based approach, this paper further proposes a novel section-wise contentbased approach.The results indicate significant precision and recall values without any manual identification of complex parameters.The study’s in-depth analysis of papers’ complete content and content within different sections suggests that content-related similarities in the abstracts of the cited and citing papers be used to classify the citing paper as an important/non-important citation for the cited paper.Therefore, the proposed approach has a great potential to be applied in citation indexes and open new horizons for future researches in citation classification.

    2 Literature Review

    The citation count is utilized to conduct various types of bibliometric analyses with multidimensional utilities.Such analyses have been used to build indexing systems [31,32] and formulate various academic policies and present awards such as Noble prizes [3].These analyses have also been used to rank researchers [6-9] and countries [10].However, researchers believe that all citations cannot be conisedered equal, and each citation should be treated according to its true standing [1,14,16,17,21,33].

    A specific study might be cited for myriad reasons.Garfield [12] was the first researcher to analyze citation behaviour [12].He identified 15 citation reasons by examining various factors such as the citation’s location in the paper and scrutinizing the differences and patterns.Some of the reasons include(1) acknowledging the contributions of predecessors, (2) highlighting fundamental contextual details,(3) extending existing work and targeting expanded objective(s), etc.Later, Liptez [34] identified various classes of citations [34].However, while both studies appropriately conceptualized the notion of citation reasons, no statistical measures were introduced [14].Nevertheless, despite this shortcoming, these studies attracted enormous attention from the research community.Consequently, many empirical investigations have been carried out to identify citation reasons.Subsequently, other studies have attempted to capture actual citation behaviour[13,35].However, common to all of these approaches is the treatment of all citations as being on the same level of importance.

    According to Zhu et al.[17], studies involving straightforward quantitative citation analysis can be enhanced by eliminating incidental citations from the citation count.Additionally, maintaining a list of only important citations can be of substantial help for scholars seeking to identify influential studies on a specific topic.Until the mid-1990s, citation reasons were manually identified.For example, a general trend at that time was to interview authors during the process of writing an article or after their proposed article had been formally published, requesting them to describe the specific reasons for citing particular works [22,23].However, differentiating scholars’ citing behaviour using cognitive approaches seemed rather impractical.Therefore, researchers have realized the need for an automated system to identify and classify citation reasons.

    Finney[24]demonstrated that the citation classification process can be automated.She created a citation function on an experimental basis.Later,she posited a relationship between cue words and citation location and combined it with the citation function.Though her approach was not fully automatic, however, it underscored the probability of developing a fully automatic citation classification mechanism in the future[25].However, other researchers were a bit reluctant to acknowledge Finney’s contribution because it was a doctoral thesis rather than a formal publication[14].

    Drawing inspiration from Finney’s approach,Garzone et al.[25]took their place among the trendsetters by creating an automated citation classification system.The authors argued that Finney’s approach had several limitations, which they addressed by creating 35 categories to classify citations.They were able to successfully solve the classification task by introducing 14 parsing rules and 195 lexical matching rules.The dataset comprised 9 biochemistry and 11 physics articles.The system found to be stable.It produced average results on unseen articles and appreciative results on previously seen articles.Though the system produced encouraging results, there was also one concern:due to many classes, the system was unable to neatly distinguish between divergent classes.Pham et al.[36] classified citations from 482 citation contexts into four categories.They employed the ripple-down rules (RDR) hierarchy using cue phrases.

    Our proposed model is similar to Zhu et al.and Valenzuela et al.[16,17].Both studies have performed a similar binary citation classification based on the same distinction between important and non-important citations as the present study.The classification task was carried out using (1) in-text count-based features, (2) similarity-based features, (3) context-based features, (4) position-based features, and (5)miscellaneous features.The fundamental idea behind the proposed technique was to recognize the set of references that has an academic influence on the citing paper.

    Zhu et al.[17]characterized academic influence as a reference that serves as a source for extracting an idea,problem,method,or experiment.They generated a total of 3143 paper-reference pairs from 100 papers extracted from the Association of Computational and Linguistics (ACL) anthology.These pairs were annotated by the authors of the citing papers.In contrast, Valenzuela et al.[16] introduced a supervised classification approach to identify important and non-important citations.The authors have extracted 465 paper-citation pairs from the ACL anthology.Two domain experts annotated these pairs as either important and non-important citations.Inter-annotator agreement was between two annotators was 93.9%.Twelve features were used to classify the citations into important and non-important classes.These features include the total number of direct citations, the number of direct citations per section, the total number of indirect citations,the number of indirect citations per section,author overlap,etc.Two different classifiers,random forest and support vector machine(SVM),were used to train these features.Both models attained 0.90 recall and 0.65 precision.Zhu’s approach[17]was criticized by Valenzuela et al.[16].The latter claimed that biased annotation cannot be ruled out if citations are coded by the citing authors.

    Another approach was proposed by Qayyum et al.[26].They utilized metadata and cue terms to discover important citations.However, a limitation of this approach is that cue terms are identified from the papers’ content and thus need to be updated for different datasets and domains.There is a need for a domain expert to manually identify cue terms for each domain and keep them updated.Nazir et al.extended Valenzuela’s approach by identifying suitable weights for in-text citation frequencies in different sections.This approach has outperformed the previous approaches.However,a critical examination highlights limitations regarding accurately mapping section headings onto logical sections and accurately identifying in-text citations.Although there are many approaches to identifying important citations [16,26,28], in order to practically apply those approaches, there is a need to accurately identify the following information:(1) Accurately identify in-text citations [16,28],(2) Accurately map section headings onto logical sections [28], and (3) Create an updated accurate list of cue terms [26].Existing approaches have reported precisions up to 0.84.However, these approaches depend on the accurate extraction of the above parameters and either ignore inaccurate results and correct the missing values manually to demonstrate the power of these parameters.The automatic extraction of such parameters is still a challenge [27,29].This motivates us to fill this research gap by creating a novel apprach that does not require these parameters to be extracted.A critical examination of the related domain of research paper recommendations motivated us to use the content of the cited and citing papers.A survey paper by Beel et al.[30] indicates that more than 55% of the more than 200 articles on research paper recommendation in the last two decades used a content-based filtering approach.

    Therefore, this research applies two types of content-based filtering methods.Firstly, the complete content of both the citing and cited paper is used to categorize the citing paper as an important/nonimportant citation for the cited paper.Furthermore, a novel section-based approach to citation classification is proposed.The results highlight the significance of the proposed approach.

    3 Methodology

    This research proposes identifying important citing papers for a cited paper by using the content of the pair (cited and citing paper).The content-based approach has been successfully applied in the last two decades for relevant research paper recommendations [30,37].Drawing inspiration from this research, this study evaluates two types of content-based comparisons between the citing and cited paper pairs.In the literature, the documents’ entire content has been used to identify relevant papers.However, in this study,we not only adapt the standard content-based approach for the task of important citation identification,but also propose a novel approach termed as section-wise content-based similarity.Fig.1 depicts the complete methodology proposed in this study.In the first step of Fig.1 a benchmark dataset is selected which provides input for both approaches (content-based approach and section-wise content-based approach).The left side of Fig.1 presents methodological steps of the section-wise content-based approach, which produces a similarity score for each section.The right side of Fig.1 depicts the contentbased approach, which produces an overall content similarity score between the two papers.Both approaches produce top-recommended papers.Thus, Fig.1 depicts a state-of-the-art evaluation and comparison strategy.The following sections elaborate on each step of this process in detail.

    3.1 Benchmark Datasets

    The dataset selected to perform the experiments is the benchmark dataset developed by Valenzuela et al.[16].This benchmark dataset is freely available online.The dataset stems from the field of information systems and encompasses 465 annotated paper-citation pairs collected from the Association of Computational and Linguistics (ACL) anthology.The ACL anthology is a digital archive of research papers in computational linguistics and a citation network containing only those papers and citations which are published in the ACL anthology itself.Tab.1 provides a clear description of the dataset.The first column represents the two domain experts who annotated the dataset, denoted “A” and “B” in the Annotator column.The second column contains the source paper ID from the ACL anthology.The third column contains the IDs of the citing papers for the source paper.The fourth column “Follow-up”contains the score assigned by the annotators (i.e., 0 for incidental and 1 for important paper-citation pairs).The dataset also contains Portable Document Format (PDF) files, which were converted into text files to extract the full content and sections of the papers.

    Table 1:Benchmark dataset

    3.2 Content-Based Approach

    The content-based approach is the most dominant method for the task of relevant paper recommendation[30,37].In this study,we propose using the content-based approach to find important citing papers for each cited paper.The implementation steps for the content-based similarity approach are shown on the right side of Fig.1.This study employs Lucene indexing.The Apache Lucene application programming interface(API) is considered the standard software for term indexing.It is widely used by researchers for indexing and finding content similarities [17].For the extraction of important terms, the papers’ full content are provided to the Apache Lucene API.Apache Lucene API indexes all terms within the content.Subsequently, the term frequency-inverse document frequency (TF-IDF) scheme is used to extract important terms from the indexed terms.The term extractor TF-IDF can be mathematically defined as given in Eq.(1).This equation is implemented for all citing and cited papers in the dataset.The basic idea of the TF-IDF technique is elaborated with the following example.For instance, the term T1frequently occurs in document D1, but T1is not found frequently in the other documents D2to Dn.Thus,the conclusion is reached that term T1is the most important term for document D1.Conversely, if any term T2frequently exists in all documents D1to Dn, it means that term T2is not important at all to distinguish documents.

    The next step is to measure the similarity of the papers’content.For this purpose,the cosine similarity technique is used.Eq.(2) shows the mathematical model for cosine vector similarity computation.The important terms extracted from document D1 are presented as vector A and the important terms from document D2 as vector B.

    Cosine similarity was computed for each document compared to all other documents.The generated similarity scores lie between 0 and 1.All text files received a similarity score using the cosine technique.After calculating the cosine similarity scores, the results were sorted in descending order to obtain a ranked list of the top 3(T@3) and top 5(T@5) recommended papers.

    3.3 Section-wise Content Similarity Approach

    The implementation steps for the section-wise content similarity approach are shown on the left side of Fig.1.The Apache Lucene API was again used to index the terms.The similarities between corresponding sections of the papers were identified.To extract the important terms, the content of corresponding paper sections were provided to the Apache Lucene API.Apache Lucene API indexed all terms in the section.Then, the TF-IDF technique was used to identify the most important indexed terms.The term extractor TF-IDF can be mathematically defined as given in Eq.(3).This equation was implemented for all corresponding sections of the citing and cited papers in the dataset.

    In Eq.(3),“t”represents the important terms whereas“s”represents the section content of the cited paper and“S”the content of the corresponding section of the citing paper.The section-wise similarity technique is used to measure the similarity between corresponding sections.The mathematical model for section-wise similarity is given in Eq.(4).The vector V1s refers to the extracted important terms for section ‘s’ of cited paper P1, while the vector V2s refers to the important terms for the corresponding section ‘s’ of citing paper P2.

    The section-wise similarity was computed for each section and was compared to all the corresponding sections.The generated section-wise similarity scores lies between 0 and 1.All text files received a similarity score using the section-wise similarity technique.After obtaining the similarity scores,the results were sorted in descending order to create a ranked list of the top 3 and top 5 recommended papers.

    3.4 Evaluation Parameters

    The proposed approach was evaluated using standard evaluation parameters used by state-of-the-art approaches, namely by (1) Valenzuela et al.[16], (2) Qayyum et al.[26], and Nazir et al.[28].The evaluation parameters used by state-of-the-art approaches are precision, recall, and F-measure.The definition of each parameter is given below:

    The formula to calculate the precision is shown in Eq.(5).The dataset contains citations classified as important and non-important.Precision in identifying important citation using the proposed technique is defined as the ratio of citations correctly classified as “important citations” to the total number of citations classified by the technique as “important citations”.

    The formula to calculate the recall is depicted in Eq.(6).Recall in identifying important citations using the proposed technique is defined as the ratio of citations correctly classified as“important citations”to the total number of citations which are in actual fact “important citations”.

    The F-measure is the harmonic mean of precision and recall and is calculated as shown in Eq.(7).

    The precision,recall,and F-measure for each cited paper were calculated against all of its citing papers by considering the classification presented in Valenzuela et al.[16].Subsequently, the average precision,recall, and F-measure were calculated for the full dataset.Then, the precision, recall, and F-measures for the state-of-the-art approaches were taken from the original published papers [16,26,28], all of which worked on the same dataset.

    4 Results and Comparisons

    This section presents the analysis of results and comparisons for both the proposed approaches.The objectives of this section are twofold.First, we sought to identify the applicability of using content to identify important citations.This refers to both approaches, i.e., using the full content and evaluating individual sections.Secondly, we sought to compare the results with existing state-of-the-art approaches proposed by Qayyum et al.[26], Valenzuela et al.[16], and Nazir et al.[28].Section 4.1 presents the results and evaluation of the two proposed approaches, whereas Section 4.2 compares the best results from the proposed approach with existing state-of-the-art approaches.

    4.1 Results and Evaluation of the Proposed Approaches

    Our first approach was to adapt a content-based filtering technique for citation classification.The second approach was to apply the same content-based technique to individual paper sections,namely the abstract,introduction, literature review, methodology, and results sections.Subsequently, we discuss which section plays the best role in classifying citations into two classes, important and non-important.Specifically, the following six similarity-based rankings were computed.

    1.Full content similarity-based ranking

    2.Abstract-section similarity-based ranking

    3.Introduction-section similarity-based ranking

    4.Literature section similarity-based ranking

    5.Methodology-section similarity-based ranking

    6.Results-section similarity-based ranking

    In this section,the section-based similarity rankings(No.2 to No.6 above)will be compared to the fullcontent-based ranking(No.1 above).

    4.1.1 Evaluation of Proposed Six Rankings

    This section presents the results for all six similarity rankings proposed in this research and listed above.The first ranking is a full-content similarity-based ranking.In this ranking, the full content of the cited document is taken and compared to the full content of the citing documents in the list.Similarity scores are calculated for comparison purposes.Afterwards, the similarity scores are sorted in descending order to rank the top 3 as well as top 5 citing documents for each cited paper.Then precision, recall, and Fmeasure scores are calculated.In the end, a cumulative F-measure for the top 3 and top 5 documents is calculated and compared to the F-measure for the top 3 and top 5 documents identified using the other similarity rankings listed above.The cumulative F-measure for the full-content similarity-based ranking was 0.63 for the top 3 documents and 0.65 for the top 5 documents, respectively.This is a significant result obtained by solely examining the content of research papers.It will be compared with existing state-of-the-art approaches in the next section.

    The second ranking was produced by computing the similarity between the abstracts of the cited and citing documents.The cumulative F-measures for the abstract-section similarity-based ranking were 0.70 for the top 3 documents and 0.69 for the top 5 documents, respectively.Recall that the cumulative F-measures for the full-content similarity-based ranking were 0.63 and 0.65, respectively.Comparisons between the abstract section similarity and full-content similarity are shown in Figs.2a and 2b for the top 3 and top 5 ranked documents, respectively.The cited paper number is listed on the x-axis and the Fmeasure score on the y-axis.The red line shows the F-measure for each cited paper using the full-content approach, whereas the blue line shows the abstract section-based F-measure.For most of the cited papers,the red line and blue line follow the same path, meaning that both approaches produced the same results in these cases.It is also clear from Figs.2a and 2b that when the results of abstract-based similarity and content-based similarity differ,abstract-based similarity produces more accurate results.

    This result clearly shows that abstract-section similarity-based ranking outperforms full-content similarity-based ranking.This is because the abstract is a concisely written paper section of just a few hundred words in which the author has to explain the whole idea of the research paper, including motivation, research gap,state-of-the-art, research question,methodology,results, and comparisons.

    Thus,the abstract has more descriptive power regarding the context and contribution of a research paper.In contrast, the full content of a paper encompasses many different sections, including the introduction,literature review, etc., which might not be feasible to compare and might not deliver such strong results.Research papers’abstracts are normally available for free for both citing papers as well as cited papers.

    The third proposed ranking involves the introduction sections of the cited and citing papers.The cumulative F-measures for the introduction-section similarity-based ranking were 0.57 for the top 3 documents and 0.59 for the top 5 documents, respectively.Recall that the cumulative F-measures for the full-content similarity-based ranking were 0.63 and 0.65, respectively.This result clearly shows that full-content similarity-based ranking outperforms introduction-section similarity-based ranking.The results are shown in Figs.3a and 3b.An in-depth analysis of the terms identified as important for the introduction sections of some of the randomly selected papers illustrates the reason for such results.The introduction section is usually an extended version of the abstract.The general flow of the introduction section is as follows:(1) background of the problem, (2) existing state-of-the-art approaches, (3) research gap, (4) methodology, and (5) results and comparisons.Accordingly, most of the content in papers’introduction sections tends to be very similar, leading towards the citing paper to be considered an important citation for the cited paper.Moreover,this section is typically not very long.

    The fourth ranking was produced by examining the content of the literature review sections of the cited and citing paper pairs.The cumulative F-measures for the literature section similarity-based ranking were 0.59 for the top 3 documents and 0.62 for the top 5 documents, respectively.Recall that the cumulative F-measures for the full-content similarity-based ranking were 0.63 and 0.65, respectively.The results are shown in Figs.4a and 4b.This result clearly shows that full-content similarity-based ranking outperforms literature-section similarity-based ranking.This is because the literature review section contains very generic terms to explain others’ work.Every author has a unique way of writing the literature review section by explaining existing approaches in the respective research areas and critical analyzing the literature.Therefore, the literature review section is not significant for identifying contextual similarities between cited and citing pairs.

    Figure 2:(a) Abstract vs.full content top 3 (b)Abstract vs.full content top 5

    The fifth ranking was achieved by examining the content of the methodology sections of both cited and citing papers.The cumulative F-measures for the methodology-section similarity-based ranking were 0.72 for the top 3 and 0.66 for the top 5 documents, respectively.Recall that the cumulative F-measures for the full-content similarity-based ranking were 0.63 and 0.65, respectively.The results are represented in Figs.5a and 5b.This results clearly show that methodology-section similarity-based ranking outperforms full-content similarity-based ranking.The results demonstrate the expressive power of the cited paper’s methodology section to identify important citations from the list of citing papers.The methodology section presents the study conceptualization of both papers, which involves the use of similar domain-related terms.Nevertheless, while the results for the methodology section were good, the results for the abstract were even better.

    Figure 3:(a) Introduction vs.full content top 3 (b)Introduction vs.full content top 5

    The sixth and final ranking proposed in this research is depicted in Figs.6a and 6b.This ranking was achieved by comparing the content of the results sections of both papers.The cumulative F-measures for the results-section similarity-based ranking were 0.64 for the top 3 and 0.63 for the top 5 documents,respectively.

    Recall that the cumulative F-measures for the full-content similarity-based ranking were 0.63 and 0.65,respectively.This result clearly shows that results-section similarity-based ranking and full-content similarity-based ranking are approximately equal.The results section is also an important section of a research paper, as it reports the important findings of both cited and citing papers.The results section similarity was found to be significant when:(1) both papers address the same topic and use a common vocabulary of terms for the specific domain, (2) both papers use the same dataset, (3) both papers apply similar evaluation metrics,and(4)both papers compare their results with the same/similar research papers.

    This section has reported the results of the six proposed rankings.The full-content-based ranking was adapted from the domain of identifying relevant research papers[30,37].Furthermore,this research proposed the novel approach of section-based similarity.Five further rankings could be calculated when applying this proposed approach.The content of the five major sections of research papers,namely abstract,introduction,literature review, methodology, and results, were systematically compared.The results indicate that examining content alone makes it possible to identify important citations for the cited paper from the list of citing papers.Furthermore, the abstract, methodology, and results sections achieved similar or better results compared to the full content of research papers.Specifically, the abstract section outperformed the full content of research papers.Therefore, we conclude that papers’ abstracts should be used to compute content-based similarity for two reasons:(1) better performance compared to the full-content approach,and (2)abstracts are normally freely available.

    Figure 4:(a) Literature vs.full content top 3 (b)Literature vs.full content top 5

    Figs.2-6 present the results of the five proposed rankings based on comparing individual paper sections.Fig.7 compares the results of six proposed ranking approaches.The complete content achieved the highest precision of 0.68;however,the abstract alone achieved a very close precision score of 0.66.Furthermore,the abstract alone was able to achieve a very high recall value of 0.94,second only to the methodology section.Due to this high recall,the abstract outperformed all other approaches in terms of F-measures.Therefore,out of the five proposed section-wise rankings,the abstract section was selected to be compared to existing stateof-the-art approaches in the following sections.

    4.2 Comparison to State-of-the-Art Approaches

    The previous section proposed six new rankings for identifying important citations for cited papers from the list of citing papers.Comparing abstracts was identified as the best ranking technique based on a critical analysis of the results for all six rankings.This section, in turn, compares the results of the best proposed ranking to the best parameter rankings achieved by current state-of-the-art approaches.Comparisons of precision and recall are depicted in Figs.8 and 9,respectively.

    Figure 5:(a) Methodology vs.full content top 3 (b)Methodology vs.full content top 5

    The proposed approach was compared to the following state-of-the-art approaches.The first approach was presented by Valenzuela et al.[16],who conducted the pioneering work in this area and have made their dataset freely available online.This is the same dataset used by the proposed approach and the other approaches represented in Figs.7 and 8.Valenzuela et al.[16] tested 12 features as identifiers of important citations for the cited papers,with the in-text citation-based feature producing the best results.

    The second state-of-the-art approach was proposed by Qayyum et al.[26].They presented a hybrid approach the uses metadata and content-based features to identify important citations.The third state-ofthe-art approach is the technique was proposed by Nazir et al.[28], who extended the approach by Valenzuela et al.[16].They assigned weights to different sections of the paper to better capture the significance of in-text citation counts.

    Fig.8 compares the precision results for the newly proposed approaches and three existing state-of-theart approaches.The x-axis lists the approaches’ names and the y-axis the precision score.The results for proposed approach #1 (utilizing the full content of the cited-citing pair) is 0.68, and the result of proposed approach #2 (examining content similarity in the abstract sections of the cited-citing pair) is 0.66.Nazir et al.[28] achieved the maximum precision of 0.84, followed by Qayyum and Afzal with a precision score of 0.72.These are the best results from a variety of feature evaluations conducted within each study.The results seem to indicate that the proposed approach outperformed Valenzuela et al.’s [16]approach but was inferior to other state-of-the-art approaches.However, this is not in fact the case.To illustrate why, let us discuss the results of each approach one-by-one.

    Figure 6:(a) Results vs.full content top 3 (b)Results vs.full content top 5

    Figure 7:Comparison of the six proposed rankings

    Figure 8:Comparing the precision of the proposed approaches with state-of-the-art rankings

    Figure 9:Comparing the recall of the proposed approaches with state-of-the-art rankings

    Valenzuela et al.[16]achieved a maximum precision of 0.37 when emplyoing only a single parameter,namely“direct citations per section”.When examining only a single parameter,the newly proposed approach focusing solely on the abstract achieved a precision score of 0.66, thus outperforming Valenzuela et al.In comparison, Valenzuela et al.achieved a precision score of 0.65 when aggregating all 12 parameters, still slightly lower than the precision score of 0.66 obtained by Proposed Approach #2.Furthermore, to compare this value, one needs to consider the following facts.Valenzuela et al.have not discussed how accurately they extracted the 12 features.For example, metadata features like keywords are only available around 50% of the time [26].Furthermore, the accurate extraction of in-text citation counts is not a trivial task and requires very sophisticated algorithms.This has been pointed out by Shahid et al.[11], who achieved 58% accuracy in extracting in-text citations.Although an approach recently proposed by Ahmad et al.[38] raises this accuracy, it still needs to be verified on journals from diverse fields and different publishers’styles.Therefore,the precision score of 0.65 achieved by Valenzuela et al.is dependent on the accurate identification of in-text citation counts.If the approach by Valenzuela et al.[16] were to extract in-text citation counts automatically using the procedures presented by Shahid et al.[11],the precision score might remain in the range of around 0.3.Standard tools such as Content ExtRactor and MINEr(CERMINE)[39] and GeneRation Of BIbliographic Data (GROBID) [40] could only achieve precision, recall, and Fmeasure scores in the range of 0.8 to 0.9 when evaluated by Ahmad et al.[38].Thus, if Valenzuela et al.[16] were to apply the best automated approach to detecting in-text citations, the precision score for finding important citations would drop from 0.65 to less than 0.5.In contrast,the proposed approach does not require any such complex parameter computations; it is based solely on the content of the abstract,which is freely available.Therefore, in terms of real applications, the proposed approach outperforms Valenzuela et al.’s approach in terms of precision score and thus can be considered a viable solution for citation indexes and digital libraries.

    The second state-of-the-art approach was presented by Qayyum et al.[26].They classified citing papers as important/non-important citations for the cited paper using metadata and the papers’ content.The best individual feature they examined achieved a precision score of 0.35.Thus, with respect to single features,the proposed approach utilizing the abstract alone outperforms Qayyum et al.[26].However, when Qayyum et al.[26] aggregated four metadata elements, the precision score using the random forest classifier reached to 0.72.Important to consider here is that this score can only be obtained when all four metadata elements are available.For example, only 58.3% of Qayyum et al.[26] dataset included keywords.This approach is not applicable in the scenarios wherein metadata is not present in equal ratio.Furthermore, cue phrases need to be identified for each individual dataset.This makes the method impractical to use in real systems.In contrast, the newly proposed approach does not rely upon defining a cue phrase dictionary or the availability of keywords.

    The third approach selected for comparison is the technique proposed by Nazir et al.[28].They used section-based in-text citation frequencies to classify citations as important or non-important.A further novel element of this approach is their identification of suitable section weights using linear regression.The approach achieved a precision score of 0.84.However, the present comparison demonstrates the pitfalls of this state-of-the-art approach.Specifically, it is necessary to calculate in-text citation frequencies, which is quite challenging to perform automatically, as noted above.Another challenge concerns mapping section headings onto logical sections (such as Introduction, Literature, Methodology,Results and Discussion).Shahid et al.[29] achieved the highest accuracy for this task which is 78%.Considering all these factors, the proposed approach is comparable to the best-known existing approach as it does not require any complex calculations to be performed unlike other state-of-the-art approches.

    The recall of both proposed approaches and existing state-of-the-art approaches is compared in Fig.8.Proposed Approach #2 (section-wise similarity between abstracts) achieved the best recall of 0.94, higher than existing state-of-the-art approaches.Content-based approaches such as those used by search engines and citation indexes are considered the best approaches to obtain maximum recall.This means that 90%of the time, important citations are identified as such by the proposed approach, with some noise.The proposed approach not only achieves better recall, its implementation is also more viable for the following two reasons:(1) it does not require complex calculations of in-text citation frequencies,mapping section headings to logical sections, the availability of all metadata fields, or identifying cue phrases for each dataset,and (2)abstracts generally available for free online.

    5 Conclusions

    Identifying the set of important citations for the cited paper from the list of citing papers is a challenge that has led the scientific community to propose a wide range of techniques.This research has critically evaluated the literature and identified three state-of-the-art approaches to classifying citations into two classes, namely important and non-important.These existing approaches have utilized a different set of features than the classification method proposed in this study.The precision of these state-of-the-art approaches range from 0.72 to 0.84.However, they are dependent on the accurate identification of some complex features, such as in-text citation frequencies, mapping section headings onto section labels,availability of metadata elements, and constructing dataset-dependent dictionaries of cue phrases.The values for the state-of-the-art approaches cited above are achieved only when all of these parameters are extracted accurately.

    However,a critical analysis shows that the accuracy of identifying in-text citations varies from 58%to 90%, as highlighted by different research and state-of-the-art tools.The accuracy of mapping section headings onto logical sections is just 78%.Keyword metadata is available only 53% of the time.Cue phrases built for one dataset need to be developed anew for another dataset.Currently, state-of-the-art approaches extract such features in a semi-automatic way, and incorrect values are corrected manually.However, when all of these features are extracted fully automatically, the precision score drops to onethird of the reported values.

    This paper presents a method that does not require the computation of such complex features.In the similar domain of identifying relevant research papers, papers’ content has been successfully used for nearly two decades to identify relevant papers.Based on these findings, this paper adopted the contentbased similarity approach to identify the similarity between pairs of cited and citing papers.Furthermore,a novel approach involving section-based similarity was proposed, implemented, and evaluated.An indepth analysis of both proposed approaches indicated that the abstract alone is sufficient to decide whether the citing paper is important or non-important for the cited paper.The proposed approach achieved precision scores of 0.68 for full content and 0.66 for the abstract section, respectively,outperforming existing state-of-the-art approaches when considering the facts presented above.Furthermore, the recall of existing state-of-the-art approaches range from 0.7 to 0.9, while the proposed approach has achieved a recall score of 0.94.Thus, the proposed approach significantly outperformed existing approaches in terms of recall, particularly when considering that inaccurate calculations of in-text citations, section mapping, metadata availability, and cue phrase construction will significantly reduce recall scores for the state-of-the-art approaches when conducted automatically.In contrast, there is no need for such complex calculations in the proposed approach.

    Acknowledgement:We acknowledge the support of Capital University of Science and Technology,Islamabad, Pakistan for providing us with the working environment to complete this research.

    Funding Statement:The author(s) received no specific funding for this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    日本免费一区二区三区高清不卡| 在现免费观看毛片| 哪里可以看免费的av片| 日韩精品有码人妻一区| 啦啦啦观看免费观看视频高清| 精品免费久久久久久久清纯| 亚洲欧洲国产日韩| 老熟妇乱子伦视频在线观看| 国产av麻豆久久久久久久| 国产免费男女视频| 国产极品精品免费视频能看的| 国产伦在线观看视频一区| 国产精华一区二区三区| 深夜a级毛片| 久久欧美精品欧美久久欧美| 国产精品久久久久久亚洲av鲁大| 不卡一级毛片| 真实男女啪啪啪动态图| 亚洲欧洲日产国产| 不卡视频在线观看欧美| 久久人人爽人人片av| 免费搜索国产男女视频| 国产视频内射| 精品99又大又爽又粗少妇毛片| 国产欧美日韩精品一区二区| 一本精品99久久精品77| 精品一区二区三区视频在线| 国产国拍精品亚洲av在线观看| 波多野结衣高清无吗| 日本一本二区三区精品| 女人被狂操c到高潮| av黄色大香蕉| 国产毛片a区久久久久| 只有这里有精品99| 精品久久国产蜜桃| 一区二区三区免费毛片| 一级毛片久久久久久久久女| 日韩欧美一区二区三区在线观看| 亚洲人成网站在线播放欧美日韩| 尤物成人国产欧美一区二区三区| 最近手机中文字幕大全| 99久久无色码亚洲精品果冻| 国产精品美女特级片免费视频播放器| 丝袜喷水一区| 插逼视频在线观看| 国产成人91sexporn| 午夜亚洲福利在线播放| a级毛片a级免费在线| 亚洲中文字幕一区二区三区有码在线看| 国产黄色视频一区二区在线观看 | 国产精品一区二区三区四区免费观看| 美女被艹到高潮喷水动态| 在线播放无遮挡| 大型黄色视频在线免费观看| 哪里可以看免费的av片| 精品欧美国产一区二区三| 亚洲天堂国产精品一区在线| 啦啦啦观看免费观看视频高清| av在线老鸭窝| 在线免费十八禁| 男女视频在线观看网站免费| 在线免费观看的www视频| 亚洲乱码一区二区免费版| 人妻夜夜爽99麻豆av| 村上凉子中文字幕在线| 乱人视频在线观看| 国产精品久久久久久精品电影小说 | 精品人妻偷拍中文字幕| 日韩,欧美,国产一区二区三区 | 国产在视频线在精品| 国产激情偷乱视频一区二区| 不卡视频在线观看欧美| 大香蕉久久网| 日韩一区二区视频免费看| 男插女下体视频免费在线播放| 变态另类丝袜制服| 麻豆国产97在线/欧美| 亚洲综合色惰| 午夜a级毛片| 99久久无色码亚洲精品果冻| 直男gayav资源| 在线观看66精品国产| 欧美区成人在线视频| 欧美日韩一区二区视频在线观看视频在线 | 婷婷六月久久综合丁香| 波多野结衣高清无吗| 人体艺术视频欧美日本| 亚洲中文字幕日韩| 成人亚洲欧美一区二区av| av黄色大香蕉| 中出人妻视频一区二区| 成人二区视频| 中文资源天堂在线| 久久99蜜桃精品久久| 一级毛片久久久久久久久女| 日韩一区二区三区影片| 久久久久九九精品影院| 久久久精品欧美日韩精品| 在线观看一区二区三区| 久久人人精品亚洲av| 熟女电影av网| 亚洲第一区二区三区不卡| 国产三级中文精品| 最近视频中文字幕2019在线8| 久久精品久久久久久久性| 搞女人的毛片| 男人舔奶头视频| 丰满的人妻完整版| 国产精品国产三级国产av玫瑰| 日本色播在线视频| 欧美精品一区二区大全| 白带黄色成豆腐渣| 色5月婷婷丁香| 国产探花极品一区二区| 边亲边吃奶的免费视频| 亚洲内射少妇av| 少妇猛男粗大的猛烈进出视频 | 亚洲欧美精品专区久久| 不卡一级毛片| 伊人久久精品亚洲午夜| 国产伦在线观看视频一区| 在线观看av片永久免费下载| 国产一区二区亚洲精品在线观看| 欧美色视频一区免费| 中文字幕免费在线视频6| 国产色婷婷99| 国产精品久久久久久久久免| 国产精华一区二区三区| 亚洲va在线va天堂va国产| 日日摸夜夜添夜夜爱| 午夜福利在线观看吧| 久久精品久久久久久久性| 日韩视频在线欧美| 人人妻人人看人人澡| 亚洲av中文字字幕乱码综合| 久久久a久久爽久久v久久| 成人综合一区亚洲| 在线播放国产精品三级| 成人综合一区亚洲| 亚洲国产精品合色在线| 九九爱精品视频在线观看| 色5月婷婷丁香| 亚洲激情五月婷婷啪啪| 亚洲av男天堂| 国产精品三级大全| 少妇丰满av| 亚洲精品日韩av片在线观看| 婷婷色综合大香蕉| 国产午夜福利久久久久久| 婷婷精品国产亚洲av| 男的添女的下面高潮视频| 国产一区亚洲一区在线观看| 亚洲人与动物交配视频| 久久久久久久久中文| 日本色播在线视频| 日韩强制内射视频| 能在线免费看毛片的网站| 美女 人体艺术 gogo| 午夜精品在线福利| av视频在线观看入口| 一级二级三级毛片免费看| 国内揄拍国产精品人妻在线| 免费av毛片视频| 一级毛片我不卡| 可以在线观看毛片的网站| 成人三级黄色视频| 一级毛片电影观看 | 波多野结衣巨乳人妻| 国产伦精品一区二区三区四那| 欧美变态另类bdsm刘玥| 亚洲经典国产精华液单| 人人妻人人澡人人爽人人夜夜 | 久久久精品欧美日韩精品| av视频在线观看入口| 观看美女的网站| 国产成人福利小说| 成人特级av手机在线观看| 天美传媒精品一区二区| 国产亚洲av嫩草精品影院| 日韩亚洲欧美综合| a级毛色黄片| 国产综合懂色| 变态另类成人亚洲欧美熟女| 日韩制服骚丝袜av| 亚洲国产欧美人成| 欧美精品国产亚洲| 国产一区二区在线av高清观看| 国产成人福利小说| a级毛片免费高清观看在线播放| 亚洲精品国产成人久久av| 亚洲欧美清纯卡通| 国产亚洲av片在线观看秒播厂 | 成人国产麻豆网| 在线免费观看的www视频| 哪里可以看免费的av片| 99久久中文字幕三级久久日本| 少妇的逼好多水| 午夜福利在线观看吧| 亚洲国产精品久久男人天堂| 国产高潮美女av| 校园人妻丝袜中文字幕| 国产成人a区在线观看| 人人妻人人澡欧美一区二区| 亚洲av熟女| 久久人人爽人人片av| 亚洲欧美日韩高清专用| 夜夜爽天天搞| 欧美色视频一区免费| 乱系列少妇在线播放| 亚洲精品456在线播放app| 特级一级黄色大片| 国产精品,欧美在线| 精品无人区乱码1区二区| 欧美最新免费一区二区三区| 国产视频首页在线观看| 成人亚洲精品av一区二区| 色综合色国产| 欧美日韩在线观看h| 天天躁夜夜躁狠狠久久av| 两个人的视频大全免费| 在线观看av片永久免费下载| 最新中文字幕久久久久| 国产亚洲91精品色在线| 97热精品久久久久久| 国内精品久久久久精免费| 蜜臀久久99精品久久宅男| 国产黄色小视频在线观看| 国产真实乱freesex| 99久国产av精品国产电影| 久久久国产成人精品二区| 日韩视频在线欧美| 欧美日韩国产亚洲二区| 欧美区成人在线视频| av国产免费在线观看| 最近手机中文字幕大全| 特级一级黄色大片| 久久久久久久久大av| 日韩成人av中文字幕在线观看| 亚洲人与动物交配视频| 国产精品伦人一区二区| 一级毛片aaaaaa免费看小| 天堂中文最新版在线下载 | a级一级毛片免费在线观看| 国产精品久久久久久亚洲av鲁大| 国产一区二区在线av高清观看| 亚洲四区av| 给我免费播放毛片高清在线观看| 午夜激情欧美在线| 精品久久久久久久久久免费视频| 乱系列少妇在线播放| 中文字幕免费在线视频6| 亚洲天堂国产精品一区在线| 精品久久久久久久久久免费视频| 黄色欧美视频在线观看| 国内精品一区二区在线观看| 波多野结衣高清无吗| 日韩强制内射视频| 一卡2卡三卡四卡精品乱码亚洲| 欧美三级亚洲精品| 变态另类丝袜制服| 久久人人爽人人片av| 国产成人aa在线观看| 成人二区视频| 国产综合懂色| 夫妻性生交免费视频一级片| av天堂在线播放| 爱豆传媒免费全集在线观看| 午夜亚洲福利在线播放| 成人综合一区亚洲| 亚洲精品亚洲一区二区| 深爱激情五月婷婷| 久久精品国产亚洲网站| 久久久久网色| 欧美丝袜亚洲另类| 国产精品永久免费网站| 欧美人与善性xxx| 美女脱内裤让男人舔精品视频 | 日韩中字成人| 久久综合国产亚洲精品| 成人一区二区视频在线观看| 3wmmmm亚洲av在线观看| 一级毛片我不卡| 国产综合懂色| 国产伦理片在线播放av一区 | 草草在线视频免费看| av在线播放精品| 成人性生交大片免费视频hd| 久久久国产成人精品二区| 天堂av国产一区二区熟女人妻| 我的老师免费观看完整版| av在线播放精品| 又爽又黄a免费视频| 日本黄大片高清| 久久精品国产亚洲av香蕉五月| 欧美激情在线99| 特大巨黑吊av在线直播| 伊人久久精品亚洲午夜| 国产精品福利在线免费观看| 男女做爰动态图高潮gif福利片| 精品免费久久久久久久清纯| 99热这里只有是精品在线观看| 国内精品一区二区在线观看| 神马国产精品三级电影在线观看| 天堂√8在线中文| 日日摸夜夜添夜夜爱| 如何舔出高潮| 在线国产一区二区在线| 在线a可以看的网站| 色吧在线观看| 中国国产av一级| 一进一出抽搐动态| 久久精品国产鲁丝片午夜精品| 两性午夜刺激爽爽歪歪视频在线观看| 又粗又爽又猛毛片免费看| 国产一区二区激情短视频| 少妇熟女欧美另类| 国产在线男女| 国产精品,欧美在线| 亚洲人与动物交配视频| 日韩大尺度精品在线看网址| 内射极品少妇av片p| 亚洲成人av在线免费| 少妇人妻一区二区三区视频| 一级毛片久久久久久久久女| 观看免费一级毛片| avwww免费| 成人性生交大片免费视频hd| 亚洲人成网站在线播放欧美日韩| 亚洲五月天丁香| 欧美激情久久久久久爽电影| 一级毛片久久久久久久久女| 天堂网av新在线| 日本黄色视频三级网站网址| 亚州av有码| 久久韩国三级中文字幕| 秋霞在线观看毛片| 麻豆一二三区av精品| 欧美一区二区精品小视频在线| avwww免费| 一个人观看的视频www高清免费观看| 又黄又爽又刺激的免费视频.| av免费观看日本| 秋霞在线观看毛片| 亚洲精品乱码久久久久久按摩| 亚洲欧美日韩高清专用| 国产成人精品久久久久久| 18禁裸乳无遮挡免费网站照片| av福利片在线观看| av在线播放精品| av在线老鸭窝| 久久精品人妻少妇| 91麻豆精品激情在线观看国产| 夜夜爽天天搞| 女人被狂操c到高潮| 国语自产精品视频在线第100页| 国产真实伦视频高清在线观看| 久久久精品欧美日韩精品| 国产单亲对白刺激| 欧美不卡视频在线免费观看| 久久久精品欧美日韩精品| 看黄色毛片网站| 91久久精品国产一区二区三区| 国产精品野战在线观看| 国产成人福利小说| 久久这里有精品视频免费| 亚洲精品粉嫩美女一区| 老师上课跳d突然被开到最大视频| 成人毛片60女人毛片免费| 色综合站精品国产| 最近手机中文字幕大全| 丝袜喷水一区| 又粗又爽又猛毛片免费看| 观看美女的网站| 免费看av在线观看网站| 一个人看视频在线观看www免费| 99久久精品热视频| 1024手机看黄色片| 久久精品91蜜桃| 最近手机中文字幕大全| 亚洲中文字幕一区二区三区有码在线看| 国产成人影院久久av| 精品不卡国产一区二区三区| 男插女下体视频免费在线播放| 久久精品91蜜桃| 久久午夜亚洲精品久久| 欧美最新免费一区二区三区| 亚洲内射少妇av| 久久精品国产99精品国产亚洲性色| .国产精品久久| 国产一区亚洲一区在线观看| 国产亚洲av嫩草精品影院| 中文字幕久久专区| 国产极品天堂在线| 国产高清有码在线观看视频| 中文在线观看免费www的网站| 91久久精品电影网| 亚洲精品国产成人久久av| av专区在线播放| 九九热线精品视视频播放| 国产大屁股一区二区在线视频| 久久久久久九九精品二区国产| 青春草亚洲视频在线观看| 中文字幕精品亚洲无线码一区| 欧美激情国产日韩精品一区| 男人舔女人下体高潮全视频| 1024手机看黄色片| 大香蕉久久网| 亚洲av一区综合| 搞女人的毛片| 免费搜索国产男女视频| 欧美区成人在线视频| 国产亚洲精品久久久com| 两个人的视频大全免费| 中文字幕精品亚洲无线码一区| 亚洲激情五月婷婷啪啪| 18禁裸乳无遮挡免费网站照片| 天堂av国产一区二区熟女人妻| 婷婷色综合大香蕉| 成人午夜高清在线视频| 99久久人妻综合| 身体一侧抽搐| 日本免费a在线| 亚洲图色成人| 国产乱人视频| 激情 狠狠 欧美| 成人美女网站在线观看视频| 国产精品久久久久久亚洲av鲁大| 午夜激情欧美在线| 亚洲国产精品sss在线观看| 一卡2卡三卡四卡精品乱码亚洲| 色综合亚洲欧美另类图片| 久久久久久久亚洲中文字幕| 久久精品综合一区二区三区| 色5月婷婷丁香| 免费大片18禁| 成人综合一区亚洲| 日日啪夜夜撸| 国产精华一区二区三区| 99精品在免费线老司机午夜| 成人国产麻豆网| 免费观看的影片在线观看| 免费av观看视频| 蜜臀久久99精品久久宅男| 亚洲aⅴ乱码一区二区在线播放| 99精品在免费线老司机午夜| 菩萨蛮人人尽说江南好唐韦庄 | 亚洲五月天丁香| 国产极品精品免费视频能看的| 一进一出抽搐动态| 美女大奶头视频| 欧洲精品卡2卡3卡4卡5卡区| 国产高清视频在线观看网站| 久久久欧美国产精品| 亚洲av一区综合| 国产黄片美女视频| 亚洲欧美日韩高清专用| 国产三级在线视频| 日本三级黄在线观看| 深夜精品福利| 亚洲精华国产精华液的使用体验 | 亚洲中文字幕一区二区三区有码在线看| 国产美女午夜福利| 日韩欧美 国产精品| 99riav亚洲国产免费| 免费在线观看成人毛片| 亚洲国产精品合色在线| 久久韩国三级中文字幕| 亚洲欧美日韩东京热| 欧美日韩乱码在线| 免费一级毛片在线播放高清视频| 在线观看一区二区三区| 免费在线观看成人毛片| 午夜福利在线观看免费完整高清在 | 国产精品免费一区二区三区在线| 秋霞在线观看毛片| av.在线天堂| 校园人妻丝袜中文字幕| 久久久久久久久中文| 久久久久久久久久黄片| 国内久久婷婷六月综合欲色啪| 少妇丰满av| 国产在视频线在精品| 国产精品三级大全| 小蜜桃在线观看免费完整版高清| 国产成人福利小说| 日本在线视频免费播放| 色噜噜av男人的天堂激情| 国产黄色视频一区二区在线观看 | 欧美日本视频| 精品久久久久久久久av| 国产v大片淫在线免费观看| 日本免费一区二区三区高清不卡| 在线观看午夜福利视频| 九草在线视频观看| 偷拍熟女少妇极品色| 日韩成人伦理影院| 免费大片18禁| 国产亚洲精品久久久久久毛片| 狂野欧美白嫩少妇大欣赏| eeuss影院久久| 国产成人a区在线观看| av在线天堂中文字幕| 三级男女做爰猛烈吃奶摸视频| 国产成人影院久久av| 黄色日韩在线| av卡一久久| 少妇高潮的动态图| 日韩成人av中文字幕在线观看| av专区在线播放| 国产精品一及| 亚洲av不卡在线观看| 波多野结衣高清无吗| 少妇人妻精品综合一区二区 | 国产午夜精品久久久久久一区二区三区| 精品一区二区三区视频在线| 色哟哟哟哟哟哟| 六月丁香七月| 夫妻性生交免费视频一级片| 美女cb高潮喷水在线观看| 99国产极品粉嫩在线观看| 国产精品永久免费网站| 淫秽高清视频在线观看| 国产成人freesex在线| 69人妻影院| 不卡视频在线观看欧美| 免费观看人在逋| 欧美日韩国产亚洲二区| av又黄又爽大尺度在线免费看 | а√天堂www在线а√下载| 国产激情偷乱视频一区二区| 男女边吃奶边做爰视频| 久久久精品大字幕| 亚洲av成人av| 男女那种视频在线观看| 亚洲熟妇中文字幕五十中出| 天堂网av新在线| 亚洲av成人av| 黄色欧美视频在线观看| 嫩草影院新地址| 国产老妇女一区| 亚洲四区av| 联通29元200g的流量卡| eeuss影院久久| 国产精品一二三区在线看| 国产精品综合久久久久久久免费| 精品一区二区三区视频在线| 国产高潮美女av| 日本欧美国产在线视频| 日本-黄色视频高清免费观看| 高清毛片免费观看视频网站| 黄色日韩在线| 欧美zozozo另类| 一个人看的www免费观看视频| 日日干狠狠操夜夜爽| 大香蕉久久网| 一边摸一边抽搐一进一小说| 亚洲电影在线观看av| 久久久国产成人精品二区| 亚洲综合色惰| 尾随美女入室| 男人和女人高潮做爰伦理| 3wmmmm亚洲av在线观看| 国产伦精品一区二区三区四那| 性插视频无遮挡在线免费观看| 1000部很黄的大片| 51国产日韩欧美| 91aial.com中文字幕在线观看| 蜜桃亚洲精品一区二区三区| 成人三级黄色视频| 夜夜夜夜夜久久久久| 亚洲国产精品sss在线观看| 嘟嘟电影网在线观看| 国产午夜精品久久久久久一区二区三区| 欧美3d第一页| 自拍偷自拍亚洲精品老妇| 国产日本99.免费观看| 欧美+亚洲+日韩+国产| 人妻系列 视频| 亚洲自拍偷在线| 人妻久久中文字幕网| 国产精品av视频在线免费观看| 天堂网av新在线| 免费人成视频x8x8入口观看| 国产精品久久久久久亚洲av鲁大| 久久人妻av系列| 日日干狠狠操夜夜爽| 免费观看在线日韩| 一进一出抽搐动态| 亚洲精品亚洲一区二区| 色视频www国产| 18禁黄网站禁片免费观看直播| 特大巨黑吊av在线直播| 亚洲欧洲国产日韩| 99久久成人亚洲精品观看| 久久久久久伊人网av| 国产又黄又爽又无遮挡在线| 麻豆国产97在线/欧美| 在线国产一区二区在线| 狂野欧美激情性xxxx在线观看| 欧美日韩国产亚洲二区| 女人十人毛片免费观看3o分钟| 日本成人三级电影网站| 国产精华一区二区三区| 神马国产精品三级电影在线观看| 国产精品麻豆人妻色哟哟久久 | 亚洲美女视频黄频| 观看免费一级毛片| 色哟哟·www| 大又大粗又爽又黄少妇毛片口| 国产精品日韩av在线免费观看| 内射极品少妇av片p| 亚洲婷婷狠狠爱综合网| 变态另类成人亚洲欧美熟女| 毛片女人毛片| 在线观看av片永久免费下载|