• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Attention-Based Bi-LSTM Model for Arabic Depression Classification

    2022-08-24 03:28:52AbdulqaderAlmars
    Computers Materials&Continua 2022年5期

    Abdulqader M.Almars

    College of Computer Science and Engineering,Taibah University,Yanbu,Saudi Arabia

    Abstract: Depression is a common mental health issue that affects a large percentage of people all around the world.Usually, people who suffer from this mood disorder have issues such as low concentration, dementia, mood swings,and even suicide.A social media platform like Twitter allows people to communicate as well as share photos and videos that reflect their moods.Therefore,the analysis of social media content provides insight into individual moods,including depression.Several studies have been conducted on depression detection in English and less in Arabic.The detection of depression from Arabic social media lags behind due the complexity of Arabic language and the lack of resources and techniques available.In this study, we performed a depression analysis on Arabic social media content to understand the feelings of the users.A bidirectional long short-term memory (Bi-LSTM)with an attention mechanism is presented to learn important hidden features for depression detection successfully.The proposed deep learning model combines an attention mechanism with a Bi-LSTM to simultaneously focus on discriminative features and learn significant word weights that contribute highly to depression detection.In order to evaluate our model, we collected a Twitter dataset of approximately 6000 tweets.The data labelling was done by manually classifying tweets as depressed or not depressed.Experimental results showed that the proposed model outperformed state-of-the-art machine learning models in detecting depression.The attention-based Bi-LSTM model achieved 0.83%accuracy on the depression detection task.

    Keywords: Depression detection; social media; deep learning; Bi-LSTM;attention model

    1 Introduction

    Depression is one of the increasingly serious public health issues in modern society.Depression can be defined as a mental illness that causes sadness and loss of interest.In its worst case,depression can sometimes result in suicide [1,2].According to World Health Organization research [3], the number of people affected by depression globally exceeds 350 million.Moreover, the risk of suicide is more than 25 times greater in people who suffer from depression than in the population without these illnesses [2].The risk factors for depression can be divided into three general categories.First, the most common factors that lead to depression are lose of friends or family,financial problems,social isolation,emotional pain,and a loss of hope.These factors can make people feel depressed,sometimes leading to suicidal behaviour[4,5].Second,traumatic experiences,such as cyberbullying,sexual and physical abuse,and academic failure,often lead to depression[6].Third,physical illnesses with little or no hope of cures,such as cancer,brain injury,diabetes,etc.,have been found in many studies to lead to depression[7].

    Social networks such as Twitter and Weibo have grown in popularity for public discussions on social matters.Twitter is a free broadcast channel that offers registered users the ability to discuss and communicate with others via 140-character texts.According to a recent Twitter report[8],a total of 1.3 billion accounts have been created,and there are approximately more than 320 million active users monthly and 140 million users who post over 450 million tweets each day.The popularity of these platforms provides individuals with a free space to talk about their feelings and moods.Fig.1 shows an example taken from Twitter of users who are depressed and are expressing negative intentions,including discussion of suicide,because of cyberbullying and COVID-19.

    Figure 1:Examples of users expressing depression using Twitter

    Several machine-learning techniques,including naive Bayes(NB,support vector machine(SVM)and logistic regression (LR) have been employed in depression identification tasks [9–11].Deeplearning techniques, such as recurrent neural networks (RNNs), convolutional neural networks(CNNs) and long short-term memory (LSTM) can also used for depression detection problems[12–14].Deep learning has already had a significant influence on the field of text categorization when compared to classic machine-learning algorithms.Many studies have shown that deep neural networks outperform more traditional machine learning in terms of accuracy and performance[15,16].

    Social media platforms provide people with a free space to share their emotions and thoughts in written form.This highlights the potential of social media as a rich and consistent source for analyzing posts in Twitter that are particularly relevant for health disorders analysis.Most current studies that focus on analyzing depression in social media do so exclusively in the English language and ignore other languages,such as Arabic.Arabic is the sixth most widely spoken language on the earth.However,depression analysis in the Arabic language is still to be done.Compared with English content,the Arabic language presents more complex challenges due to the use of unstructured nongrammatical slang.In addition,depression is a global issue,and the Arab community is also affected by it;however,there have been few studies on the detection of Arabic depression.

    Hence, in this study, we were interested in filling this gap by introducing a robust model for capturing behavioural health disorders from Arabic content.To achieve this, we present an effective deep learning algorithm to automatically detect depression in social media by combining Bi-LSTM and attention mechanism.The underlying intuition is to learn the discriminative features for depression detection by adding an attention mechanism to Bi-LSTM.Moreover, employing an attention mechanism can help select informative words that are more likely to contribute toward depression detection in social media posts.In this work,we mainly contribute the following.

    ·We introduce a novel deep learning model that learns to perform depression detection automatically from Arabic posts.

    ·The attention mechanism is embedded into Bi-LSTM to enable informative feature extraction from texts.As a result,the deep learning classifier can detect depression better in Arabic content.

    ·Our experiment results indicate that our model outperformed the other models and enhanced the accuracy of the predictions.

    The remainder of the paper includes the following sections:Related work is presented in Section 2,the proposed methodology is introduced in Section 3,experimental results are presented in Section 4,and then we summarize and propose possible future directions in Section 5.

    2 Related Work

    We discuss,In this section,the literature work on depression detection and the attention mechanism briefly.

    2.1 Depression Detection

    The task of predicting depression can be considered a classification problem.Several previous research studies have been concerned with identifying depression.One popular method is to ask people to fill questionnaires or participate in interviews on social media[17,18].For example,Park et al.[18]analyzed the mental disorder of people by exploring the use of languages of users on social media.However,these types of techniques are costly and time-consuming.In the past decade,social media platforms have been widely used for mental disorder analysis.The multimodal depressive dictionary learning model(MDL)for mental health disorder detection on Twitter was introduced by Shen et al.[19].They first applied rule-based heuristic methods to build a well-labelled depression dataset.Then,MDL is used to learn the sparse user representation from each feature group as a single modality.The experimental results demonstrate that the model outperformed the baseline methods in terms of F1-Measure.

    Kim et al.[20] used XGBoost and convolutional neural networks (CNN) to analyze texts from Reddit to determine whether a user has depression, anxiety, or a personality disorder.Mustafa et al.[21]utilize the fourteen psychological attributes in Linguistic Inquiry and Word Count(LIWC)to analyze emotions and identify depression.Based on the weights assigned by LIWC, a machine learning classifier was trained to classify users into three categories of depression.This model was evaluated using a dataset including 179 depressive individuals.A model of public reactions on Twitter was developed by Noor et al.[22]to identify the people’s emotions related to the Covid-19 pandemic.SPM methods are employed to discover frequent words/patterns within tweets, along with their relationships.

    More recently, a hybrid model has been introduced that analyzes user’s textual posts [23].In a hybrid model,Bidirectional Long Short-Term Memory(Bi-LSTM)with various word representation methods and were employed to detect depression, which gave good results.Orabi et al.[24] built three models that use CNN, and the last one uses RNN on the top of the word-embeddings to detect depression of Twitter users.However, the number of tweets used to evaluate the model is relatively small.An early risk detection error(ERDE)metric is proposed by Losada et al.[25].ERDE technology can be applied in a variety of applications,including depression identification.The system provides scores to users who are classified as depressed based on observations from the data.The main drawbacks of this model are that it includes four different meta-parameters that need to be defined before its use, and it needs to observe enough data before making a prediction.Sadeque et al.[26]addressed the drawback of ERDE by proposing a new metric called latency weighted F1 for analyzing social media posts and identifying whether a user was depressed.Other successful models in the text classification task have been proposed that apply different n-gram weighting techniques.Mac Kim et al.[27] utilized term frequency-inverse document frequency (TF-IDF) and post-level embedding for feature representations.The system employed TF-IDF weightings of unigrams to generate feature representations of the labelled dataset, whereas the embeddings were obtained using sent2vec [28].Malmasi et al.[29] utilized lexical features to represent words using different n-grams and syntactic features for deeper linguistic analysis,using the Stanford NLP library.However,these studies applied simple classification models,such as linear SVM classifiers and the SGDClassifier.

    Few studies have been done to detect depression in Arabic content.Almouzini et al.[30]applied several machine learning methods (Random Forest, Na?ve Bayes, AdaBoostM1, and Liblinear) to analyze Arabic tweets and classify them as depressed or not.The results illustrate that the Liblinear algorithm archives the highest accuracy compared with other methods.However,the dataset used in this paper was unbalanced since the dataset contains 27 depressed users and 62 non-depressed users.Recently,Al-Laith et al.[31]introduced a hybrid framework to monitor people’s mental and emotional issues regarding the Covid-19 pandemic.First,the rule-based method was utilized to annotate tweets using an Arabic emotion lexicon.The tweets are then classified into six categories(fear,joy,sorrow,disgust, anger) and two types (symptom and non-symptom tweets) using long-short-term memory(LSTM).

    2.2 Attention Mechanism

    An attention mechanism was developed by Bahdanau et al.[32] to boost the performance of a neural network’s encoder-decoder architecture.The core idea of attention is that it allows the model to put more attention to certain words when processing the data.In other words,the model uses only the relevant information of the input data to predict the output words.Vinyals et al.[33]have extended and improved the previous model by introducing a mechanism that computes attention vectors to reflect how much weight should be assigned to different words to enhance accuracy and performance on large-scale translations.The attention mechanism has been utilized to increase prediction accuracy in classification tasks [34–36].Liu et al.[35] presented a deep-learning model based on attention to handling the issue of high dimensionality and sparsity in text data.Six sentiment datasets and a question dataset are used for experimental verifications.Ran et al.[36] introduced a Short-Term Long Memory(LSTM)with an attention mechanism to detect mental health disorders from interview transcriptions.The evaluation results show that the model performed the best in terms of performance compared with LSTM and SVM.Sharma et al.[37] added a SoftMax function to the hidden state of LSTM to capture more valuable information for visual recognition.However, all mentioned papers applied an attention mechanism for videos and speech depression detection and translation tasks[33,38–40].

    In conclusion, although many applications for text classification have successfully utilized the attention mechanism.This is the first study we know that uses an attention mechanism to analyze textual data related to depression.Moreover,the focus of the current research is on English with fewer contributions in Arabic.Due to the complexity of Arabic,working with Arabic text can be challenging.To overcome this issue,we found that adding an attention mechanism to a deep-learning model can help better analyze Arabic content and detect depression in terms of performance.

    3 Proposed Model

    In this section,we explain in detail the proposed model for depression prediction.Fig.2 illustrates the flow diagram of the proposed model.In the first step,we collected and labelled the data manually.Second,we pre-process and clean the dataset and convert the word strings to a sequence of integers.Third, we then pre-trained our model to build the vector representations of the words.Fourth, we explain the attention mechanism that can be added to the Bi-LSTM to focus on particular features and enhance the accuracy of predictions for the depressed and not depressed binary classification.In the following subsections,we explained the steps in detail.

    Figure 2:The flow diagram of the proposed model

    3.1 Data Collection

    A dataset was collected using the Twitter Developer Application Programming Interface(API).Unfortunately,due to a Twitter policy restriction,we only collected approximately 6000 tweets related to depression written in the Arabic language.The dataset was collected between the 17thof February and the 8thof April.Tab.1 shows the statistical details of the dataset.To balance the dataset, we gathered a list of tweets related to depression using the following hashtags:(1)(depress), (2)(upsets),and(3)(sad).In addition,to balance the ratio of depressed and not depressed,we used the following hashtags:(1)(happy) and (2)(optimistic).After collecting the dataset, extensive cleaning and preprocessing were done to removenoise and outliers.Each tweet was then labelled as one for depressed or zero for not depressed.In the following subsections,we describe the annotation step and the preprocessing step in detail.

    Table 1:Statistical details of dataset

    3.2 Annotation

    Two native PhD students who are experts in human emotions are hired to annotate our dataset.In the beginning, a set of annotation guidelines were given to annotators to have homogeneous annotation results and improve the accuracy of our depression detection dataset.In the first round,the dataset is divided into half, and we asked annotators to manually label the tweets.The data has been labelled{1}as depressed and{0}as non-depressed(normal).Tweets are labelled based on their language,emotional level,suicidal tendencies,and factors related to depression.In other words,users are labelled as depressed if they use specific words related to depression, otherwise non-depressed(normal).In the second round,we asked annotators to revise the annotation process collected from different annotators to have the best level of contingency in results.In case we have different annotation samples,a third annotator is requested to annotate those samples.Finally,the results from different annotators are merged to have a single dataset.Tab.2 shows a piece of the annotation process from different annotators.

    Table 2:Sample of annotation process from different annotators

    3.3 Data Pre-Processing

    Data pre-processing is a critical stage in our model since it translates Arabic text into a format suited for depression detection tasks.In addition,the Twitter text is naturally noisy,as it contains less meaningful“words”for our task,such as hashtags,URLs,replies,etc.Therefore,the main objectives of these steps in our model are to remove noisy and non-useful words and improve the text classification’s prediction.The main pre-processing of Arabic tweets consists of the following steps:

    ·Remove all non-Arabic words.

    ·Remove hashtags,URLs,replies,emojis,digits and punctuation.

    ·Remove non-meaningful stop-words such as(“he”),(“she”),(“in”),and(“to”).

    ·Eliminate repetitive words,for example,replaced with.

    In addition to the previous steps,we also apply a steaming process to the words.Stemming is the step of converting words to their original forms by removing any prefixes,suffixes,and infixes using the Arabic light stemmer[41].After the data preprocessing is done with the Twitter data,we use the Keras tokenizer class to convert a list of string words into sequences of integers.The tokenizer indexes all the words in the text,and each word has a unique index.By completing this step,you can create a dictionary mapping word to an index,which will be needed for the word embedding step.To make the data uniformly distributed,we also apply the same scale on the data by padding with zeros.In other words,each sentence with a sequence length shorter than the mean length is padded with zeros,while longer sentences are reduced to the mean sequence length.

    3.4 Word Embedding

    Word embedding is a popular approach to extracting word associations and relations from a large text corpus in natural language processing.These words are visualized as vectors of real numbers.Such a model can help to find synonymous words or suggest a list of words for a partial sentence.Various models have been introduced for learning word embeddings from raw text, such as GloVe [42] and the Word2Vec model [43].Word2Vec has two model structures for learning feature representations,the continuous bag-of-words(CBOW)and the skip-gram.Pre-trained word embedding is a popular techniques used to improve the accuracy of neural network models.In Arabic content domains,there are a few open-source pre-trained models.In our experiment, we use the pre-trained AraVec [44]to learn distributed word representations.This model uses Word2vec to learn word vectors from sources such as the Twitter platform,Wikipedia articles,and other Internet content.Moreover,AraVec contains a rich vocabulary that can help our model capture more semantic signals from the pre-trained embeddings.The total number of vectors used to build AraVec is approximately 3.300.000.000.

    3.5 Bidirectional Long Short-Term Memory(Bi-LSTM)

    LSTM is a deep-learning model that has been developed and implemented successfully for a variety of tasks such as text categorization, picture processing, and natural language processing.It has shown great performance when compared with other deep-learning methods[45].Unidirectional LSTM only keeps and remembers contextual information of the past, and it neglects the future contextual data [45] In contrast, bidirectional LSTM (Bi-LSTM) [46] addresses this issue by using two independent LSTMs(forward and backward).The forward LSTM processes data in the forward direction (from the past to the future), whereas the backward LSTM processes data in the opposite direction (from the future to the past).Fig.3 illustrates the proposed attention-based deep learning model.In our proposed model, we employ the Bi-LSTM network to learn the discriminative word representations for depression detection.The structure of LSTM consists of(1)forget gateft(2)input gateit(3)output gateotand(4)cell memory statect.Given the input featuresx={x1,x2,x3,...,xt},the hidden state can be computed as follows:

    where isαis a sigmoid function;Wf,Wi,Woare the weighted matrix andbf,bi,boare the biases(hyperparameters)of the LSTM,which will be learned during the training process;andxtis the inputs of the LSTM,andhtis the vector of the hidden layer.The forward hidden stateis computed using the forward LSTM,while the forward hiddenis computed using the backward LSTM.The output of hidden state of the model is then produced by concatenatingIt is important to mention that our problem is a binary classification task.The sigmoid function is used here for the binary depression classification.

    Figure 3:Attention-based Bi-LSTM model

    3.6 Attention Layer

    When dealing with text classification tasks, it is important to understand the relations between words to correctly label the text to either positive“depress”or negative“not-depress”.However,not all words in a sentence are equally significant in the text classification task.To overcome this issue,several types of attention mechanisms have been proposed and applied to various tasks.The idea behind the attention mechanism is that it places greater emphasis on important content by assigning higher weights to specific keywords.This mechanism has shown an excellent performance in several text classification applications.

    In our model, we implemented an attention mechanism similar to the implementation of Bahdanau et al.[32]for analyzing Arabic tweets.By adding an attention layer,we allow our model to decide which words should receive “attention”.In the same sentence, some words are highly related to depression while others are irrelevant.For example, words related to depression such as “sad,depressed, unhappy...etc.”can receive more weight and enables the classifier to perform better in depression detection.Hence,the attention layer can help to focus on specific information.To achieve this, our model first receives the hiddenhtvectors at each time step from the Bi-LSTM model; the model then learns the hidden representationmtand context vectormw.The context vectormwhere represents the weights(importance)of different words over the word sequences.The attention-based model uses a Softmax function to calculate the weighted mean of the stateht.In other words, we assign a single value to each vector of representation,which we call an attention weight.The formula is described as follows:

    whereαtis the attention weight vector,and the weighted hidden representations ares.The outputs are then passed to a fully connected Sigmoid layer which is then used to determines wither the input text is depressed or not.

    4 Experiments

    In this section,we discuss the evaluation of the performance and accuracy of our suggested model using a dataset collected from Twitter.We first describe the experimental settings.Then,our model’s performance was calculated by comparing it to other approaches using precision,recall,F(xiàn)-measure,and accuracy.

    4.1 Experimental Settings

    In our model, there are a number of parameters to tune.Several experiments were conducted to determine the model’s best parameters.The performance of the proposed model in different hyperparameters is illustrated in Tab.3.The following optimal configurations were used.The dropout rate was 0.2; the hidden layer size was 64, and the number of epochs was 20.We chose the Adam optimizer and set the learning rate at 0.001.In our model,the loss function(binary cross-entropy)was used.Early stopping was employed to assess the model’s performance on the dataset.The total number of parameters of the model are shown in Tab.4.To set up our model,we first divide our dataset into 70%for training and 30%for testing.As we are dealing with short texts,we fixed the word embedding size to 50 with a length of 30.

    4.2 Evaluation Metrics

    The performance of our model was measured using four well-known metrics:precision, recall,F(xiàn)-measure,and accuracy.The equations are defined as follows:

    where (TP) is the number of genuine samples accurately labelled as positive.(FP) denotes the total number of false samples that are wrongly labelled as positive.The total number of false samples that are correctly labelled as negative is(TN),whereas the number of real samples that are wrongly labelled as negative is(FN).

    Table 3:Performance of the proposed model in different hyperparameters

    Table 4:The total number of parameters of the model

    4.3 Experimental Results

    Our goal in this section is to evaluate the proposed depression detection model in Section 4 against the following state-of-the-art baselines that utilize textual content for depression detection:

    ·Machine learning models like na?ve bayes,support vector machines(SVM)and Random Forests have been used recently to detect mental health conditions in user posts[9–11,47].

    ·XGBoost[48]is a machine learning technique that has been used recently to analyze texts from social media in an attempt to identify whether a user has a specific mental health disorder.

    ·The LSTM model has been widely explored in previous depression detection studies[31,49].

    ·CNN:Several previous studies have used CNN to detect depression automatically from Twitter[24].We train two CNN models for baselines:CNNWithMax and MultiChannelCNN.

    ·Bi-LSTM model that was used in a recent study for depression classification of text[23,24].

    Fig.4 shows the most frequent words portrayed in each category(positive is“depressed”;negative is“not depressed”),as visualized using the WordCloud module.Positive was defined in terms of the words that were the target of this study not the emotional valence of the words.As we mentioned earlier,our proposed model attempts to identify depression from tweets.Tweets are classified based on their language,emotional level,suicidal tendencies,and factors associated with depression.Users are categorized as being depressed if they use specific words related to depression,otherwise non-depressed(normal).Tab.5 displays the prediction results for our model compared with others.As illustrated in the table,in terms of F-measure,naive Byes performed the lowest when compared with the other models.It had an F-measure of 0.77 in the negative tweets(not-depressed)and 0.74 in the positive ones(depressed).For the SVM and MultiChannelCNN,the results show that they performed better than the Na?ve Byes in depression detection with F-measures of 0.82 and 0.77 for the negative and positive tweets,respectively.Random forest,XGBoost,CNNWithMax,LSTM and Bi-LSTM achieved better results, indicating the ability of those models to detect depressions from social media.However, the model’s performance was enhanced by combining the Bi-LSTM with an attention mechanism in precision,recall and F-measure.When compared to the baseline models,the attention-based LSTM model generated the highest F-measure for positive and negative tweets.As Fig.5 shows,the proposed model obtains the greatest accuracy of 83%compared with the other models.

    Figure 4:WordClouds for most used positive(“depressed”)and negative(“not depressed”)words

    Furthermore, Fig.6 shows the accuracies of our model over different epochs.The accuracy of the proposed model confirmed our assumption that embedding an attention layer with a Bi-LSTM would achieve better results when dealing with Arabic content.To sum up,two main observations can be concluded from the outcomes of our model.First,the deep-learning model outperformed classic machine learning models such as naive Bayes and SVM.Second,adding an attention mechanism to the Bi-LSTM can help it focus on discriminative words and enhance the model’s performance,which indicates that the model can successfully recognize depressed and non-depressed tweets.

    Table 5:Performance of our proposed model compared with others

    Figure 5:The accuracy of the proposed model compared with other

    Figure 6:Accuracy and loss of the proposed model in Arabic dataset

    5 Conclusion

    In this study,we addressed the issue of identifying depression on Twitter.Tackling this problem is essential for understanding the reasons for depression expressed in social media.In this paper,we propose an attention-based Bi-LSTM to analyze and detect depression in Arabic content.The proposed model can successfully analyze and classify tweets as depressed or not depressed by adding an attention mechanism to a Bi-LSTM model.A WordCloud was then used to visualize the most frequently found Arabic words to express depression in our social media dataset.The effectiveness and accuracy of our model were evaluated with the Twitter dataset.The findings of our experiments indicate that our proposed model outperformed the other tested classification approaches.As future work,an exciting and possible direction would be incorporating non-textual features into our proposed model to improve depression detection in social media.Moreover, we anticipate reporting the performance of our model on a large-scale dataset.

    Funding Statement:The author received no specific funding for this study.

    Conflicts of Interest:The author declares that they have no conflicts of interest to report regarding the present study.

    久久热在线av| 一区福利在线观看| 51午夜福利影视在线观看| 制服诱惑二区| 久久热在线av| 国产亚洲精品久久久久久毛片| 久久国产精品男人的天堂亚洲| 99在线人妻在线中文字幕| 亚洲黑人精品在线| 99久久综合精品五月天人人| 国产熟女xx| 少妇粗大呻吟视频| 国产亚洲欧美在线一区二区| 无人区码免费观看不卡| 亚洲av成人一区二区三| 免费看a级黄色片| 亚洲视频免费观看视频| 中文欧美无线码| 色综合欧美亚洲国产小说| 国产熟女xx| 乱人伦中国视频| 国产黄a三级三级三级人| 亚洲人成电影观看| 久久热在线av| 丝袜在线中文字幕| 国产三级黄色录像| x7x7x7水蜜桃| 国产成+人综合+亚洲专区| 好男人电影高清在线观看| 亚洲成人国产一区在线观看| 日韩视频一区二区在线观看| 精品国产乱码久久久久久男人| 久久久精品欧美日韩精品| 99在线人妻在线中文字幕| 国产欧美日韩精品亚洲av| videosex国产| 免费女性裸体啪啪无遮挡网站| av网站免费在线观看视频| 成人黄色视频免费在线看| 女人精品久久久久毛片| 久久伊人香网站| 亚洲色图综合在线观看| 欧美日韩瑟瑟在线播放| 国产99白浆流出| 久久久国产欧美日韩av| 国产免费av片在线观看野外av| 男女床上黄色一级片免费看| 久久中文字幕人妻熟女| 嫁个100分男人电影在线观看| 免费不卡黄色视频| 免费观看精品视频网站| 亚洲av五月六月丁香网| 97碰自拍视频| 琪琪午夜伦伦电影理论片6080| 久9热在线精品视频| 免费在线观看日本一区| 久久欧美精品欧美久久欧美| 欧美日韩乱码在线| 多毛熟女@视频| 老司机在亚洲福利影院| 人人妻人人爽人人添夜夜欢视频| 香蕉国产在线看| 淫妇啪啪啪对白视频| 动漫黄色视频在线观看| 人人妻人人爽人人添夜夜欢视频| 热99re8久久精品国产| 窝窝影院91人妻| 黄色 视频免费看| 亚洲久久久国产精品| 91国产中文字幕| 久久热在线av| 亚洲av成人不卡在线观看播放网| 亚洲av第一区精品v没综合| 91麻豆av在线| 又黄又爽又免费观看的视频| 欧美老熟妇乱子伦牲交| 在线观看免费午夜福利视频| 国产亚洲精品一区二区www| 91成人精品电影| 国内毛片毛片毛片毛片毛片| 日韩欧美在线二视频| 亚洲色图av天堂| 在线观看免费视频网站a站| 欧美在线一区亚洲| 亚洲情色 制服丝袜| 亚洲精品在线美女| 一个人免费在线观看的高清视频| 91麻豆av在线| 在线观看免费视频日本深夜| 欧美日韩中文字幕国产精品一区二区三区 | 婷婷六月久久综合丁香| 久久午夜亚洲精品久久| 咕卡用的链子| 精品一品国产午夜福利视频| 国产成人影院久久av| 黑人欧美特级aaaaaa片| 99久久久亚洲精品蜜臀av| 天堂√8在线中文| 精品熟女少妇八av免费久了| www日本在线高清视频| 亚洲国产看品久久| 如日韩欧美国产精品一区二区三区| 黄片播放在线免费| 国产深夜福利视频在线观看| 在线看a的网站| 午夜久久久在线观看| 在线观看66精品国产| 88av欧美| 国产亚洲欧美98| 欧美 亚洲 国产 日韩一| 亚洲成人精品中文字幕电影 | 91精品国产国语对白视频| 日韩精品青青久久久久久| 99精国产麻豆久久婷婷| 多毛熟女@视频| 丰满饥渴人妻一区二区三| 久久精品影院6| 无人区码免费观看不卡| 国产精品98久久久久久宅男小说| 女生性感内裤真人,穿戴方法视频| 久久久国产成人免费| 午夜日韩欧美国产| 欧美黑人精品巨大| 成人手机av| 亚洲色图综合在线观看| 国产不卡一卡二| 中亚洲国语对白在线视频| 国产精品一区二区在线不卡| 国产精品综合久久久久久久免费 | 满18在线观看网站| 国产av又大| 窝窝影院91人妻| 神马国产精品三级电影在线观看 | 国产精品国产高清国产av| 啦啦啦免费观看视频1| 99国产精品一区二区蜜桃av| 亚洲男人的天堂狠狠| 中文亚洲av片在线观看爽| 天堂√8在线中文| 99香蕉大伊视频| 怎么达到女性高潮| 日本欧美视频一区| 99精品久久久久人妻精品| 在线国产一区二区在线| 精品一区二区三卡| 欧美色视频一区免费| 黄色毛片三级朝国网站| 无遮挡黄片免费观看| 精品福利永久在线观看| 黄色丝袜av网址大全| 黄色 视频免费看| 亚洲av电影在线进入| 成年女人毛片免费观看观看9| 老司机深夜福利视频在线观看| av欧美777| 久久精品成人免费网站| 麻豆一二三区av精品| 免费少妇av软件| 久久精品91无色码中文字幕| 久久久久久久久中文| 国产不卡一卡二| 日韩高清综合在线| 亚洲午夜理论影院| 国产免费av片在线观看野外av| 久久香蕉激情| 欧美乱码精品一区二区三区| 国产精品国产av在线观看| 国产熟女午夜一区二区三区| 免费日韩欧美在线观看| 女生性感内裤真人,穿戴方法视频| 亚洲成a人片在线一区二区| 欧美精品啪啪一区二区三区| 国产精品免费一区二区三区在线| 很黄的视频免费| 女性被躁到高潮视频| 欧美黑人欧美精品刺激| 精品久久久久久电影网| 亚洲中文字幕日韩| 久久久精品欧美日韩精品| av网站在线播放免费| 超碰97精品在线观看| 精品久久久久久久毛片微露脸| 少妇的丰满在线观看| 变态另类成人亚洲欧美熟女 | 欧美成狂野欧美在线观看| 丝袜美足系列| 国产精品国产av在线观看| 别揉我奶头~嗯~啊~动态视频| 亚洲中文日韩欧美视频| 午夜免费观看网址| 亚洲欧美激情综合另类| 亚洲欧美日韩另类电影网站| 亚洲九九香蕉| 欧美久久黑人一区二区| 亚洲第一av免费看| 欧美精品一区二区免费开放| 一本综合久久免费| 久久中文看片网| 国产极品粉嫩免费观看在线| 亚洲精品一卡2卡三卡4卡5卡| 亚洲熟妇熟女久久| 熟女少妇亚洲综合色aaa.| 日韩中文字幕欧美一区二区| 黄频高清免费视频| 一区二区三区国产精品乱码| 久久天堂一区二区三区四区| 欧美丝袜亚洲另类 | 另类亚洲欧美激情| 性少妇av在线| 夜夜看夜夜爽夜夜摸 | 精品国产乱子伦一区二区三区| 亚洲精品在线观看二区| 黄片大片在线免费观看| 亚洲va日本ⅴa欧美va伊人久久| 亚洲,欧美精品.| 日本vs欧美在线观看视频| av免费在线观看网站| 亚洲熟妇熟女久久| 法律面前人人平等表现在哪些方面| 国产精品综合久久久久久久免费 | 久久九九热精品免费| 国产成人精品久久二区二区免费| 久久精品国产清高在天天线| 亚洲成国产人片在线观看| 日本黄色视频三级网站网址| 国产成人精品久久二区二区91| 黑人欧美特级aaaaaa片| 精品人妻在线不人妻| 亚洲一卡2卡3卡4卡5卡精品中文| 精品福利观看| 99国产综合亚洲精品| 丰满人妻熟妇乱又伦精品不卡| 18禁观看日本| 啪啪无遮挡十八禁网站| 久久精品亚洲精品国产色婷小说| 国产午夜精品久久久久久| 自拍欧美九色日韩亚洲蝌蚪91| 999久久久精品免费观看国产| 中文字幕人妻丝袜制服| 国产精品野战在线观看 | 国产精品电影一区二区三区| 久久精品国产亚洲av香蕉五月| 欧美日韩福利视频一区二区| av免费在线观看网站| 亚洲av日韩精品久久久久久密| 国产精品久久电影中文字幕| 亚洲成人免费av在线播放| a在线观看视频网站| 咕卡用的链子| 视频区欧美日本亚洲| 免费av中文字幕在线| 97碰自拍视频| 成人黄色视频免费在线看| 女性被躁到高潮视频| 亚洲精品粉嫩美女一区| 高清在线国产一区| 午夜福利欧美成人| 国产97色在线日韩免费| 999精品在线视频| 大型av网站在线播放| 免费观看精品视频网站| 99热只有精品国产| √禁漫天堂资源中文www| 国产深夜福利视频在线观看| 波多野结衣一区麻豆| 欧美精品啪啪一区二区三区| 免费观看人在逋| 亚洲欧美日韩高清在线视频| 午夜免费观看网址| 高清av免费在线| 午夜视频精品福利| 如日韩欧美国产精品一区二区三区| www日本在线高清视频| 久久久久久大精品| 人妻丰满熟妇av一区二区三区| 在线观看一区二区三区激情| 美女 人体艺术 gogo| 又紧又爽又黄一区二区| 亚洲aⅴ乱码一区二区在线播放 | 欧美精品亚洲一区二区| 夜夜看夜夜爽夜夜摸 | 亚洲人成77777在线视频| 国产欧美日韩综合在线一区二区| 国产精品野战在线观看 | 在线观看免费午夜福利视频| 亚洲国产欧美网| 久久人人精品亚洲av| 国产成人精品无人区| 国产av一区在线观看免费| 国产亚洲av高清不卡| 99国产精品免费福利视频| 久久香蕉激情| 亚洲中文日韩欧美视频| 日本欧美视频一区| 国产精品久久久久久人妻精品电影| √禁漫天堂资源中文www| 日韩中文字幕欧美一区二区| 亚洲精品美女久久av网站| 99久久国产精品久久久| 欧美乱色亚洲激情| 久久亚洲精品不卡| 人成视频在线观看免费观看| 女性被躁到高潮视频| 91成年电影在线观看| 最近最新中文字幕大全电影3 | 在线观看舔阴道视频| 亚洲男人天堂网一区| 校园春色视频在线观看| 日本五十路高清| 久久精品91无色码中文字幕| 一区二区三区国产精品乱码| 18禁国产床啪视频网站| 一级片'在线观看视频| 国产精品亚洲一级av第二区| 男女之事视频高清在线观看| 91在线观看av| 亚洲av电影在线进入| 日韩大尺度精品在线看网址 | 国产精品久久久久久人妻精品电影| 国产成+人综合+亚洲专区| 久久久久久大精品| 一级毛片高清免费大全| 国产黄a三级三级三级人| 久久人人爽av亚洲精品天堂| av视频免费观看在线观看| 99久久综合精品五月天人人| 国产成人系列免费观看| 亚洲免费av在线视频| 亚洲 国产 在线| 制服诱惑二区| 999久久久精品免费观看国产| 日韩高清综合在线| 在线观看免费视频日本深夜| 久久久久九九精品影院| 变态另类成人亚洲欧美熟女 | 午夜久久久在线观看| 亚洲自拍偷在线| 97超级碰碰碰精品色视频在线观看| 老熟妇仑乱视频hdxx| 男人的好看免费观看在线视频 | 99精品久久久久人妻精品| 亚洲中文字幕日韩| 18美女黄网站色大片免费观看| 又黄又爽又免费观看的视频| 高清黄色对白视频在线免费看| 欧美人与性动交α欧美软件| 中文字幕人妻丝袜制服| 国产免费男女视频| 精品免费久久久久久久清纯| 久久精品亚洲精品国产色婷小说| 51午夜福利影视在线观看| 91九色精品人成在线观看| 久久精品影院6| 丰满迷人的少妇在线观看| 少妇被粗大的猛进出69影院| 国产片内射在线| 国产精品香港三级国产av潘金莲| 欧美中文综合在线视频| 日韩精品免费视频一区二区三区| 国产激情久久老熟女| 国产一区在线观看成人免费| 人妻久久中文字幕网| 美女午夜性视频免费| 久久精品国产99精品国产亚洲性色 | 欧美日韩瑟瑟在线播放| 高清黄色对白视频在线免费看| 久久精品亚洲精品国产色婷小说| 亚洲七黄色美女视频| 国产熟女xx| 成在线人永久免费视频| 91在线观看av| 一夜夜www| 中文字幕精品免费在线观看视频| 一夜夜www| 国产精品乱码一区二三区的特点 | 桃色一区二区三区在线观看| 美女大奶头视频| 性欧美人与动物交配| 天天躁狠狠躁夜夜躁狠狠躁| 亚洲欧美精品综合久久99| 波多野结衣av一区二区av| 亚洲成国产人片在线观看| 亚洲成人免费av在线播放| 欧美中文日本在线观看视频| 国产又爽黄色视频| 人妻久久中文字幕网| av在线播放免费不卡| 成人亚洲精品av一区二区 | 精品一区二区三区av网在线观看| 深夜精品福利| 黄色a级毛片大全视频| 亚洲av美国av| 免费观看精品视频网站| 亚洲av日韩精品久久久久久密| 啦啦啦在线免费观看视频4| 国产精品一区二区三区四区久久 | 制服诱惑二区| 国产精品综合久久久久久久免费 | 99久久综合精品五月天人人| 国产精品香港三级国产av潘金莲| svipshipincom国产片| 婷婷六月久久综合丁香| 国产aⅴ精品一区二区三区波| 亚洲精品一二三| 国产av一区在线观看免费| 另类亚洲欧美激情| 热re99久久精品国产66热6| 99riav亚洲国产免费| 黑人巨大精品欧美一区二区mp4| a在线观看视频网站| 亚洲中文字幕日韩| 麻豆国产av国片精品| 悠悠久久av| 精品熟女少妇八av免费久了| 乱人伦中国视频| 操美女的视频在线观看| 精品久久久久久,| 一级片'在线观看视频| 亚洲片人在线观看| 日韩欧美在线二视频| 国产精品国产av在线观看| 9色porny在线观看| 亚洲精品国产精品久久久不卡| 久久国产亚洲av麻豆专区| 91成年电影在线观看| 久久精品亚洲精品国产色婷小说| 亚洲av熟女| 亚洲午夜精品一区,二区,三区| a级毛片在线看网站| 麻豆成人av在线观看| 欧美精品一区二区免费开放| 韩国精品一区二区三区| 99久久精品国产亚洲精品| 亚洲国产精品sss在线观看 | 天天躁夜夜躁狠狠躁躁| 色婷婷av一区二区三区视频| 啦啦啦免费观看视频1| 88av欧美| 黄色毛片三级朝国网站| 免费女性裸体啪啪无遮挡网站| 久久久久国产精品人妻aⅴ院| 麻豆av在线久日| 亚洲中文字幕日韩| 日本wwww免费看| 神马国产精品三级电影在线观看 | av视频免费观看在线观看| 亚洲自拍偷在线| 美女 人体艺术 gogo| 国产av在哪里看| 午夜老司机福利片| 亚洲精品中文字幕在线视频| 午夜福利在线观看吧| 亚洲一区二区三区不卡视频| 国产一区二区三区综合在线观看| 久久草成人影院| 午夜视频精品福利| 亚洲aⅴ乱码一区二区在线播放 | 一二三四在线观看免费中文在| 俄罗斯特黄特色一大片| 日日夜夜操网爽| 男女床上黄色一级片免费看| √禁漫天堂资源中文www| 欧美av亚洲av综合av国产av| 色婷婷av一区二区三区视频| 丝袜在线中文字幕| 1024视频免费在线观看| 91九色精品人成在线观看| 亚洲熟妇熟女久久| 黄片播放在线免费| 男女下面进入的视频免费午夜 | 看免费av毛片| www.自偷自拍.com| av超薄肉色丝袜交足视频| 欧美激情久久久久久爽电影 | 嫁个100分男人电影在线观看| 亚洲国产精品合色在线| 午夜福利在线观看吧| 亚洲精品中文字幕在线视频| 欧美丝袜亚洲另类 | 日韩av在线大香蕉| 国产视频一区二区在线看| 麻豆国产av国片精品| 午夜影院日韩av| 精品日产1卡2卡| 国产精品免费一区二区三区在线| 色播在线永久视频| 国产精品久久久久久人妻精品电影| 精品一区二区三区视频在线观看免费 | 大码成人一级视频| 最好的美女福利视频网| 免费看十八禁软件| 女人精品久久久久毛片| 在线观看免费视频日本深夜| 最近最新免费中文字幕在线| 曰老女人黄片| 亚洲aⅴ乱码一区二区在线播放 | 久久人人97超碰香蕉20202| 久久香蕉激情| 日本免费一区二区三区高清不卡 | 正在播放国产对白刺激| 丰满的人妻完整版| 精品第一国产精品| 天堂动漫精品| 91老司机精品| 午夜福利,免费看| 亚洲欧美日韩高清在线视频| 亚洲av成人av| 亚洲欧美日韩高清在线视频| 国产一区二区三区综合在线观看| 亚洲色图综合在线观看| 日本欧美视频一区| 高清av免费在线| 日韩精品免费视频一区二区三区| 级片在线观看| 精品国产美女av久久久久小说| 成人18禁在线播放| 国产欧美日韩一区二区三| 免费av中文字幕在线| 久久精品国产清高在天天线| 成人影院久久| 日本精品一区二区三区蜜桃| 久久 成人 亚洲| 亚洲中文av在线| 精品高清国产在线一区| 日韩三级视频一区二区三区| 黑人欧美特级aaaaaa片| 国产高清激情床上av| 欧美另类亚洲清纯唯美| 午夜a级毛片| 亚洲男人的天堂狠狠| 男女下面进入的视频免费午夜 | 久久久久亚洲av毛片大全| 亚洲人成伊人成综合网2020| 嫩草影视91久久| 精品国产一区二区久久| 日韩免费av在线播放| 欧美日韩福利视频一区二区| 麻豆久久精品国产亚洲av | 一进一出抽搐gif免费好疼 | 国产精品综合久久久久久久免费 | a级毛片在线看网站| 久久国产精品影院| 久久九九热精品免费| 国产欧美日韩一区二区三区在线| 久99久视频精品免费| 成人国产一区最新在线观看| av片东京热男人的天堂| 免费在线观看视频国产中文字幕亚洲| 亚洲 国产 在线| 日本wwww免费看| 欧美av亚洲av综合av国产av| 两人在一起打扑克的视频| 成人国产一区最新在线观看| 18禁黄网站禁片午夜丰满| www.自偷自拍.com| 中文欧美无线码| 999久久久精品免费观看国产| 久9热在线精品视频| 神马国产精品三级电影在线观看 | 中文字幕色久视频| 国产又爽黄色视频| 精品久久久久久成人av| 一进一出好大好爽视频| 久久影院123| 黄片小视频在线播放| 最新美女视频免费是黄的| 天堂动漫精品| 国产精品电影一区二区三区| 美女高潮到喷水免费观看| 国产熟女xx| 亚洲欧美一区二区三区黑人| a级毛片黄视频| 黄片大片在线免费观看| 久久婷婷成人综合色麻豆| 免费在线观看黄色视频的| 乱人伦中国视频| 亚洲欧美精品综合久久99| 欧美黑人精品巨大| 中文字幕最新亚洲高清| 亚洲人成77777在线视频| 国产精品亚洲一级av第二区| 女人精品久久久久毛片| 亚洲男人天堂网一区| 国产亚洲精品一区二区www| 91国产中文字幕| 国产精品久久久av美女十八| 两个人看的免费小视频| 久久人妻福利社区极品人妻图片| 国产无遮挡羞羞视频在线观看| 狂野欧美激情性xxxx| 日本精品一区二区三区蜜桃| 一本大道久久a久久精品| 久久九九热精品免费| av欧美777| 丝袜美足系列| 欧美不卡视频在线免费观看 | 亚洲国产毛片av蜜桃av| 91老司机精品| 国产成人精品无人区| 好男人电影高清在线观看| 成人18禁高潮啪啪吃奶动态图| 欧美日韩乱码在线| 午夜精品国产一区二区电影| 搡老乐熟女国产| 国产精品av久久久久免费| 成人av一区二区三区在线看| 久久久国产一区二区| 天天添夜夜摸| 超碰97精品在线观看| 夫妻午夜视频| 国产成人精品在线电影| 国产99白浆流出| 亚洲avbb在线观看| 嫩草影院精品99| 国产一区二区三区视频了|