• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Development of Social Media Analytics System for Emergency Event Detection and Crisis Management

    2021-12-14 06:04:08ShaheenKhatoonMajedAlshamariAmnaAsifMdMarufHasanSherifAbdouKhaledMostafaElsayedandMohsenRashwan
    Computers Materials&Continua 2021年9期

    Shaheen Khatoon,Majed A.Alshamari,Amna Asif,Md Maruf Hasan,Sherif Abdou,Khaled Mostafa Elsayed and Mohsen Rashwan

    1Department of Information System,College of Computer Science and Information Technology,King Faisal University,Al-Ahsa,31982,Saudi Arabia

    2Faculty of Computers and Artificial Intelligence,Cairo University,Egypt

    3Faculty of Computing&Information,Cairo University,Egypt

    4Faculty of Engineering,Cairo University,Egypt

    Abstract:Social media platforms have proven to be effective for information gathering during emergency events caused by natural or human-made disasters.Emergency response authorities,law enforcement agencies,and the public can use this information to gain situational awareness and improve disaster response.In case of emergencies,rapid responses are needed to address victims’requests for help.The research community has developed many social media platforms and used them effectively for emergency response and coordination in the past.However,most of the present deployments of platforms in crisis management are not automated,and their operational success largely depends on experts who analyze the information manually and coordinate with relevant humanitarian agencies or law enforcement authorities to initiate emergency response operations.The seamless integration of automatically identifying types of urgent needs from millions of posts and delivery of relevant information to the appropriate agency for timely response has become essential.This research project aims to develop a generalized Information Technology(IT)solution for emergency response and disaster management by integrating social media data as its core component.In this paper,we focused on text analysis techniques which can help the emergency response authorities to filter through the sheer amount of information gathered automatically for supporting their relief efforts.More specifically,we applied state-of-the-art Natural Language Processing (NLP),Machine Learning (ML),and Deep Learning(DL)techniques ranging from unsupervised to supervised learning for an in-depth analysis of social media data for the purpose of extracting real-time information on a critical event to facilitate emergency response in a crisis.As a proof of concept,a case study on the COVID-19 pandemic on the data collected from Twitter is presented,providing evidence that the scientific and operational goals have been achieved.

    Keywords:Crisis management;social media analytics;machine learning;natural language processing;deep learning

    1 Introduction

    During the last few decades,the frequency of the occurrence of natural disasters has increased,according to Emergency Event Database [1].These disasters include floods,earthquakes,sandstorms,disease epidemics,and human-made disasters such as industrial accidents and terrorism incidents.According to the same report,Saudi Arabia is prone to three major natural disasters,namely floods,epidemics,and sandstorms,besides industrial hazards and accidents.The kingdom has recorded 14 natural disasters affecting nearly 30,000 people and resulting in an economic loss of approximately $450 million over the past three decades [2].The Presidency of Meteorology and Environment along with the Civil Defense Directorate,have taken the step to advance Disaster Risk Management (DRM).They identified the following three priorities for hazard prevention and mitigation:1) performing risk assessments and updating preparedness strategies 2) strengthening urban resilience and planning and 3) ensuring effective operation of early warning systems.Due to the increased threats of terrorism-related activities and incidents,the Ministry of Interior sponsors several initiatives for strengthening national security [2].Several awareness programs have been launched to educate the public on how to respond during and after natural and human-made disasters and violent extremism.However,the country has yet to develop a comprehensive DRM framework for automatically predicting natural hazards,early warning,and risk assessment systems.

    Recently,the ubiquitous connectivity and proliferation of social networks opened up new opportunities for crisis management through crowdsourcing.One such crowdsourcing tool is Ushahidi [3].It was initially developed to visualize crowdsource reports of violence in Kenya after the post-election violence in 2008.Since then,it is expanding and has become an excellent example of crowdsourcing for raising awareness on different social issues.In modern days,whenever any crisis occurs,the national security agencies,civil defense,emergency and humanitarian professionals,media outlets,and the general public join hands and collaborate using modern technologies such as social media.The potential use of social media had caught the attention of the crisis management research community when close attention was given to social media by Red Cross in its Emergency Social Data Summit held in August 2010 [4].Since then,the research community has started exploring how to monitor,analyze and display data extracted from social media.A recent survey [5]on the role of big data in disaster management shows that social media has been used as the most prominent information source in various disaster management phases,such as long-term risk assessment and reduction,forecasting and prediction,monitoring and detection,early warning,and post-disaster coordination and response.However,extracting meaningful information from such a large,diverse,dynamic,but potentially useful data source is a big challenge which has just begun being to be addressed by the research community.

    In this project,we investigate a cloud computing-based big data framework that will enable us to utilize heterogeneous data sources and sophisticated machine learning techniques to gather and process information intelligently,and provide emergency workers useful insights for making informed decisions,as well as guide the general public on how to stay safe during emergencies.Such a comprehensive framework will help the Kingdom to develop comprehensive Disaster Risk Management (DRM) capability for automatically predicting hazards,early warning,risk assessment,and risk mitigation,including coordination of emergency activities and evacuation.The main thrust is to develop an information system product by dynamically extracting data from diverse social media channels,and storing,managing,analyzing,and interpreting these data in a real-time fashion.Additionally,to disseminate resultants information to decision-makers in a format appropriate for carrying out their tasks.The proposed framework performs the following tasks.First,it dynamically captures multilingual (Arabic,English,and other languages)multimodal (text,audio,video,and images) data from various social media channels in real-time.Second,it applies language-specific models to translate multilingual and multimodal contents into a unified ontology (knowledge-base).Third,it uses machine learning and artificial intelligence techniques for intelligent inference from the knowledge base.Last,it interprets results,present information on an interactive dashboard,and disseminate information to relevant stakeholders.

    As a part of the project,this paper investigates the existing platforms developed to support crisis-related activities and highlights their benefits and shortcomings.An end-to-end disaster support system is proposed to build upon existing platforms on social media use during mass emergencies and disasters.Then,a methodological approach based on the state-of-the-art AI techniques ranging from unsupervised to supervised learning for an in-depth analysis of social media data collected during disasters is described.

    Next,a case study where we apply the proposed methodology on the Twitter data collected during the COVID-19 pandemic for a comprehensive understanding of these real-world devastating crisis events is also described.Specifically,we use topic modeling techniques to understand the different topics discussed in disaster-related posts.To help concerned authorities quickly sift through big crisis data,we employ clustering techniques to group semantically similar messages and find high-level categories.To help humanitarian organizations fulfill their specific information needs,we use deep learning-based classification techniques to classify social media messages into humanitarian categories.

    The rest of the paper is organized as follows:Section 2 investigates the existing platforms developed to support crisis-related activities and highlights their benefits and shortcomings.Section 3 discusses the architecture of the IT solution proposed to develop DRM capability for end-to-end crisis response and management.Section 4 describes the machine learning pipeline for in-depth analysis and mapping social media posts to relevant humanitarian categories for effective response.The proposed machine learning pipeline results are discussed in Section 5 using a COVID-19 case study.Finally,the conclusion and future directions are presented in Section 6.

    2 Related Work

    This section discusses different social media platforms developed to extract crisis-related information from social media to support disaster-related activities.

    The platform “Tweek the Tweet”[6]is a two-way communication channel over Twitter,which connects people and functional authorities to communicate crisis-related information.A series of information processing techniques are used to extract geo-locations,create incident reports on Tweets,and classify them into different categories based on a hashtag.The reports are then published on the Web in various formats,where users can see aggregated information on the type of disaster and places where information has been reported.Tweak the Tweet has been deployed for more than 40 events between 2010 and 2013,including the 2010 Haiti Earthquake,Hurricane Irene in 2011,Colorado wildfires in 2012,and Hurricane Sandy.This tool is useful for humanitarian organizations to sift through evolving events categorized across different hashtags.However,identifying the type of help required by the victims for emergency response requires manual intervention,and there is great potential to apply machine learning to facilitate the work of first responders.

    Likewise,MicroMappers [7]is a crowdsourcing platform that enables people to contribute voluntarily during disasters for humanitarian organizations’effective response.The driving notion is that social media contents may contain information related to a disaster,such as people asking for or offering help,reporting injuries or death,cautioning others to stay away from the danger,and so on.To sort through what is useful and what is not useful,MicroMapper utilizes volunteers to label geotagged contents into four categories:“request for help,” “infrastructure damage,” “not relevant,” and “not English.” The relevant contents are grouped to construct a map sent to agencies providing aid on the ground that need the most up-to-date status of an area.This platform was actively used when Hurricane Matthew hit areas across the Western Atlantic in 2016,collecting over 250,000 images and six million text messages.This tool tries to convert users’posts into actionable information.However,the system’s success largely depends on volunteers,who manually tag the posts into different categories.

    Ushahidi [3]is another open-source crowdsourcing platform developed to monitor violence during a crisis and raise awareness about attackers.It was initially designed to visualize crowdsource reports of violence in Kenya after the post-election violence in 2008.Since then,it is expanding and has become an excellent example of crowdsourcing for raising awareness on different social issues.This tool offers an admin and public interface,where admins create a survey by defining the disaster type with various fields relevant to a particular disaster event and workflows for various tasks needed to be performed or coordinated.Users can submit a survey manually using predefined survey templates by filling in the relevant information.Admin checks and edits the user’s posts as it arrives and performs the necessary tasks before publishing it to the public.For example,for a traffic accident with human injury,the user may report the date,time,location,and injury information.At the same time,the admin will verify the information,notify the authority (traffic police and hospital),and,if necessary,follow up and coordinate with them accordingly.The formal response agencies,humanitarian organizations,and other end users can use the client interface to identify,filter,and organize the overwhelming amount of social media data for any given crisis to make time-critical decisions.Besides,the effectiveness of the tool in crisis monitoring,largely depends on volunteers to manually create and fill in templates with relevant information before making them available to the public.The system has the great potential to integrate AI and Machine learning to annotate such semi-structured data or extract useful pieces of knowledge and information from them to support humanitarian organizations.

    SensePlace2 [8]is another web-based platform which was developed to support situational awareness in crisis management.It was designed to collect crisis-related tweets with place and time information and then visualize them using maps and time plots.Geotags are extracted using geo-location in the tweet,location of users’posting the Tweet,place-based hashtags,and named entity extraction.An interactive visual interface allows users to search and filter tweets by concept,time,and spatial query specification and visualize heat-maps based on geotags.

    Although,the system has a strong focus on situational awareness,which is useful to visualize temporal,spatial,and thematic aspects of a crisis,however,it does not create actionable reports for crisis management and decision making.survey manually using predefined survey.

    Emergency Situation Awareness (ESA1https://esa.csiro.au/aus/index.html) [9]is a web-based system that aims to enhance situational awareness,particularly for earthquakes.It collects,filters,and analyses Tweets from a specific region of interest and presents information in time and space maps.Event detection,text classification,online clustering,and geotagging are applied to detect tweets of interest for emergency managers.The emergency manager can sift through interesting tweets to identify the type of event or its severity.Similarly,Twitcident [10]is another situation awareness tool,which filters crisis-related tweets by semantically enriching the incidents from Twitter and Web.The semantic enrichment module uses named entity recognition and external Web resources to extract attribute-value pairs of relevant tweets.It then classifies tweets based on hand-crafted rules for attribute-value pairs into different categories such as damage,risk,injuries etc.Following that,users can create reports on incident-related messages by querying attribute-value pairs.

    AIDR (Artificial Intelligence for Disaster [11]is a web application designed to filter crisisrelated tweets and utilize crowdsourcing to classify tweets in one of the predefined categories,e.g.,donation,damage,and so on.It then employs a small set of labeled tweets to train the classifier.The system generates an output of messages sorted into categories to create crisis maps and other types of reports.Twitter for Disaster Response-Tweedr [12]applies keyword and regionally based queries to collect crisis-related information from Twitter.It then applies clustering and classification techniques to extract actionable information for disaster relief.Several classification algorithms,such as sLDA,SVM,and logistic regression,are used to recognize tweets reporting losses or damage.However,labeled examples are needed to classify new event in one of the predefined categories.Similarly,CrisisTracker [13]taps into social media to track a set of keywords related to crisis and uses clustering to construct related stories.It engages volunteers to further curate stories by manually tagging them according to the disaster category reported in related stories.Yet,the system’s success depends on volunteers’size and motivations to correctly label each story.

    Most of the systems discussed above are built around the concept of a visual display according to temporal,spatial,and thematic aspects of crisis-related social media data for situational awareness.The visual elements are powered by different computational capabilities,such as extracting relevant information using specific criteria and Natural Language Processing (NLP)techniques,including Named Entity Recognition (NER) and linking entities to concepts.The findings from these platforms suggest that the more situational awareness the formal humanitarian organizations and people have,the better prepared they are to make informed decisions.Some of these platforms focus on creating actionable reports to support disaster response and relief activities.However,the information needed to create actionable reports requires crowdsourcing to manually tag predefined categories,which is neither scalable nor feasible.Furthermore,there is a lack of literature on cohesive information extraction pipeline to automatically extract relevant information,create actionable reports,and deliver information seamlessly to relevant response agencies.This paper proposes a machine learning pipeline that extracts relevant information for disaster relief workers through a cohesive pipeline,consisting of information extraction,clustering,and classification.

    3 Architecture of Social Media Analytics for Emergency Event Detection and Crisis Management

    This research project aims to develop a cloud-based integrated solution for disaster and emergency management using social media analytics.The main thrust is augmenting existing sensor-based Disaster Risk Management (DRM) systems with social media capabilities by keeping the public in the loop (human sensors).The development of such a system will enable relevant disaster management authorities to integrate and access data from several internet-based social data sources and apply semantic analysis to generate actions to be executed in response to presented contents.The generated results will be used by relevant emergency monitoring and disaster management agencies for emergency response,early warning,risk assessment,and risk mitigation,including coordination of emergency activities.

    Fig.1 depicts the overall research framework and highlights its three major components:(1) event (incident) identification and representation using real-time data stream coming from heterogeneous sources such as social networks,(2) automatic reasoning and Inferencing by applying intelligent techniques to derive knowledge and actionable insights from data,(3) incident monitoring and visualization by developing intuitive interfaces to provide emergency professionals with intelligence derived from the knowledge-base.In the following section,the detail of each component is provided.

    3.1 Event Identification and Representation

    The workflow can either be triggered automatically upon detecting an event.Alternatively,it can be initiated manually by an operator on deployment.

    Several parameters need to be specified to start the crawler.The location-based crawler requires a predefined area of interest and time window size for all social media networks configured for crawling.The location will be provided using static location coordinates (i.e.,longitude,latitude) through Google API.An operator needs to specify search-terms or use predefined terms stored in the language database to initiate the search in the keyword search.The crawler will then start searching the posts matching with search terms.The multi-language component’s goal is to provide the capability of crawling the content uploaded in different languages.For achieving that,our system provides a language translation service using Google/Microsoft language translation APIs to translate posts in the target language and store language-specific keywords in the language knowledge-base.Upon setting up these parameters,the application starts receiving data from social media sites,such as Twitter,Facebook,YouTube,a news feed,etc.The content from one or more sources may contain various forms,including text (posts,comments on blog posts,news,etc.) and media uploaded with the post (image and video contents with rich metadata information such as location and text).Following the data collection and translation,the data crawled is transformed into a suitable format to apply preprocessing techniques on text,image,and video,before applying semantic analysis.

    3.2 Automatic Reasoning and Inference

    The data is transformed in a format suitable to apply subsequent analysis and stored in the event database following preprocessing.The automatic reasoning and inference module performs analysis upon the event data,such as semantic analysis,topic extraction,classification,image,and video analysis.As a result,the system identifies themes and cluster together similar posts,topics,conversations,and content.Classification allows the contents to map into predefined categories.However,as social media contents change quickly,it is not practically feasible to define every disaster category.A disaster ontology will be developed to map metadata derived from social data to the matching class in ontology.Hence,ontology creation and alignment are the immediate future work under this research project.At this stage,we assume that ontology will have classes for different disaster types,as well as information about relevant relief organization and the location of those organizations.

    3.3 Incident Monitoring and Visualization

    To effectively monitor and visualize crisis-related social media contents,a web-based interface will be developed with the following functionalities:

    Selected Dissemination of Information (SDI):To effectively share the right information seamlessly with the right people,an SDI component will be developed.Such an efficient filtering system will identify only the right contents to be shared out of a massive volume of information.For example,if the extracted knowledge is about a traffic accident,the traffic police will be notified to summarize the location,the number of injuries and deaths,the services needed among others.Similarly,the information related to medical emergencies will be shared with emergency services,such as ambulance services.While national security incidents will be shared with the National Intelligence,and so on.

    Interactive Visualization:A comprehensive dashboard to visualize incident-related information will be developed.This information will be visualized using maps,time plots,incidents graphs showing a relationship between events,and statistics about the different incidents.

    Query and Navigation Interfaces:Intuitive query and navigation interfaces will be developed to allow users to filter through the detail of incidents and events.The query interface would let users get additional detail on each incident type (such as location,damage type etc.).

    4 Machine Learning Pipeline for Extraction and Classification of Emergency Events

    As a part of the bigger project,described in Section 3,in this paper,we proposed a machine learning pipeline to uncover useful patterns of emergency events using one source of social media data (Twitter) on one type of data (Text).We used several natural language processing techniques to process social media posts.Our machine learning pipeline includes data preparation and textual analysis.Data preparation consists of two steps:(1) data collection and preprocessing(2) feature extraction.After preprocessing the raw dataset and extracting the features,we proceed to the textual analysis stage,including (1) topic modeling,(2) clustering,and (3) classification.As a result,disaster responders can sift through interesting social media data according to a specific humanitarian category.The following sub-sections explain the specific steps,equations,and algorithms in the proposed pipeline.

    4.1 Data Collection and Preprocessing

    The first module is prepared to collect and prepare the data for analysis.We used Twitter’s streaming API using filters in terms of search terms.Twitter data is unstructured and varies significantly in readability,grammar,and sentence structure.The accuracy of any natural language processing techniques largely depends on the informal nature of the language in tweets.The first step in constructing this pipeline is to prepare the dataset in which the noise is reduced to an acceptable level,and relevant tweets should not be omitted at the preprocessing stage.So,we perform preprocessing of tweet text before using them to conduct language analysis.

    We remove URL’s,mentions,hashtags,emojis,smileys,special characters,and stop words as they do not add valuable semantic information for further analysis.The remaining words are tokenized and converted into lowercase to decrease the influence of typos.In the next step,a normalization component is developed to make the tweets more formal.Tweet textual contents are usually in ill-form and full of misspellings,e.g.,earthquak (earthquake),missin (missing),ovrcme(overcome),short words,e.g.,pls (please),srsly (seriously),govt (government),msg (message) and repeated words e.g.soooo depressed (so depressed),okkk (ok).Additionally,people use phonetic substitution,e.g.,2nd (second),2morrow (tomorrow),4ever (forever),4g8 (forget),w8 (wait),and words without spaces,e.g.,prayforIran (pray for Iran),weneedsdonations (we need donations) to stuff more words in given limitation of the Twitter post under 140 charters [14].

    The first module employs a series of regular expressions and linguistic utilities to normalize the text of the tweets.Spelling correction function,for example,utilized scspell (spell checker source code),2https://pypi.org/project/scspell3k/a spelling checking dictionary which allows to match each word in the given text against the English dictionary and replace misspelled and short words with its correct form.We manually build an initial dictionary of 1559 words with abbreviations and short cuts and their correct forms for abbreviations and phonetic variations,e.g.,gbtw,get back to work;2easy,too easy.We used Netlingo3https://www.netlingo.com/acronyms.phpand Webopedia4https://www.webopedia.com/quick_ref/textmessageabbreviations.aspdictionaries available online to prepare a predefined list of ill-form and respective correct-form pair.The normalization function uses the above resources to replace each ill-form word from input text with each of its correct forms.

    After normalization,we found several unrecognized terms which are not available in the English dictionary.For those terms,we used the WordNet5http://wordnetweb.princeton.edu/perl/webwn:NLTK’s English dictionary and the English database WordNetlexicon to look up for words’synset and eliminate those words the synset of which is not found.Finally,WordNet synset is used to extract the synonym of words (e.g.,calamity,mishap,tragedy mapped into an accident) to increase the output significance.We eliminated tweets containing less than three words after removing stop words from the final corpus,as they were deemed them too short for training subsequent language models.

    4.2 Features Extraction

    After basic preprocessing,we created the text corpus for further linguistic analysis by converting preprocessed text corpus into uni-grams and bigram’s bag-of-words vector.Previous studies have found that these two features outperform when used for similar classification tasks [15,16].Finally,we used tf-idf (short for term frequency-inverse document frequency) to calculate the weight for each term to obtain terms’vector [17].In a tf-idf scheme,the basic vocabulary is formed by choosing the number of times a term appears in the document collection,here called term frequency (tf).After length normalization of term frequency,inverse document frequency is counted,for measuring the number of occurrences of a word in the entire corpus.As a result is a document-by-term matrix,is created,each row of which contains the tf-idf values for every document in the corpus.The advantage of tf-idf scheme is that it reduces documents of arbitrary length to fixed-length lists of numbers,and identifies the set of words that are more discriminative for documents in the collection.Since this paper aims to identify topics and find dominant words for each topic,finding some informative but infrequent words is more suitable than treating each word equally as in simple term frequency-based bag-of-words model in which infrequent words might rank low or even left undetected.Term frequency and inverse document frequency is calculated as:

    wheredftis the number of documents containing the termtappears,andNis the number of documents in the corpus.Then,the weight (Wt,d) vector for each termtin documentdis calculated as:

    By calculating all terms in the corpus,we create the vectors for each document.In this study,posts are considered as documents for constructing the terms vectors.We denote each document by a vectorD∈Rv,wherediequals the tf-idf scoreof theith word in the vocabularyvoccurs within the corpus.For example,the word distribution of documentdjis represented as a vectorwj=[wj1,wj2,...,wjv]in avdimensional vector space Rv,where,wiis the weight indicate the importance oftiindj.Termst1,t2,...,tnconstitute a set of features shared across all documents.In other words,each weightwjiindicates how much the termticontributes to the semantic content ofdj.The length of a document is equal to the total number of words/terms in a document:Now corpus is ready for further linguistic analysis explained in the next sections.Our approach includes a number of preprocessing,feature selection,and weighting steps ultimately lead to considerable gains in terms of accuracy on a given text analysis task,such as topic identification and accuracy in classification.

    4.3 Topic Modelling

    The next module in the pipeline facilitates the automatic extraction of thematic structure from the preprocessed textual corpus.This is done with the use of topic modeling,which is the most commonly unsupervised learning technique used for the above-mentioned purpose.We used LDA [18]as a topic model to extract latent topics from a large Tweeter dataset.Unlike supervised learning,which requires prior annotations or labeling,LDA is an unsupervised learning algorithm,which uses probabilistic clustering based on the data itself,allowing us to conduct exploratory analyses on large unstructured textual data without requiring any prior knowledge.

    LDA is a generative probabilistic model which assumes that each topic is a mixture of a set of words,and each document is a mixture of probable topics.LDA has the capability to identify the topic by considering the topics as latent random variables valued from the text document by using a hierarchical Bayesian analysis technique [18].We used LDA to draw topics from a large corpus of text and then predict labels for a new document.Its use enables us to efficiently categorize large bodies of data based on latent topics and predicting the label for incoming new data without any human intervention.Initially,we treat the topic proportions of a document as a draw from LDA Dirichlet distribution.Next,we obtain the words in the document by repeatedly choosing a topic assignment from those proportions,then drawing a word from the corresponding topic.

    LDA’s main weakness problem here is that it returns topics as numbered distributions rather than topic names.To overcome this,Ramage et al.[19]proposed a variation of LDA known as Labelled Latent Dirichlet Allocation (L-LDA) that associates a document with easily-interpretable topics.However,the topics and their corresponding labels must be defined manually.It shall be noted that one semantic topic can span across several topics extracted by LDA due to the various combination of words topics are consisted of;hence manual topic interpretation requires domainspecific knowledge.Moreover,it is tedious and time-consuming.The most generic approach to automatic labeling is drawing top n-words,which are represented in a topic distribution induced by LDA and selecting the most frequent words corresponding to a given topic.However,previous work has shown that top terms are not enough for interpreting the coherent meaning of a topic [20].Furthermore,top terms do not necessarily reveal certain words relevant to the information clustered by a topic.More recent approaches have explored the use of external sources such as Wikipedia and WordNet for deriving candidate labels by means of lexical based algorithms applied to these sources [20,21].However,those topic labeling approaches focus on topics deriving from well-formatted and static documents.On the other hand,the assignment of appropriate labels to topics which derive from tweets presents different challenges since tweets often contain ill-formed words.Moreover,the use of Twitter as the “what’s happening right now”tool introduces new event dependent relations among words that might not have a counterpart in existing knowledge sources (e.g.,Wikipedia).

    In contrast to previous approaches,this paper addresses the labeling of topics exposing event-related content that might not have a counterpart on existing external sources.Based on the observation that semantic similarities of a collection of documents can serve as a label representing that collection,we propose the generation of topic label candidates based on the semantic relevance of a document in the observed corpus.In our automatic label identification approach,the task is to discover a sequence of words for each topicK∈θk,whereθkis a mixture of topics derived from LDA.As topics induced by LDA span over different documents and each document contains a mixture of topicsK,the relevant document for a given topic k can be found in the observed corpus D as follows:Given the set of documentsθdrelevant to topicθk,we can find the prominent document with topic k.For example,corresponding to topic-9,the prominent documents are doc-7,doc-9,doc-20,doc-23 and so on.The word distribution of documentdjis represented as a vectorwj=wj1,wj2,...,wjv,in avdimensional vector space Rv,where,wiis the weight indicate the importance oftiindj.The next step is assigning weight to each document,so as to extract top documents that reflect the salient documents within the candidate documents for the target topic.Computing the weight of a document is fairly straightforward.For instance,the document weight is the sum of the individual terms that appear in a document.Terms can be words,phrases,or other lexical features.To compute terms weight,we compute the tf-idf score for each term in the previous step.One problem with tf-idf is that the document weight can be influenced by its length since longer documents contain more terms than shorter ones.To address this issue,we normalize the documents by dividing them with the normalization factor.Based on this assumption,the document weight is calculated as follows:

    Once each document has been weighted,the documents can be ranked by their weights from which topmdocuments with the most weight can be drawn to choose the sequence of words being used for the topic label.For the topmrelevant document,we draw topic words distribution rank by marginal probability ofwiin topic j as follows:

    wherewnis the ranked list of words for topicjin the relevant document setkd.In each iteration,the above equation is applied for picking the sequence of highest weighted terms out of the highest weighted documents.For each wordwi∈wjn,the most salient words in the target document set are picked as shown below:

    The topxhighest weighted sequence of terms is selected to generate a label.The length of the label could be the top two to four words depending on the desired length of the label.We use an update function that penalizes words that have already been picked.The results of the topic label extraction are assessed and compared with manually labeled data.The accuracy of the topic model is explained in the results section.The summarization method reveals words relevant to the information cluster of a topic that might not appear in top topic keywords merely by extracting top terms of a topic and using them as a topic label.In this way,the labels generated for topics belonging to different categories,generally extend the information provided by the topics’top terms.

    4.4 Topic Induced Clustering

    One of the limitations of topic modeling is that individual documents can be represented with a mixture of latent semantic topics in a document collection.It is essential to classify the posts in terms of events related to different humanitarian categories,which will help the emergency operators to sift through the top-ranked posts under the same category for understanding further the issues reported in a post.These classes should ideally be representative of the underlying data,in the sense that they reflect the problems caused by the disaster event.For this purpose,we use clustering techniques on the potentially relevant topics obtained from the topic model in the previous step.Automatically generated clusters are then manually observed by human experts who assign a category name/label to each cluster.

    We obtained the topic proportion vector of each document from the LDA topic matrix to apply cluster analysis.Previous studies have demonstrated clustering effectiveness induced by topic modeling [15,22].The workflow of topic model-derived clustering is shown in Fig.2.

    In the first method,LDA-derived topics are used as the corpus for cluster analysis,where each document (tweet)diis a vector of k topics in a document-topic matrix (Fig.2a).Each document is assessed for its topic mixture proportion vectorθk.A document is assigned to cluster x based on the maximum probability in the row of the document-topic matrix,i.e.,x=argmaxkθk.

    Figure 2:The workflow of the topic model-derived clustering methods

    In the second method,we applied feature-based cluster analysis.The algorithm starts by extracting features from the LDA-induced topic-word matrix (Fig.2b),where each column is a vector ofnwords (features) corresponding to semantic categories (topics).The words with the highest probability for each topic are selected to create feature vectors (dataset).Following that,the top fifteen words for each topic are selected for clustering analysis.To perform clustering,we first applied the Principal Component Analysis (PCA) [23]to reduce the feature vector’s dimension since the feature vector’s size is equal to the number of unique words in the corpus,which is too large to handle.While using PCA,we fixed a cumulative variance of 50% for the first component.We then applied conventional K-mean clustering [24]on the PCA spanned feature vectors to classify each text in the LDA-induced corpus.We compared the performance of each method and discussed it in the result section.

    4.5 Fine-Tuning BERT Classifier

    Social media users tend to post real-time situational information that could be used by disaster relief and response agencies for effective decision making.However,it is essential to classify the posts into different humanitarian categories for effective processing.After the classification,the dataset would be more informative for assessing the situation in terms of a specific response agency.Several studies have used [25,26]different classifiers to classify disaster-related social media posts using deep learning techniques,such as Convolutional Neural Network (CNN) [27],Gated Recurrent Unit (GRU) [28],and Long Short-term Memory (LSTM) [29].A key element that impairs the performance of these techniques is the input embedding.Existing studies have commonly used pre-trained embedding packages to encode the textual information.However,most of the pre-trained packages are unidirectional with fixed parameters;hence they may not adapt to different disaster categories without tuning them.

    Devlin et al.[30]recently proposed a state-of-the-art BERT model to enhance language representation.The model is designed to pre-train deep bidirectional representations from the unlabeled data by jointly conditioning both left and right contexts in all layers [30].Unlike the unidirectional language model,BERT’s input representation is constructed by summing the corresponding token,segment,and position embeddings.Due to enhanced language representation,we adopted pre-trained BERT (BERTbase-uncased) as the underlying model with 12 transformer layers(12-layer,768-hidden,12-heads,110M parameters) and an output layer with SoftMax to perform the classification task on social media posts (see Fig.3).Each transformer comprises a multi-head self-attention mechanism and a position-wise fully connected feed-forward network,followed by a normalization layer [31].The output of each sub-layer is LayerNorm(x+Sublayer(x)),where Sublayer(x) is the function implemented by the sub-layer itself.

    Figure 3:The architecture of the Fine-tuned BERT-base classifier

    We adopted multiple baseline models to assess the performance of fine-tuned BERT model.The baseline models include CNN with GlobalMaxPooling,Deep Pyramid Convolutional Neural Network (DPCNN),and Recurrent Neural Network (RNN) with GRU units and attention layer and hidden dense layer on top.The performance of the different model is evaluated by calculating accuracies on validation and test dataset:

    We label each post with a humanitarian category using the classifier so that new posts on a given disaster type can be automatically classified with no manual effort.The performance of the BERT-based classifier is discussed in the results section of the case study.

    5 Preliminary Results on the COVID-19 Case Study

    The proposed architecture has been validated on Twitter data collected from recent biological disasters i.e.,COVID-19.We decided to focus on biological disasters and their impact on society provided the world’s current pandemic situation.However,the system can be extended to any other type of natural or human-made disaster.For all experiments,we used the Python programming language.

    5.1 Data Collection and Preprocessing

    We used Twitter Streaming API to collect data using keywords related to covid-19,Following keywords are used to collect data:coronavirus,corona virus,ncov19,ncov,covid,corvid-19,COVID-19,corona virus outbreak,ncov-19,pandemic,Saudi Arabia coronavirus update,Saudi Arabia COVID-19,coronavirus transmission,corona virus news,coronavirus outbreak,coronavirus weapon,coronavirus epidemic,coronavirus pandemic,coronavirus conspiracy,coronavirus news,nCov2019,#coronavirus,nCov,novel corona virus,Wuhan virus,coronavirus bio-weapon,social distancing,lock down,curfew news.

    A total of 150,000 tweets were collected for experimentation purposes.After initial preprocessing and removing duplicates,the dataset size was reduced to 80,000.For each tweet,original post (tweet text),timestamp,and geographic information (coordinates) were obtained.All these fields are considered useful in deriving the patterns and identifying the purpose of the posts.

    5.2 Topic Modeling

    In the next step,the topic modeling algorithm LDA was applied to extract semantic topics from the preprocessed social data.The workflow of topic modeling consists of data processing,model training,parameter tuning,and identifying relevant topics.To select and realize the most efficient workflow,several topic models are trained by orchestrating preprocessing steps and applying different combinations of hyperparameters.For calculating the most appropriate number of topics,we used the genism coherence model.We choose 8 topics returned by LDA for the selected dataset because this number of topics achieved the highest coherence score (0.45).

    We examined the document-term matrix for 8 topics and obtained the top 10 keywords for each topic.Tab.1 presents the results of the identified topics and the most popular words within each topic.

    Unlike the traditional clustering approach,where each data point belongs to one cluster/topic,in topic modeling,it can belong to different topics.For example,a tweet can talk about different symptoms and treatments.For this reason,in Tab.1,it can be observed that the same keywords are appearing in multiple topics.Additionally,semantic categories are grouped under topic numbers such as topics 1 and 2.In order to identify the semantic category,such as a topic referring to death reports or treatment,a domain-specific knowledge must be provided.One method is the manual analysis of the words appearing with high probability in the topic-word distribution while another automatic approach is based on matching keywords.For example,crisis-related social media posts are identified and classified in 23 information-specific categories providing a list of keywords dictionary that could also be used for topic identification [32].

    Table 1:Latent topics and their top 10 keywords

    In this scenario,the keyword-based approach faces difficulties,since social media posts contain anything happening in the world,thus creating an exhaustive list of keywords for every event is not practically feasible.The other approach uses top topic words/terms to assign a semantic label to a topic.However,top topic words do not always reveal words relevant to the information cluster of a topic.Our proposed tf-id based summarization algorithm generates the labels for topics belonging to different categories by extending the top topic words to more topic coherent words in the highest-ranked documents under the target topic.In Tab.2,we compared the top 3 terms sets extracted from LDA for the first two topics (listed under TT),with the label extracted by tf-idf based topic summarization method (listed under tf-idf).We can notice that TT does capture specific information related to an event;yet,it does not provide information regarding the transmission.In this sense,the topic label generated by tf-idf describes the posts related to disease transmission,more accurately.This experiment shows that frequency-based summarization techniques outperform top terms and relevance-based summarization techniques for generating topic labels that improve/can be improved by/upon the top-terms baseline without relying on any external sources.This is an attractive property for automatically generating topic labels for tweets where their event-related content might not have a counterpart on existing external sources.

    Table 2:Comparison of top-term vs. tf-idf based topic label

    To evaluate the accuracy of the automatic labeling algorithm,we manually label 2000 tweets based on human interpretation and compare results to understand whether it is realistic to achieve/generate/assign/create/give comparable human labels.For this purpose,we randomly generate some Tweets to be used as samples in each topic in order to understand the themes of these topics.Our research team of 20 members discussed terms in each topic and grouped them into nine common themes (see Tab.4).

    Tab.3 shows the topic number induced by LDA along with the corresponding label generated by the automatic labeling algorithm,and a possible/comparable human interpretation.We selected the same tweets to be labeled by the automatic labeling algorithm and compared results with human-labeled tweets to assess the accuracy of the proposed approach.

    Table 3:Comparison of top-term vs.tf-idf based predicted labels and comparable human interpretation

    A two-dimensional confusion matrix demonstrates the accuracy of the proposed automatic labeling algorithm by showing the number of actual and predicted labels.The confusion matrix consists of four values:true positives (TP) manually and automatically classified as positive),false negatives (FN) (automatically classified as negative),false positives (FP) only automatically classified as positive and true negatives (TN) both manually and automatically classified as negative(see Fig.4a).The confusion matrix is used to evaluate the proposed topic model’s performance.It can be observed from Fig.4b that the model is capable of predicting the correct labels with an error rate of only 3%-5% with the precision and recall of 95% and 96%,respectively,where precision and recall are calculated as follows:

    Figure 4:Error analysis of the topic model using a confusion matrix.(a) Confusion matrix,(b) Topic modeling evaluation matrix

    5.3 Clustering

    The semantic categories for each document induced by the topic model are distributed over multiple topics.We used clustering (described in Section 4.4) to classify documents into distinct clusters.Each clustering function takes the LDA generated corpus as input and assigns documents to disjoint clusters.The results of the 8 topics are shown in Fig.5.For each such cluster,the system tries to detect and extract only the main semantic class via analyzing all documents in the cluster.Manual analysis by expert annotators shows that cluster ‘0’represents ‘impact,’cluster‘1’represents ‘authorities,’cluster ‘2’represents ‘transmission,’cluster ‘3’represents ‘death reports,’cluster ‘4’represents ‘preventions,’cluster 5 represents ‘sign and symptoms,’cluster ‘6’represents‘treatment’and cluster ‘7’represents other useful information.

    Figure 5:(a) Scatter plots represents the clustering results using the highest probable topic assignment;(b) Scatter plots illustrates the clustering results using feature extractions and k-mean clustering

    At first,we used the document topic-matrix generated by the topic model in the previous step to assign each document to its highest probable topic.The scatter plot of the results on 8 topics shown in Fig.5.It can be observed from Fig.5a that documents on topics 2 and 5 are heavily misclassified under topic 0 (shown with “+”in Fig.5a),and that there is overlap between clusters‘0’and ‘1’,‘0’and ‘7’,‘0’and ‘4’and,‘0’and ‘6’.In the second set of experiments,we applied feature-based cluster analysis using LDA-induced word-topic matrix and applied K-mean clustering to improve classification results (see Fig.5b).From Fig.5b,we can see that misclassification is almost reduced to zero except for a few documents from cluster ‘7’which were misclassified under cluster 0.Also,we can observe more disjoint clusters with a slight overlap between clusters 2 and 3,2 and 1,2,and 7.The observation indicates that feature-based clustering is performed better than the highest probable topic assignment;however,creating coherent categories from tweets can be computationally difficult due to noisy data.In fact,sometimes it is difficult even for experts quite difficult to identify a semantic category accurately.

    5.4 Tweet Classification

    Classifying the social media posts into humanitarian categories is important for capturing the affected areas’needs.We adopted the humanitarian categories and labeled data from Alam et al.[33]in order to for train and validate the/our models.We defined nine categories and manually labeled 2000 tweets (see Tab.4) as described below:

    1) Reports:the tweets are related to different reports,e.g.,reports of deaths due to the disease,confirmed cases,and the number of cases reported.

    2) Signs or symptoms:text contains symptoms such as fever,cough,diarrhea,and shortness of breath or questions related to these symptoms.

    3) Disease transmission:text or questions related to disease transmission.

    4) Disease prevention:questions or suggestions related to the prevention of the disease or refer to a new prevention strategy,e.g.,travel ban,self-isolation,wash hands,social distancing,quarantine.

    5) Disease treatment:questions or suggestions with regard to disease treatments.

    6) Impact:e.g.,unemployment,economic impact,reports about people affected by the disease.

    7) Authorities:text related to government policies,WHO initiatives,etc.

    8) Other useful information:other helpful information that is related to the disease itself.

    9) Not related or irrelevant:text irrelevant to the situation.

    We trained a classifier by fine-tuning a pre-trained BERT transformer [30].To encode input sentences in accordance with BERT input format,we first used BERT tokenizer to split sentences into tokens,add special characters,i.e.,[CLS]and [SEP],map tokens to IDs.,pad,or truncate sentences to the fixed length and create an attention mask.The maximum length of a sequence was defined as 100,with shorter sequences padded and longer sequences truncated to the maximum length.The input sequence was represented by the token,segment,and position embeddings.

    We used the BERTbase-uncasedas the underlying BERT model with 12 transformer layers(12-layer,768-hidden,12-heads,110M parameters) and an output layer with SoftMax to perform the classification task on social media posts.The model was trained using 90% posts in the training set and 10% posts in the test set.Training data were further shuffled into 80% training and 20% validation sets.As shown in Tab.4,the number of labeled tweets for different humanitarian categories are uneven.The imbalanced data may lead to learning bias in our machine learning models.Hence,we employed stratified sampling to mitigate imbalance training examples.We trained the classifier with a batch size of 32 for 5 epochs.The dropout probability was set to 0.2 for all layers.Adam optimizer with a learning rate of 2e-5 and epsilon 1e-8 was used.

    Table 4:Labeled dataset for each category used for training and testing the models

    We used some standard classification models such as GRU,CNN with GlobalMaxPooling and GRU,Deep Pyramid Convolutional Neural Network (DPCNN),Recurrent Neural Network(RNN) with GRU units and attention layer,and hidden dense layer on top.The validation and testing accuracy for the fine-tuned BERT and baseline models are shown in Tab.5.

    Table 5:Labeled dataset for each category used for training and testing the models

    The fined-tuned BERT classifier achieved a validation accuracy of 91.78%,which is much higher than the standard models’accuracy,ranging from 71% to 81%.The fine-tuned BERT model also outperforms other baseline deep learning models in terms of test accuracy.It is worth mentioning that this is a multi-classification problem (nine class);hence the validation accuracy is not very high compared to binary classification tasks.Also,due to the limited labeled dataset(2000 examples),the model would not learn features from the unseen text,hence achieving a lower test accuracy of 76.52%.A rigorous dataset labeling is required to address these problems.However,the classification results are reasonably acceptable for capturing most of the information categories and providing evidence for the capability of a fine-tuned BERT model used for precise classification.

    6 Conclusion and Future Work

    This work presented a machine learning pipeline to automatically map the disaster events across the different humanitarian organizations for supporting their relief efforts.The pipeline integrates topic modeling,clustering,and classification with the capability to detect evolving disaster events and map them across different humanitarian categories using social media data.The proposed pipeline’s application was demonstrated in a case study related to the COVID-19 virus by using Twitter data.The results suggest the following:(1) the integrating enhanced topic summarization method is useful for detecting coherent topics and can be used to predict corresponding labels for real-time situation analysis;(2) In comparison to traditional clustering techniques the topic induced clustering method is more useful for grouping social media posts across different classes;(3) the fine-tuned BERT-based classifier performs better than standard deep-learning classifiers in classifying the tweets into different humanitarian categories.

    The proposed machine learning pipeline offers important directions for future research.First,the proposed pipeline is validated on the COVID-19 case study.To extend the proposed pipeline capabilities to other disaster domains such as flood,traffic industrial accidents,and earthquakes can be further integrated into the pipeline.However,lack of broad humanitarian categories’availability across different disaster domains might impair the process.Second,the integration of other intelligent techniques to uncover more situation awareness such as disaster-affected areas,entities mentioned in the posts,and additional useful information would better support response organizations in their relief work.Third,integration of other data sources such as different social media channels (Facebook,Instagram etc.),news articles,and remote sensing of data would help to gain better awareness for a disaster-affected region.However,integrating heterogeneous data into a unified format is a challenge.Finally,the integration of techniques used to identify social media posts’credibility and information bias is an important area to be addressed.

    Funding Statement:The authors extend their appreciation to the Deputyship for Research &Innovation,Ministry of Education in Saudi Arabia for funding this research work through the Project Number 523.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.The sponsor had no role in the study’s design;collection,analyses,or interpretation of data,in the writing of the manuscript,or in the decision to publish the results.

    80岁老熟妇乱子伦牲交| 国产精品一区二区在线不卡| 一本一本综合久久| 一级爰片在线观看| 精品国产一区二区三区久久久樱花| 亚洲色图 男人天堂 中文字幕 | 青春草亚洲视频在线观看| 好男人视频免费观看在线| 青春草国产在线视频| 久久久精品免费免费高清| 男女免费视频国产| 又粗又硬又长又爽又黄的视频| 成人无遮挡网站| 美女视频免费永久观看网站| 国产一区二区三区综合在线观看 | 日韩强制内射视频| 麻豆乱淫一区二区| 在线精品无人区一区二区三| 婷婷色综合大香蕉| 男男h啪啪无遮挡| 成人二区视频| 人妻一区二区av| 亚洲精品自拍成人| 伦理电影免费视频| 建设人人有责人人尽责人人享有的| 亚洲精品久久成人aⅴ小说 | 王馨瑶露胸无遮挡在线观看| 国产在线一区二区三区精| 大香蕉久久网| 99热国产这里只有精品6| 人人澡人人妻人| a级毛色黄片| 国产精品人妻久久久久久| 人成视频在线观看免费观看| 色哟哟·www| 一本一本综合久久| 欧美丝袜亚洲另类| 国产av一区二区精品久久| 国产在线视频一区二区| 伊人亚洲综合成人网| 亚洲av福利一区| 最近手机中文字幕大全| 青春草视频在线免费观看| 久久人妻熟女aⅴ| 成人国产av品久久久| 亚洲中文av在线| 观看av在线不卡| xxxhd国产人妻xxx| 久久久久久人妻| 少妇高潮的动态图| 女人精品久久久久毛片| 国内精品宾馆在线| 午夜免费观看性视频| 91久久精品电影网| 中国国产av一级| 三级国产精品片| 日韩在线高清观看一区二区三区| 国产一区二区在线观看日韩| 成人无遮挡网站| 国产日韩欧美亚洲二区| 久久久国产一区二区| 能在线免费看毛片的网站| 国产乱人偷精品视频| 久久久久久久久久久免费av| 亚洲精品乱久久久久久| 国产精品99久久99久久久不卡 | 国产视频首页在线观看| 999精品在线视频| 高清视频免费观看一区二区| 日韩中字成人| 人妻人人澡人人爽人人| av专区在线播放| 久久热精品热| 亚洲精品国产色婷婷电影| 欧美精品亚洲一区二区| 日日摸夜夜添夜夜添av毛片| 久久久久精品久久久久真实原创| 亚洲精品国产av成人精品| 亚洲欧美色中文字幕在线| 亚洲欧美一区二区三区黑人 | 国产国语露脸激情在线看| 亚洲成人手机| 乱人伦中国视频| 99视频精品全部免费 在线| 最近最新中文字幕免费大全7| 国语对白做爰xxxⅹ性视频网站| 涩涩av久久男人的天堂| 天天躁夜夜躁狠狠久久av| 激情五月婷婷亚洲| 久久久久久久大尺度免费视频| 岛国毛片在线播放| 亚州av有码| 五月天丁香电影| 午夜免费观看性视频| 日韩av不卡免费在线播放| 久久ye,这里只有精品| 日韩一区二区三区影片| 国产精品99久久久久久久久| 亚洲怡红院男人天堂| 一区二区三区乱码不卡18| 亚洲av成人精品一区久久| 满18在线观看网站| 9色porny在线观看| 久久精品国产a三级三级三级| 99国产综合亚洲精品| 少妇人妻久久综合中文| 亚洲欧美一区二区三区国产| 亚洲人成网站在线观看播放| 视频区图区小说| 日韩欧美精品免费久久| 免费看不卡的av| av播播在线观看一区| 免费观看性生交大片5| 亚洲人成网站在线播| 国产av精品麻豆| 久久久久视频综合| a级毛色黄片| 日韩,欧美,国产一区二区三区| 日韩不卡一区二区三区视频在线| 亚洲国产精品成人久久小说| 国产不卡av网站在线观看| 国产黄频视频在线观看| 91精品伊人久久大香线蕉| av黄色大香蕉| av在线观看视频网站免费| 亚洲,一卡二卡三卡| 99久久人妻综合| 狠狠婷婷综合久久久久久88av| 777米奇影视久久| 熟女人妻精品中文字幕| 国产欧美亚洲国产| 高清av免费在线| 青春草国产在线视频| 国产精品一区二区三区四区免费观看| 亚洲国产精品专区欧美| 亚洲欧美日韩另类电影网站| 看免费成人av毛片| av又黄又爽大尺度在线免费看| 在现免费观看毛片| 不卡视频在线观看欧美| 一本色道久久久久久精品综合| 亚洲色图综合在线观看| 亚洲国产色片| 香蕉精品网在线| 热99久久久久精品小说推荐| 美女大奶头黄色视频| 色婷婷av一区二区三区视频| 亚洲色图综合在线观看| 中文字幕制服av| 亚洲欧美色中文字幕在线| 日本-黄色视频高清免费观看| 69精品国产乱码久久久| 男女高潮啪啪啪动态图| 亚洲欧美清纯卡通| 在线观看www视频免费| 成人综合一区亚洲| av有码第一页| 国产黄色免费在线视频| 亚洲av男天堂| 久久ye,这里只有精品| 在现免费观看毛片| 久久精品国产a三级三级三级| 99re6热这里在线精品视频| 久久99一区二区三区| 欧美日韩在线观看h| 十分钟在线观看高清视频www| 国产精品久久久久久av不卡| 亚洲丝袜综合中文字幕| 亚洲欧美成人综合另类久久久| 国内精品宾馆在线| 精品酒店卫生间| 精品久久蜜臀av无| 亚洲av不卡在线观看| 一级毛片黄色毛片免费观看视频| 久久青草综合色| 午夜福利视频在线观看免费| 黄片无遮挡物在线观看| 欧美性感艳星| 成人二区视频| 夫妻性生交免费视频一级片| 青春草亚洲视频在线观看| 熟妇人妻不卡中文字幕| 国产成人一区二区在线| 国产熟女午夜一区二区三区 | 18+在线观看网站| 一区二区三区精品91| 丝袜美足系列| 欧美+日韩+精品| 日日撸夜夜添| 久热久热在线精品观看| 天天操日日干夜夜撸| 人人妻人人添人人爽欧美一区卜| 久久精品熟女亚洲av麻豆精品| 街头女战士在线观看网站| 精品国产国语对白av| 久热这里只有精品99| 超色免费av| 亚洲精品久久午夜乱码| 99久久精品一区二区三区| 欧美日韩精品成人综合77777| 国产精品久久久久成人av| 91精品一卡2卡3卡4卡| 亚洲欧美成人综合另类久久久| 简卡轻食公司| 最近中文字幕2019免费版| 免费av中文字幕在线| 狠狠精品人妻久久久久久综合| 满18在线观看网站| 久久青草综合色| 欧美97在线视频| av免费观看日本| av不卡在线播放| 麻豆成人av视频| 中国三级夫妇交换| 亚洲精品中文字幕在线视频| 亚洲精品aⅴ在线观看| 99九九线精品视频在线观看视频| 熟女人妻精品中文字幕| 在线观看三级黄色| 国产精品人妻久久久影院| 久久久久久久精品精品| 欧美日韩av久久| 极品少妇高潮喷水抽搐| 日韩熟女老妇一区二区性免费视频| 国产精品国产三级专区第一集| 亚洲国产精品专区欧美| 久久精品久久精品一区二区三区| 中文字幕精品免费在线观看视频 | 一本一本综合久久| 午夜免费观看性视频| 九九在线视频观看精品| 中文欧美无线码| 午夜久久久在线观看| 在现免费观看毛片| 亚洲少妇的诱惑av| 美女脱内裤让男人舔精品视频| 少妇被粗大猛烈的视频| 9色porny在线观看| 晚上一个人看的免费电影| 岛国毛片在线播放| 你懂的网址亚洲精品在线观看| 五月伊人婷婷丁香| 观看av在线不卡| 亚洲怡红院男人天堂| 人妻人人澡人人爽人人| 欧美97在线视频| 91午夜精品亚洲一区二区三区| 日韩强制内射视频| 欧美亚洲日本最大视频资源| 人妻制服诱惑在线中文字幕| 国产伦精品一区二区三区视频9| 精品少妇内射三级| 久久精品国产亚洲av天美| 亚洲人成网站在线播| 亚洲精品一区蜜桃| 午夜免费观看性视频| 国产成人午夜福利电影在线观看| 高清视频免费观看一区二区| 母亲3免费完整高清在线观看 | 天美传媒精品一区二区| 热re99久久国产66热| 免费观看性生交大片5| 日韩免费高清中文字幕av| 综合色丁香网| 91在线精品国自产拍蜜月| 制服人妻中文乱码| 综合色丁香网| 在线观看www视频免费| 国产高清三级在线| 久久久久久久久久人人人人人人| 日本爱情动作片www.在线观看| av专区在线播放| 色婷婷av一区二区三区视频| 欧美成人午夜免费资源| 黄片无遮挡物在线观看| 少妇人妻精品综合一区二区| 天堂中文最新版在线下载| 国产成人免费无遮挡视频| 国产不卡av网站在线观看| 精品久久蜜臀av无| 午夜福利视频在线观看免费| 少妇人妻久久综合中文| 亚洲精品乱码久久久v下载方式| 久久久午夜欧美精品| 欧美 亚洲 国产 日韩一| 国产精品99久久久久久久久| 欧美性感艳星| 欧美激情 高清一区二区三区| 亚洲综合色网址| 亚洲图色成人| 99热国产这里只有精品6| 国产精品一二三区在线看| 天天操日日干夜夜撸| 高清不卡的av网站| 久久精品夜色国产| 国产色婷婷99| 久久毛片免费看一区二区三区| 亚洲久久久国产精品| 久久久久久久久久成人| videosex国产| 肉色欧美久久久久久久蜜桃| 亚洲精品亚洲一区二区| 九色成人免费人妻av| 欧美 亚洲 国产 日韩一| 日本91视频免费播放| 两个人的视频大全免费| 狂野欧美激情性xxxx在线观看| 999精品在线视频| 国产有黄有色有爽视频| 国产精品欧美亚洲77777| 国产一级毛片在线| 人体艺术视频欧美日本| 国产国拍精品亚洲av在线观看| 国产精品一二三区在线看| 成人免费观看视频高清| 亚洲精品乱久久久久久| 一级黄片播放器| .国产精品久久| 国产av国产精品国产| 尾随美女入室| 伦理电影免费视频| 两个人免费观看高清视频| 日韩三级伦理在线观看| 亚洲第一av免费看| 国产男人的电影天堂91| 亚洲三级黄色毛片| 搡女人真爽免费视频火全软件| 亚洲综合精品二区| 国产深夜福利视频在线观看| 国产成人精品福利久久| 亚洲国产精品一区二区三区在线| 岛国毛片在线播放| 久久热精品热| 亚洲精品色激情综合| 成人漫画全彩无遮挡| 免费观看无遮挡的男女| 国产片特级美女逼逼视频| 亚洲国产毛片av蜜桃av| 欧美日韩综合久久久久久| 日本猛色少妇xxxxx猛交久久| 亚洲精品成人av观看孕妇| 亚洲,欧美,日韩| 桃花免费在线播放| 水蜜桃什么品种好| 日韩熟女老妇一区二区性免费视频| 精品午夜福利在线看| 久久精品久久久久久久性| 久久国产精品大桥未久av| 国产亚洲午夜精品一区二区久久| 欧美日韩亚洲高清精品| 天堂中文最新版在线下载| 国产精品女同一区二区软件| 高清在线视频一区二区三区| 亚洲精品乱码久久久久久按摩| 好男人视频免费观看在线| 午夜免费观看性视频| 亚洲天堂av无毛| 黄色毛片三级朝国网站| 亚洲精品乱码久久久v下载方式| 亚洲精品亚洲一区二区| √禁漫天堂资源中文www| 国产精品一二三区在线看| 夫妻性生交免费视频一级片| a级毛片免费高清观看在线播放| 特大巨黑吊av在线直播| 天天操日日干夜夜撸| 久久 成人 亚洲| 日韩一区二区视频免费看| 精品国产乱码久久久久久小说| 久久精品久久精品一区二区三区| 九色亚洲精品在线播放| 久久精品熟女亚洲av麻豆精品| 欧美激情国产日韩精品一区| 国产成人精品久久久久久| 最新中文字幕久久久久| 少妇 在线观看| 夫妻性生交免费视频一级片| 一级二级三级毛片免费看| 久久久a久久爽久久v久久| 国产熟女午夜一区二区三区 | 黑丝袜美女国产一区| 美女xxoo啪啪120秒动态图| 夜夜爽夜夜爽视频| 日韩精品有码人妻一区| 一级黄片播放器| 亚洲欧美精品自产自拍| 国精品久久久久久国模美| 国产精品国产av在线观看| 我的女老师完整版在线观看| 亚洲精品日本国产第一区| 色5月婷婷丁香| 国产精品国产三级国产av玫瑰| 亚洲精品456在线播放app| 永久免费av网站大全| 久久精品久久精品一区二区三区| 男女无遮挡免费网站观看| 精品国产一区二区久久| 欧美3d第一页| 天天操日日干夜夜撸| 人妻人人澡人人爽人人| 黄色配什么色好看| 最近最新中文字幕免费大全7| 国产一区二区在线观看av| 久久 成人 亚洲| 午夜福利视频在线观看免费| a级毛片在线看网站| 美女内射精品一级片tv| 亚洲久久久国产精品| 高清av免费在线| 少妇人妻精品综合一区二区| 91久久精品电影网| 在现免费观看毛片| 一级毛片黄色毛片免费观看视频| 搡老乐熟女国产| 国产淫语在线视频| 18禁裸乳无遮挡动漫免费视频| 极品少妇高潮喷水抽搐| 人人妻人人澡人人看| 美女视频免费永久观看网站| 成人毛片60女人毛片免费| 精品久久久精品久久久| 久久久久久久久久久免费av| 色视频在线一区二区三区| 国产成人精品一,二区| 哪个播放器可以免费观看大片| av女优亚洲男人天堂| 人体艺术视频欧美日本| 日日爽夜夜爽网站| 人妻一区二区av| 一级a做视频免费观看| 丝袜喷水一区| 69精品国产乱码久久久| 国产精品国产三级国产专区5o| 纵有疾风起免费观看全集完整版| 精品国产国语对白av| 国产在线视频一区二区| 国产精品一区二区在线不卡| 精品一品国产午夜福利视频| 各种免费的搞黄视频| 欧美 日韩 精品 国产| 久久99精品国语久久久| 简卡轻食公司| 久久人妻熟女aⅴ| 午夜91福利影院| 久久久国产精品麻豆| 久久人人爽av亚洲精品天堂| 久久久精品免费免费高清| 欧美精品国产亚洲| 亚洲少妇的诱惑av| 街头女战士在线观看网站| 国产精品99久久99久久久不卡 | 看免费成人av毛片| 久久av网站| 国产黄色免费在线视频| 精品国产国语对白av| 精品少妇黑人巨大在线播放| 中国三级夫妇交换| 黑人欧美特级aaaaaa片| 欧美变态另类bdsm刘玥| 国产成人av激情在线播放 | 国产精品秋霞免费鲁丝片| 久久久a久久爽久久v久久| 国产亚洲午夜精品一区二区久久| 国产探花极品一区二区| 成人无遮挡网站| 交换朋友夫妻互换小说| 亚洲精品456在线播放app| 精品一区二区免费观看| 久久婷婷青草| 久久精品久久精品一区二区三区| 国产精品久久久久久久电影| 韩国av在线不卡| 免费观看a级毛片全部| 国产亚洲精品第一综合不卡 | 午夜激情福利司机影院| 久久精品国产自在天天线| 国产亚洲av片在线观看秒播厂| 国产国拍精品亚洲av在线观看| 亚洲国产精品999| 嫩草影院入口| 久久久久久久久久久久大奶| 亚洲精品456在线播放app| 七月丁香在线播放| 久热这里只有精品99| 免费观看性生交大片5| 熟女电影av网| 日韩一区二区三区影片| 日韩制服骚丝袜av| 精品视频人人做人人爽| 欧美日韩视频高清一区二区三区二| 欧美日本中文国产一区发布| 男男h啪啪无遮挡| 人妻夜夜爽99麻豆av| 久久精品久久久久久久性| 亚洲国产av影院在线观看| 久久婷婷青草| 亚洲,一卡二卡三卡| 欧美xxⅹ黑人| 婷婷色av中文字幕| 亚洲精品乱久久久久久| 色5月婷婷丁香| 精品国产国语对白av| 国产精品 国内视频| 超碰97精品在线观看| 99久久中文字幕三级久久日本| 在线观看免费视频网站a站| 最近中文字幕高清免费大全6| 中文字幕人妻丝袜制服| 麻豆乱淫一区二区| 日本vs欧美在线观看视频| 考比视频在线观看| 中文字幕av电影在线播放| 寂寞人妻少妇视频99o| 热re99久久精品国产66热6| 一级毛片 在线播放| 男人添女人高潮全过程视频| 久久久久久久精品精品| 国产免费福利视频在线观看| 丝袜在线中文字幕| 菩萨蛮人人尽说江南好唐韦庄| 男女无遮挡免费网站观看| 国产黄片视频在线免费观看| 永久网站在线| 国产精品久久久久久久久免| 啦啦啦在线观看免费高清www| 亚洲性久久影院| 午夜福利影视在线免费观看| 最新的欧美精品一区二区| 老熟女久久久| 国产高清国产精品国产三级| 国产av精品麻豆| 国产乱来视频区| 日日撸夜夜添| 人妻一区二区av| 美女cb高潮喷水在线观看| 蜜臀久久99精品久久宅男| 国产深夜福利视频在线观看| 国产视频首页在线观看| 99精国产麻豆久久婷婷| 建设人人有责人人尽责人人享有的| 国产成人免费观看mmmm| 免费看不卡的av| 日韩欧美精品免费久久| 国产精品偷伦视频观看了| 精品亚洲成国产av| 欧美日韩视频高清一区二区三区二| 狠狠婷婷综合久久久久久88av| 国产69精品久久久久777片| 免费不卡的大黄色大毛片视频在线观看| 久久av网站| 大香蕉97超碰在线| 国产熟女午夜一区二区三区 | 校园人妻丝袜中文字幕| 高清黄色对白视频在线免费看| 国产精品蜜桃在线观看| 一级a做视频免费观看| 国产免费现黄频在线看| 在线免费观看不下载黄p国产| 亚洲怡红院男人天堂| 97超碰精品成人国产| 一区二区三区免费毛片| 只有这里有精品99| 亚洲av成人精品一区久久| 最近中文字幕高清免费大全6| 我的老师免费观看完整版| 亚洲av不卡在线观看| 国产亚洲午夜精品一区二区久久| 免费av不卡在线播放| 一边摸一边做爽爽视频免费| 国产精品国产三级国产av玫瑰| 精品少妇黑人巨大在线播放| 亚洲国产精品成人久久小说| 搡女人真爽免费视频火全软件| 国产不卡av网站在线观看| 日韩免费高清中文字幕av| 欧美日韩视频精品一区| 男女无遮挡免费网站观看| 嫩草影院入口| 在线观看免费日韩欧美大片 | 在线观看美女被高潮喷水网站| 亚洲av日韩在线播放| 国产精品蜜桃在线观看| 欧美另类一区| 国产有黄有色有爽视频| 91精品国产九色| 免费久久久久久久精品成人欧美视频 | 亚洲不卡免费看| 国产高清国产精品国产三级| 成年美女黄网站色视频大全免费 | 欧美精品国产亚洲| 国产白丝娇喘喷水9色精品| 99热6这里只有精品| 亚洲精品乱久久久久久| 岛国毛片在线播放| 秋霞在线观看毛片| 精品久久蜜臀av无| 日韩熟女老妇一区二区性免费视频| 一边摸一边做爽爽视频免费| av又黄又爽大尺度在线免费看| 黄色一级大片看看| 久久精品夜色国产| 各种免费的搞黄视频| 精品酒店卫生间| 国产乱来视频区| 又粗又硬又长又爽又黄的视频| 国产精品一区二区在线观看99| 在线观看www视频免费| 免费久久久久久久精品成人欧美视频 | 国产不卡av网站在线观看| 国产精品久久久久成人av| 日韩伦理黄色片| 日日摸夜夜添夜夜添av毛片| 男人爽女人下面视频在线观看| 美女国产视频在线观看| 不卡视频在线观看欧美|