• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Global-Attention-Based Neural Networks for Vision Language Intelligence

    2021-06-18 03:27:30PeiLiuYingjieZhouDezhongPengandDapengWu
    IEEE/CAA Journal of Automatica Sinica 2021年7期

    Pei Liu, Yingjie Zhou,, Dezhong Peng,, and Dapeng Wu,

    I. INTRODUCTION

    IMAGE captioning aims at giving a natural language description for a given image automatically by computer[1], [2], which lies at the intersection of computer vision (CV)and natural language processing (NLP). Practical applications of automatic caption generation include leveraging descriptions for image indexing and retrieval and helping those with visual impairments by transforming visual signals into information that can be communicated via test-to-speech technology. Most of the current approaches employed the encoder-decoder framework, in which encoder usually consists of convolutional neural networks (CNNs), in order to obtain a fixed length of feature vectors from input image as visual representation; while the decoder adopts recurrent neural networks (RNNs) [1], [3]–[6] or attention-based neural network (such like Transformer) [7]–[9], [10], [11] to decode image visual features into flexible length captions.

    Recently, many captioning approaches tend to take region proposal features as their input features inspired by the progress made in the CV [12]–[14]. Region proposal features refer to a set of visual features extracted for each region proposal through an object detection model. Compared to a fixed single vector, the object features are stored separately and integrally, therefore this kind of representation method could provide more accurate and complete visual information.However, there are some challenges when taking region proposal features as inputs in the captioning task, because the intrinsic gap naturally exists between the tasks of object detection and image captioning, and further cause a result that the region proposal features obtained from the former task may not serve well for the latter directly. Specifically, only a small portion of region proposals are essential for captioning at each step of word generation, which means that there are many noises and distractions to mislead the decoder, and it is difficult to let the decoder keep attention on these correct areas when “l(fā)ooking for” the needed visual information in the process of caption generation.

    In order to address the above-discussed challenges, many works adopted various attention mechanisms [15] and more complicated sequence generation networks. The attention mechanism was introduced in captioning tasks to leverage the above problem by helping the decoder only pay attention to most related areas. Meanwhile, with the recent successes in NLP [7], [16], [17], the self-attention mechanism is powered by assigning different weights to previously got words, and the Transformer proposed in [7] has the ability to summarize the “past” (previously obtained words) much more accurately.Most of captioning approaches only try to “ look inside”,including already got words, state of the model, etc., in order to extract the needed visual information from region proposal features, but ignores the external latent contribution (i.e., the degree of the inherent importance of region proposal for captioning task) that actually inherently existed in the image and could affect the result of caption generation.

    To overcome this gap, our motivation aims at exploring the latent contributions for region proposals and utilize these clues in the decoder to help visual feature extract. The latent contribution could provide external guidance for the language module so that the decoder could better attend on these critical proposals and less chance to be misled. This inherent characteristic hints that these areas with high latent contribution would have a higher chance to be attended and mentioned, while the chance of the rests are relatively much lower. An example is given in Fig. 1 . The areas in blue bounding boxes, i.e., tagged “fireplace”, “stocking” and“chair”, are easier attended by us to take as caption contents,while the areas in orange bounding boxes, i.e., tagged“flower”, “clock”, “l(fā)amp” and “door” might be ignored,which is also verified in the ground-truth captions of this image in the MS-COCO dataset. This phenomenon illustrates the situation that region proposals have inherent different degree impacts for image captioning. We human beings could get these latent contributions of region proposals at the first glimpse of the input image.

    Fig. 1. An example shows that local regions with different probabilities being mentioned in caption text, in which blue boxes are easier being mentioned than blue boxes.

    In our work, we proposed a novel global-attention-based neural networks (GANN) to incorporate the latent contributions in decoder. First, we extract such a global caption feature by taking region proposal features as input in encoder, which could used for predicating the caption objects(i.e., the objects mentioned or will be mentioned in captions)directly; after that, the final weights in attention module of each decoder layer is obtained by combining latent contributions (global weights) and local weights, in which the former reflect the importance of each region proposal“globally” by taking feature caption feature as query vector,while the latter reflects the importance of each region proposal“l(fā)ocally” by taking hidden state vector as query vector at each specific time step of word generation.

    The main contributions are summarized as follows:

    1) We proposed a global-attention-based neural network for image captioning, in which we could incorporate the latent contributions of region proposals into the attention module of decoder so that the decoder can better attend on correct and related region proposals in the process of caption generation.

    2) We proposed a novel method for global caption feature extraction, which is treated as a multi-class multi-label classification problem and jointly trained with the caption model.

    3) We perform comprehensive evaluations on image captioning dataset MS-COCO, demonstrating that the proposed method outperforms several current state-of-the-art approaches in most metrics, and the proposed GANN could improve previous approaches.

    The rest of the paper is organized as follows: Section II introduces related works; Section III describes the global caption feature extraction and presents the GANN;implementation details and experiments are described in Section IV; finally, Section V concludes the paper.

    II. RELATED WORKS

    1) RNN vs. Transformer:Most of previous captioning approaches are built based on recurrent neural networks(RNNs) or Transform neural networks [7]. a) For the first category, approaches usually take a fixed length feature vector extracted from pre-trained convolutional neural networks(CNNs). For example, the works [6], [18] first employ RNNs and long short term memory (LSTM) as their language model in image captioning respectively, and the work [19] consider the training images as the references and proposed a reference based long short term memory (R-LSTM) model. b) For the second category, Transformer based model could achieve better performance in many sequence-to-sequence tasks through the multi-head self-attention mechanism, like reported in many machine translation works [20], [21]. For example,the work [22] built its model based on transformer to explore the intra-modal interactions, and the work [23] tries to incorporate the geometric clues for image captioning. The results from these recent works illustrate that transformer based model could give more accurate captions.

    2) Attention Mechanism:The attention mechanism [7], [15],[24] now is often adopted to help the decoder to extract needed visual information. We can understand these attention modules in captioning approaches from two aspects, i.e., the query vector, and multiple attentions fusion method. a) for the query vector, most of models [4], [25]–[27] adopt the hidden state as the query vector in attention module at each step of caption generation for given image or video, the works [11]propose a two layers attention approach, in which the top layer takes the output of bottom layer as query vector to generate the information vector and attention gate. b) for the fusion method of multiple attention, the work [28] compute the temporal attention on frames and spatial attention on image regions both using hidden state as query vector, and fuse these two kinds of attention in an adaptive way, and the work [23] fuse the position attention and self-attention in encoder in order to get geometry information involved. In our work, multiple attention weights are computed by taking hidden output of previous layer and global caption vector as query vector respectively, and fuse these two kinds of attention for involving the latent contributions in each step of visual information extraction.

    Fig. 2. An example shows that local regions with different probabilities being mentioned in caption text, in which blue boxes are easier being mentioned than blue boxes.

    3) Semantic Clues:Another approach [3], [29]–[31]considers incorporating semantic clues. The usage and extraction method for semantic clues vary in there approaches.a) for semantic clue extraction, the work [29], [32] tries to extract scene vector independently with caption model, while the work [33] and our work extract the semantic clues simultaneously along with the captioning model. b) for the usage of semantic clues, the work [33] uses a semantic embedding network layer to accept semantic clues as input and feed its output for decoder, the work [29] utilize the tag vector as semantic clues to guide the ensemble of parameters in language model, however, in our work, we take the semantic clues to calculate the global attention weights, which is hoped to help the encoder focus on the correct areas as knowledge priors.

    III. GLOBAL-ATTENTION-BASED NEURAL NETWORK

    In this section, the implement details and intuition of our proposed global-attention-based neural network (GANN) will be described and discussed. Fig. 2 shows the structure of our proposed approach. In the encoding phrase, region proposal features are extracted from the input image through the pretrained Faster-RCNN model, then feed into the encoder as well as an inserted empty vector (same role of “sos” embedding vector in BERT). After that, the encoder outputs the encoded features and the global caption feature, which are followed by additional classification layers for caption objects predication as a multi-class multi-label problem. In decoding phrase, the global caption features would be employed to explore the latent contribution for each encoded region proposal feature in multi-head global attention layer, to help the decoder better attend on these most related region proposal. The word embedding layer and position embedding layer in decoder are ignored in the above figure to avoid ugly layout. The arrangement of the section is as follows: first in Section III-A, we will introduce the extraction of global caption feature; and then we will depict a standard Transformer neural networks structure in Section III-B; after that, the Section III-C will focus on introducing the details of the GANN, finally, the objective functions will be given in the Section III-D.

    A. Global Caption Feature Extraction

    In this subsection, we will introduce the extraction of the global caption feature in encoder which is illustrated in top of Fig. 2. Our encoder takes the work [23] as bias and takes the region proposal features as inputs, which is extracted through a pre-trained Faster-RCNN [13] (identical to the work [4]),and output the global caption feature as well as the encoded region proposal features. Additional FFN networks is following the encoder to predicate the caption objects by taking the global caption feature as input. In this way, the global caption feature is endowed the caption object information which by give the region proposal occurring caption object to be more chance noticed in the caption generation process.

    We will first introduce the selection of caption object labels and then create the caption object label vector as ground truth for each image in training set. After the caption vocabulary is obtained using all ground-truth captions from training set, we select these noun words from this vocabulary with the occurrence is higher than 100, here we get 359 caption object labels for the dataset MS-COCO. In order to eliminate duplicate words and ambiguities, we re-category these words into different classes, including combining singular and plural forms, removing gender attribute, for example, we cast the words “he”, “him”, “she”, “her” into the same class, saying“people”. After this process, we get 220 unique caption object labels. In each position of caption object label vector, we set its value as 1 if corresponding caption object appear in given captions, otherwise, set its value as 0. The obtained labels in training data will be used to guide the global caption feature extraction for predicating the caption objects through auxiliary classification layers.

    When predicting the caption object labels for a given image.Motivated by [29], we treat this problem as multi-class multilabel classification problem. Here, we will define a loss function for this task, and the corresponding network structure could be found in Fig. 2. Suppose there are an input image in training set, andl=[l1,...,lK]∈{0,1}Kis its label vector,Kis the dimension of caption object label vector, thenlk=1 if the image is annotated with the labelk, otherwiselk=0. we denote the input image representation for encoder asR=[r0,r1,r2,...,rn], in which we inject a special empty featurer0(zero vector) into region proposal featuresri,i∈1,2,...,n, andnis the number of region proposal. Ther0is the embedding of placeholder token for extracting global caption feature, playing the same role of “ 〈 s os 〉” embedding vector in NLP model, like BERT [16]. For convenient, we denotetheoutputsofencoder asYˉ=[y0,y1,...,yn],andY=Y??{y0}isencodedfeatures,wherey0is the corresponding output in position ofr0. FFN means feed forward networks following the encoder which is used for projecting the global caption feature into label distribution. We give following equations:

    B. Standard Transformer

    In this section, we will introduce the architecture of the standard Transformer network given by work [7], which is the based model for our work. We adopt the same approach to build our the word embedding layer and the position embedding layer as the transformer [7] work which has been ignored in Fig. 2 for page courtesy.

    The encoder and decoder in Transformer both consist a stack of layers (we set to 6 in our model), and each layer is composing by a multi-head self-attention layer, and a multihead attention layer, pulsing a FFN layer, where the first two layers both are composed by 8 identical heads with independent parameters and the network structure is the same as [7]. The network structure of head is shown on the right side of Fig. 3. Inspired by previous work, such like [23] in captioning task, we adopt the similar configuration as used in transformer. In each head of self-attention, the inputs are provided by the output of previous outputs (except the first layer take the sum of masked word embedding layer and position layer as input), however in multi-head attention layer of decoder, the inputs of each head provided by the encoded region proposal featuresYand the output of previous layerXat the same time.

    For convenience, we denoteA∈{X,Y}, in whichXis the output of previous layer, andYis encoded features from encoder. We take thatAequalsXfor these attention head in multi-head attention layer of encoder, otherwiseAequalsY.At the attention calculation process in each head, the queryQ,keyK, and valueVwill be calculated first for each of input tokens using the following linear projections:

    Fig. 3. The architecture of proposed global-based attention neural network(GANN) compared the scaled dot-product attention. Compared to (b), our proposed neural networks (a) incorporate external latent contributions of region proposals into attention calculation process.

    whereX∈RN×nmodel,WQ,WK,WA∈Rdmodel×nmodelare learned matrices for getting the query, key and value respectively, anddmodelis the dimension of internal vector in each head.

    We will introduces the process of the attention weight calculation in each head as follows:

    Equations (4) to (6) are calculated independently in each head. The final output of each layer is got by concatenating the output of all heads to one single vector, whose dimension is 512, and then multiplied with a learned projection matrixWOas the final output of this attention layer, i.e.,

    The next component of in each encoder and decoder layers is the point-wise F FN , which is applied to each outputxof the attention layer.

    whereW1,b1, andW2,b2are the weights and biases of two fully connect linear projection layers. In addition, skipconnections and layer-norm are applied to the outputs of the self-attention and the feed-forward layers.

    C. Global-Attention-Based Neural Network

    At this section, we will introduce the proposed GANN and give some intuition explanations about the reason why the global attention mechanism works. First, The global caption feature obtained from the encoder carries the caption object information, which could offer some help for the model to attend the most related region proposals and therefore benefit for caption generation. We let the extracted global caption feature get involved in the attention weight calculation process in each head of every decoder layer, because we believe that the attention could be affected not only by the “internal”factors, like current state and memory of model, also by the“external” latent contribution of each region proposals. We compute these two kinds of attention weight separately, and then try to combine two weights together using the method used in (12).

    We will depict the definition of the GANN here. Supposed that we got global feature vector from encoder annotated byy0for input image, i.e., the corresponding output for the ejected empty inputs of encoder, and we denoteYis the encoded region proposal features though encoder, i.e., the set of the output from encoder exclude the global feature vectory0. We define the global attention weight as follows:

    At the same time, we use hidden output at current time step by the multi-head self-attention module in each decoder layer,i.e., the content information, to obtain the local attention weight matrix using the equations from (4) to (5), like most of models do at the decoder module, the weight is calculated in following equation:

    whereWijmeans the element of the final attention weight matrixW. As for decoder structure, we employ the similar neural network structure as transformer [7], each layer is constructed by a self-attention layer, and a GANN followed by a feed-forward layer, the corresponding neural network structure is shown in Fig. 2. and the outputVˉ of the head is then calculated as

    The global attention encoding diagram shown in Fig. 2 shows the multi-head global attention layer of GANN, which would be repeatedly applied in each layer of decoder in the same way as the work [7].

    D. Training and Objectives

    For fair comparisons with the works trained using selfcritical sequence training (SCST) [34], we also report the results optimized for CIDEr [35]. Initializing from the corssentropy trained model, we seek to minimize the negative expected score

    IV. EXPERIMENTAL EVALUATION

    We conducted our experiments on the benchmark dataset for image captioning, and quantitatively and qualitatively compare the results with some state-of-the-art approaches in Sections IV-C and IV-D, respectively. Also, we conduct some ablative analysis about the effect encoder structure for global caption feature extraction, the effect of GANN on decoder,and the effect of the value of trade-off factor λ.

    A. Dataset and Metrics

    Our experiments are conducted on the Microsoft COCO(MS-COCO) 2015 Captions dataset [36], which totally includes 123 287 images. All of our experiments result on the Karpathy validation and test splits [37] that are widelyadopted in other image captioning approaches. There are 5 K images in Karpathy test and validation sets respectively. The final caption results are evaluated using the CIDEr-D [35],SPICE [39], BLEU [40], METEOR [41], ROUGE-L [42]metrics.

    TABLE I COMPARATIVE ANALYSIS TO EXISTING STATE-OF-THE-ART APPROACHES. WE COMPARE THE ORIGINAL MODEL WITH THESE MODELS WITH THE PROPOSED GANN. TWO KINDS OF RESULTS ARE PRESENTED, I.E., SINGLE MODEL AND ENSEMBLE METHOD. ALL RESULTS ARE OBTAINED ON MS-COCO “KARPATHY” TEST SPLIT, WHERE M, R, C AND S ARE SHORT OF METEOR, ROUGE-L, CIDER-D AND SPICE SCORES. ALL VALUES ARE REPORTED AS PERCENTAGE (%)

    B. Implementation Details

    Our algorithms was developed in PyTorch taking the implementation in [23] as our basis. We conduct experiments on NVIDIA Tesla V100 GPU. The ADAM optimizer with learning rate is initialized by 2×10–4and annealed by 0.95 every 3 epochs. We increase the scheduled sampling probability by 0.05 every 5 epochs [43]. We adopted the same warm up method as the work in [23], and set our batch size to 16. Before starting training the decoder, we took a pre-training procedure on encoder separately using the caption object prediction loss function defined in Section III-A in first 10 epochs. After that, we jointly trained the encoder and decoder in future epochs usingLXEjoint. We also trained another version model with self-critical reinforcement learning method (SCST) [34] using the objectives defined in 19 for CIDEr-D score when the loss score on the validation split does not improve for some training steps. In the cross entropy training and CIDEr-D score optimization method, early stop mechanism is adopted in all phrase of training.

    C. Quantitative Analysis

    In this subsection, we will show our experiment results,which demonstrate that the proposed global attention method can enhance the performance for the task of image captioning.First, we select three most representative state-of-the-art methods, which include both the regular image captioning decoder and Transformer based decoders. The comparison methods are briefly described as follows: 1) Up-Down [4], in which a separate LSTM is used to encode the “past” word in accumulate way, and an attention mechanism is utilized at each step of word generation; 2) RFNet [38], which tries to fuse encoded features from multiple CNN models; 3) GCNLSTM [26], which predicates visual relationship clues in each entity pairs and encodes this information into feature vectors through message passing mechanism; 4) Att2all [7], which maximum reserves the structure from the Transformer design for machine translation; 5) AoANet [11], which proposes an“attention on attention” module to determine the relevance between attention results and queries; 6) ObjRel [23], the geometric is involved in encoder for a purpose of exploring the spatial relationships among region proposals.

    For fair comparison, the models in Fig. 1 are trained in the same method using first trained XE loss and then optimized for CIDEr-D score, and two kinds of results, i.e., single model and ensemble models, are shown in Table I. In this table, we can see that our single model achieves the highest scores on BLEU-1, BLEU-4, CIDEr-D, and SPICE metrics, and on all metrics in ensemble models. As for the result using CIDER-D optimization, we achieve the best performance on BLEU-1,BLEU-4, ROUGE-L, SPICE in single model, and get 39.1 score on BLEU-4 score in ensemble models, which illustrates that our proposed GANN could enhance the performance for captioning task.

    D. Qualitative Analysis

    In Fig. 4 shows a few examples with images and captions generated by comparison models and compared results with our proposed GANN. We derive the captions using the same setting and training method, i.e., SCST optimization [34].From these examples, we find that these “base” models areless accurate for the image content, while our method could give relatively better captions. More specifically, our proposed global attention method is superior in the following two aspects: 1) our proposed GANN could help language model correctly focus on the caption objects. For example, our proposed method could help the model attend on the “flight”in the first example, and “tv” on the second example in Fig. 4,while the other models choose to ignore them. 2) our proposed GANN could help the language model count objects of the same kind more accurately. For example, there are two benches and two motorcycles in the third and fourth image in Fig. 4, but the comparison models only give one in their captions. However, our proposed model only tend to recognize concrete objects while ignores the relatively abstract concept compared to the ground truth captions, such as in the fourth image, our model could not give “autumn”, “fall” these season words, this limitation of understanding higher concepts will be our next research target in future work.

    TABLE II THE EXPERIMENT RESULTS OF MODELS COMPOSED BY DIFFERENT METHODS OF GLOBAL CAPTION FEATURE EXTRACTION AND DECODERS ON MS-COCO “KARPATHY” TEST SPLIT. ALL OF THESE EXPERIMENTS ARE TRAINING IN THE SAME METHOD USING XE-LOSS. ALL VALUES ARE REPORTED AS PERCENTAGE (%)

    Fig. 4. Examples of captions generated by the comparison approaches and our proposed approach along with their ground truth captions.

    In Fig. 5, we visualize region proposal with the highest attention weight in each time step of caption generation process. We select [7] as the “base” model and “base” +GANN as comparison, and training process are exactly the same using XE-loss which has been discussed above subsection. Observing the attended image region in Fig. 5, we find that the “base” model was attracted by the car behind the motorcycle, and the caption decoder generate the caption directly based on the attention result, therefore the corresponding caption is about that car. In contrast, the “base”+ GANN caption model could correctly focus on the “man”and “motorcycle” with the help of global caption feature and the proposed GANN, because it plays a role of reminder for language module at each time step of word predication.

    E. Ablative Analysis

    We conducted additional experiments to study three kinds of impacts for the caption generation results: 1) the method of global caption feature extraction, 2) the effect of GANN on decoder, 3) trade-off factor λ in the joint training process. All the results are obtained using the XE-loss and the same training method.

    1) Method of Global Caption Feature Extraction:We choose two additional kinds of methods besides our method mentioned in Section III-A to extract the global caption feature. More specifically, the mean method and CNN-based network are adopted, whose network architecture are shown in Fig. 6. In the experiments, we use Up_Down [4], LSTM and transformer [7] as decoder for comparison. The result presented in the Table II, we can see that the global caption feature extracted through Transformer based encoder could obtain a better result compared to another two kinds of encoder network structure on almost all metrics. For example,the model [7] with Transformer based encoder obtained 6.2 and 3.1 higher score on metric of CIDEr-D compared to the models with Mean and CNN based encoder respectively.

    Fig. 5. Visualization of attention regions in 6th layer of decoder for the “baseline” model and “baseline” with GANN. The “baseline” model can be easily misled by irrelevant attention while “baseline” with GANN is less like so.

    TABLE III THE EXPERIMENT RESULTS OF SEVERAL “BASE” MODELS AND “BASE” MODELS WITH GANN ON MS-COCO “KARPATHY” TEST SPLIT.ALL OF THESE EXPERIMENTS ARE TRAINING IN THE SAME METHOD USING XE-LOSS. ALL VALUES ARE REPORTED AS PERCENTAGE (%)

    Fig. 6. Two neural network structure of encoder for global caption feature extraction, i.e., the mean operation and convolutional neural networks, are adopted to obtain the global caption feature respectively in above figure.

    Fig. 7. Two neural network structure of decoder, i.e., the LSTM based decoder and and UP_Down based decoder, are adopted respectively in above figure in order to study the effect of GANN on decoder.

    2) Effect of GANN on Decoder:Besides our Transformerbased decoder, we choose another two kinds of network structures as language module as comparison to explore the effect of GANN on decoder. The another two kinds of decoders are LSTM and Up_Down, which is shown in Fig. 7.In all experiments, the global caption feature extraction method is the same, i.e., the Transformer based encoder. We compare these three “base” model with “base” + GANN model, and the results are shown in the Table III. We find that our proposed the GANN could improve the performance compared to original caption models. For example, our model(the last line of Table III) achieves 1.3 and 2.3 increment over the metrics of METOR and CIDEr-D respectively.

    Fig. 8. The performance affected by the trade-off factor of λ. In above we show 4 main metrics curve obtained by different value of λ sampled from 0.1 to 0.5,and 0.05 as separation distance.

    3) Effect on the Trade-off Factorλ:We also study the effect of λ on final performance, i.e., the trade-off factor between the loss for caption objects predication and XE-loss for caption generation, which is defined in Section III-D. The result is shown in the Fig. 8, in which the value of λ is chosen from the from 0 .1 to 0.5 and 0 .05 as separation distance. From the result, we can see that the λ value actually affects the final performance of our model, and it is also hard to get such a value for λ that all metrics could be optimized to maximum score. We take BLEU-4 as our bias, and select the λ to 0.2 in our experiments, in which we can obtain the BLEU-4 score 37.8, and corresponding results are shown in Table I.

    V. CONCLUSION

    In this paper, we proposed a global-attention-based neural network (GANN) for the captioning task, which enhances conventional attention mechanisms. GANN explores the latent contributions of region proposals and helps the resulting model correctly focus on the most relevant areas, which further improves the performance of the captioning model.Extensive experiments conducted on the MS-COCO dataset demonstrate the effectiveness of our proposed method.

    国产成人福利小说| 国产高清不卡午夜福利| 99热这里只有精品一区| 亚洲av一区综合| 国产午夜精品久久久久久一区二区三区| 亚洲在久久综合| 国产综合懂色| 国产美女午夜福利| 18禁裸乳无遮挡动漫免费视频 | 国产成人免费无遮挡视频| 777米奇影视久久| 亚洲成人中文字幕在线播放| 伦理电影大哥的女人| 老师上课跳d突然被开到最大视频| 午夜爱爱视频在线播放| 国产精品女同一区二区软件| 亚洲综合色惰| av在线观看视频网站免费| 国产成人精品福利久久| 男女无遮挡免费网站观看| 成人鲁丝片一二三区免费| 国产片特级美女逼逼视频| 99九九线精品视频在线观看视频| 精品久久久久久久久亚洲| 国产精品无大码| 成人毛片60女人毛片免费| 精品视频人人做人人爽| 亚洲欧美日韩另类电影网站 | 中文资源天堂在线| 免费看日本二区| 国产成人免费无遮挡视频| 日韩亚洲欧美综合| 亚洲精品国产色婷婷电影| 精品酒店卫生间| 精品少妇黑人巨大在线播放| 麻豆成人av视频| 国产成人freesex在线| 在线精品无人区一区二区三 | 成年女人看的毛片在线观看| 在线观看av片永久免费下载| 国产精品福利在线免费观看| www.色视频.com| 久久久精品94久久精品| 精品久久久精品久久久| 最近的中文字幕免费完整| 免费黄网站久久成人精品| 国产精品麻豆人妻色哟哟久久| 80岁老熟妇乱子伦牲交| 18+在线观看网站| 久久精品国产亚洲av涩爱| 伊人久久精品亚洲午夜| 国产毛片在线视频| 白带黄色成豆腐渣| 成年版毛片免费区| 成人高潮视频无遮挡免费网站| 人人妻人人澡人人爽人人夜夜| 国产成人精品福利久久| 国产熟女欧美一区二区| 久久久午夜欧美精品| 久久精品久久久久久噜噜老黄| av在线蜜桃| 国产精品一区二区在线观看99| 国产亚洲av片在线观看秒播厂| 国产 精品1| 亚洲,一卡二卡三卡| 亚州av有码| 免费高清在线观看视频在线观看| 国产高清不卡午夜福利| 午夜免费观看性视频| 欧美成人精品欧美一级黄| 大香蕉97超碰在线| 一级黄片播放器| 亚洲精品一区蜜桃| 精品久久久噜噜| 国产精品国产av在线观看| 18+在线观看网站| 伦精品一区二区三区| 白带黄色成豆腐渣| 边亲边吃奶的免费视频| 国产女主播在线喷水免费视频网站| 日韩人妻高清精品专区| 久久久亚洲精品成人影院| 国产精品偷伦视频观看了| 91aial.com中文字幕在线观看| 黄片wwwwww| 一级毛片 在线播放| 偷拍熟女少妇极品色| 欧美xxxx黑人xx丫x性爽| 亚洲欧美清纯卡通| 国产淫片久久久久久久久| 2021少妇久久久久久久久久久| 国模一区二区三区四区视频| 中文乱码字字幕精品一区二区三区| 人妻系列 视频| 99久久人妻综合| 亚洲av在线观看美女高潮| 欧美xxⅹ黑人| 亚洲欧美成人精品一区二区| 免费看不卡的av| 日日摸夜夜添夜夜添av毛片| 亚洲经典国产精华液单| 亚洲国产高清在线一区二区三| 国产免费福利视频在线观看| 欧美精品一区二区大全| 欧美日韩视频精品一区| 亚洲伊人久久精品综合| 日日撸夜夜添| 一边亲一边摸免费视频| 日韩成人av中文字幕在线观看| 日本av手机在线免费观看| 国产av国产精品国产| 2021少妇久久久久久久久久久| 国产人妻一区二区三区在| 国产视频内射| 久热久热在线精品观看| 久久久久久久大尺度免费视频| 久久精品久久精品一区二区三区| 观看美女的网站| 国内揄拍国产精品人妻在线| 国产精品久久久久久久电影| 午夜福利网站1000一区二区三区| 午夜免费男女啪啪视频观看| 国产成人午夜福利电影在线观看| 国产色婷婷99| 亚洲精品一二三| av福利片在线观看| 寂寞人妻少妇视频99o| 日韩av在线免费看完整版不卡| 婷婷色综合大香蕉| 少妇人妻 视频| 色哟哟·www| 欧美日韩视频高清一区二区三区二| 毛片一级片免费看久久久久| 99久国产av精品国产电影| 国产中年淑女户外野战色| 亚洲精品成人av观看孕妇| 高清毛片免费看| 国产色婷婷99| 亚洲av在线观看美女高潮| 久久久久久久大尺度免费视频| 人妻系列 视频| 国产91av在线免费观看| 国产成人午夜福利电影在线观看| 国产精品国产av在线观看| 国产精品不卡视频一区二区| 色网站视频免费| 亚洲激情五月婷婷啪啪| 草草在线视频免费看| 亚洲av.av天堂| av一本久久久久| 欧美一区二区亚洲| 一级毛片久久久久久久久女| 国产精品久久久久久av不卡| 久久久久九九精品影院| 一本一本综合久久| 欧美 日韩 精品 国产| 极品少妇高潮喷水抽搐| 亚洲国产精品国产精品| 亚洲天堂国产精品一区在线| 久久精品熟女亚洲av麻豆精品| 午夜精品一区二区三区免费看| 亚洲人成网站高清观看| 亚洲av成人精品一区久久| 日韩欧美一区视频在线观看 | 草草在线视频免费看| 97在线视频观看| 午夜老司机福利剧场| 麻豆乱淫一区二区| 国产亚洲一区二区精品| 亚洲天堂国产精品一区在线| 国产一区二区三区综合在线观看 | 99久国产av精品国产电影| 麻豆精品久久久久久蜜桃| 新久久久久国产一级毛片| 中文资源天堂在线| 美女脱内裤让男人舔精品视频| 中文欧美无线码| 亚洲精品,欧美精品| 亚洲欧美成人综合另类久久久| 99久久中文字幕三级久久日本| 国产精品国产三级国产av玫瑰| 最近最新中文字幕免费大全7| 日韩制服骚丝袜av| 一级片'在线观看视频| xxx大片免费视频| 亚洲国产欧美在线一区| 夫妻午夜视频| 久久99热这里只频精品6学生| 亚洲av一区综合| 啦啦啦中文免费视频观看日本| 一区二区三区乱码不卡18| 久热久热在线精品观看| 在线观看一区二区三区激情| a级一级毛片免费在线观看| 国产精品一区二区在线观看99| 日韩av不卡免费在线播放| 国产精品秋霞免费鲁丝片| 男人爽女人下面视频在线观看| 色视频在线一区二区三区| 亚洲精品日本国产第一区| 成年av动漫网址| 欧美老熟妇乱子伦牲交| 三级男女做爰猛烈吃奶摸视频| 国产女主播在线喷水免费视频网站| 男女下面进入的视频免费午夜| 欧美性猛交╳xxx乱大交人| 热99国产精品久久久久久7| 亚洲成人一二三区av| videos熟女内射| 麻豆成人av视频| 国产精品久久久久久久电影| 国产极品天堂在线| 久久久久久伊人网av| 一级二级三级毛片免费看| 老司机影院成人| 亚洲人成网站在线播| 亚洲丝袜综合中文字幕| 国产男女超爽视频在线观看| 国产精品福利在线免费观看| 在线观看三级黄色| 一级毛片久久久久久久久女| 深爱激情五月婷婷| 欧美成人a在线观看| 久久精品夜色国产| 丝袜喷水一区| 亚洲经典国产精华液单| 欧美日韩在线观看h| 国产黄a三级三级三级人| 日本av手机在线免费观看| 一区二区三区乱码不卡18| 毛片女人毛片| 国产69精品久久久久777片| 老女人水多毛片| 搞女人的毛片| 国产黄色免费在线视频| 久久久国产一区二区| 女人十人毛片免费观看3o分钟| 亚洲人与动物交配视频| 别揉我奶头 嗯啊视频| 国产一区二区三区综合在线观看 | 亚洲,一卡二卡三卡| 亚洲精品国产av成人精品| 建设人人有责人人尽责人人享有的 | 久久久精品免费免费高清| 日本熟妇午夜| 男人舔奶头视频| 五月玫瑰六月丁香| 91精品伊人久久大香线蕉| 欧美3d第一页| 亚洲精华国产精华液的使用体验| 亚洲国产日韩一区二区| 亚洲成色77777| 日韩av在线免费看完整版不卡| 亚洲av.av天堂| av卡一久久| 男人狂女人下面高潮的视频| 免费看不卡的av| 国产精品.久久久| 国产女主播在线喷水免费视频网站| 日本色播在线视频| 国产亚洲91精品色在线| 久久久久久久精品精品| 下体分泌物呈黄色| 亚洲av在线观看美女高潮| 伊人久久精品亚洲午夜| 22中文网久久字幕| 色吧在线观看| 好男人视频免费观看在线| 国产91av在线免费观看| 日本熟妇午夜| 国产一区二区三区av在线| 中文精品一卡2卡3卡4更新| 精品久久久久久久末码| 男人狂女人下面高潮的视频| 亚洲无线观看免费| 男人舔奶头视频| 成人国产麻豆网| 免费大片18禁| 人妻系列 视频| 日韩电影二区| 午夜福利在线观看免费完整高清在| 免费观看a级毛片全部| 青春草亚洲视频在线观看| 久久精品久久久久久久性| 伊人久久精品亚洲午夜| 午夜福利在线在线| 色吧在线观看| 亚洲aⅴ乱码一区二区在线播放| 国产成人aa在线观看| 久久久久久久大尺度免费视频| 赤兔流量卡办理| av在线蜜桃| 亚洲国产欧美在线一区| 丝袜脚勾引网站| 亚洲精华国产精华液的使用体验| 全区人妻精品视频| 又大又黄又爽视频免费| 精品久久久久久久久av| 国产在视频线精品| 国产精品麻豆人妻色哟哟久久| 欧美三级亚洲精品| 七月丁香在线播放| 国产在视频线精品| 久久精品熟女亚洲av麻豆精品| 寂寞人妻少妇视频99o| 国产av不卡久久| 国产精品99久久99久久久不卡 | 三级国产精品片| 精品99又大又爽又粗少妇毛片| 九草在线视频观看| 国产黄色免费在线视频| 中文精品一卡2卡3卡4更新| 好男人视频免费观看在线| 免费看av在线观看网站| 少妇的逼水好多| 亚洲第一区二区三区不卡| 精品国产三级普通话版| 日韩强制内射视频| 亚洲国产精品成人久久小说| av在线app专区| 免费大片黄手机在线观看| 日韩av不卡免费在线播放| 黑人高潮一二区| 欧美成人一区二区免费高清观看| 欧美丝袜亚洲另类| 男人添女人高潮全过程视频| 亚洲成人av在线免费| 26uuu在线亚洲综合色| 国产精品精品国产色婷婷| 男人和女人高潮做爰伦理| 国产久久久一区二区三区| 内射极品少妇av片p| 亚洲一区二区三区欧美精品 | 中文字幕免费在线视频6| 欧美+日韩+精品| 国产精品国产三级专区第一集| 久久人人爽人人爽人人片va| 女人十人毛片免费观看3o分钟| 干丝袜人妻中文字幕| kizo精华| 人妻夜夜爽99麻豆av| 啦啦啦在线观看免费高清www| 亚洲欧美清纯卡通| 久久久久久久久久久丰满| 国产白丝娇喘喷水9色精品| 深夜a级毛片| 国产视频首页在线观看| 亚洲成人久久爱视频| 新久久久久国产一级毛片| 亚洲av二区三区四区| 精品人妻视频免费看| 18禁在线播放成人免费| 国产乱来视频区| 国产在线男女| 亚洲欧美精品自产自拍| 日韩大片免费观看网站| 亚洲婷婷狠狠爱综合网| 99热这里只有精品一区| 日韩,欧美,国产一区二区三区| 欧美亚洲 丝袜 人妻 在线| 91久久精品国产一区二区三区| 一本一本综合久久| 国产成人a区在线观看| 在线看a的网站| 爱豆传媒免费全集在线观看| 久久久久九九精品影院| videos熟女内射| av.在线天堂| 国产成人免费观看mmmm| 国产亚洲精品久久久com| 最近最新中文字幕大全电影3| 熟女电影av网| 国产欧美另类精品又又久久亚洲欧美| 春色校园在线视频观看| 国产亚洲av片在线观看秒播厂| 日本-黄色视频高清免费观看| 成人国产av品久久久| 国产视频首页在线观看| 秋霞在线观看毛片| 亚洲人成网站在线观看播放| 自拍偷自拍亚洲精品老妇| 国产综合精华液| 香蕉精品网在线| 精品酒店卫生间| 草草在线视频免费看| 国产免费一区二区三区四区乱码| av黄色大香蕉| av.在线天堂| 国产亚洲91精品色在线| 国产精品一区www在线观看| 精品久久久久久久人妻蜜臀av| 婷婷色麻豆天堂久久| 青青草视频在线视频观看| 欧美丝袜亚洲另类| 亚洲在久久综合| 26uuu在线亚洲综合色| 亚洲高清免费不卡视频| 国产综合懂色| 成人毛片60女人毛片免费| 又粗又硬又长又爽又黄的视频| 黄色视频在线播放观看不卡| 少妇的逼水好多| 黄片无遮挡物在线观看| 欧美激情在线99| 极品少妇高潮喷水抽搐| av在线蜜桃| av一本久久久久| 欧美丝袜亚洲另类| 亚洲最大成人中文| 精品一区二区免费观看| 国内精品宾馆在线| 欧美日本视频| 精品人妻视频免费看| 91精品伊人久久大香线蕉| 乱码一卡2卡4卡精品| 纵有疾风起免费观看全集完整版| a级毛片免费高清观看在线播放| 女人被狂操c到高潮| 国产白丝娇喘喷水9色精品| 我的女老师完整版在线观看| 国产片特级美女逼逼视频| 国产成人午夜福利电影在线观看| 欧美日韩在线观看h| 精品人妻熟女av久视频| 国产毛片a区久久久久| 久久久久久伊人网av| 水蜜桃什么品种好| 91狼人影院| 亚洲美女视频黄频| 99久国产av精品国产电影| 国产精品国产av在线观看| 97人妻精品一区二区三区麻豆| 欧美成人a在线观看| 黄色欧美视频在线观看| 国产永久视频网站| 精品一区在线观看国产| 麻豆乱淫一区二区| 黄色一级大片看看| 一本色道久久久久久精品综合| 久久久久国产精品人妻一区二区| 一个人看视频在线观看www免费| 精品久久久久久电影网| 亚洲四区av| 免费av观看视频| 又大又黄又爽视频免费| 精品一区二区三区视频在线| 中文资源天堂在线| 九九久久精品国产亚洲av麻豆| 国产老妇伦熟女老妇高清| 色5月婷婷丁香| 亚洲欧美成人综合另类久久久| 99久国产av精品国产电影| 色5月婷婷丁香| 国产成人91sexporn| 最近2019中文字幕mv第一页| 久久鲁丝午夜福利片| 亚洲精品乱码久久久v下载方式| 国产精品一区二区在线观看99| 国产亚洲最大av| 成人二区视频| 欧美三级亚洲精品| av在线天堂中文字幕| 日日啪夜夜爽| 婷婷色麻豆天堂久久| 毛片一级片免费看久久久久| 久久这里有精品视频免费| 男女下面进入的视频免费午夜| 精品一区二区三区视频在线| 日本熟妇午夜| 一级片'在线观看视频| 91精品国产九色| 纵有疾风起免费观看全集完整版| 国产黄片视频在线免费观看| 免费少妇av软件| 国产精品成人在线| 亚洲国产精品成人综合色| 大又大粗又爽又黄少妇毛片口| 边亲边吃奶的免费视频| 国产一区亚洲一区在线观看| 又爽又黄无遮挡网站| 人妻一区二区av| 2021天堂中文幕一二区在线观| 国产在视频线精品| 久久亚洲国产成人精品v| 国产精品一及| 国产美女午夜福利| 黄色日韩在线| 青春草亚洲视频在线观看| 国产综合懂色| 亚洲精品成人久久久久久| 少妇人妻精品综合一区二区| 人人妻人人澡人人爽人人夜夜| 亚洲精品国产成人久久av| av黄色大香蕉| 欧美日韩视频精品一区| 亚洲国产色片| 大陆偷拍与自拍| 亚洲色图综合在线观看| 少妇的逼水好多| 丰满人妻一区二区三区视频av| 超碰97精品在线观看| av.在线天堂| 一二三四中文在线观看免费高清| 可以在线观看毛片的网站| 精品国产三级普通话版| 亚洲欧美日韩另类电影网站 | 极品教师在线视频| 成人毛片a级毛片在线播放| 国模一区二区三区四区视频| 亚洲无线观看免费| 欧美精品国产亚洲| 亚洲久久久久久中文字幕| 日日啪夜夜撸| 国内精品宾馆在线| 成人免费观看视频高清| 国产精品爽爽va在线观看网站| 久久久a久久爽久久v久久| 欧美bdsm另类| 国产视频内射| 交换朋友夫妻互换小说| 最后的刺客免费高清国语| 日韩国内少妇激情av| 一级毛片电影观看| 国产一区有黄有色的免费视频| 欧美日韩精品成人综合77777| 少妇人妻精品综合一区二区| 成年女人看的毛片在线观看| 卡戴珊不雅视频在线播放| 91精品一卡2卡3卡4卡| 午夜激情久久久久久久| 国产在线一区二区三区精| 国产 一区精品| 亚洲欧美一区二区三区国产| 精品熟女少妇av免费看| 天天躁日日操中文字幕| 国产伦理片在线播放av一区| 少妇的逼好多水| 乱码一卡2卡4卡精品| 免费观看a级毛片全部| 日韩一区二区三区影片| 午夜福利网站1000一区二区三区| 丰满少妇做爰视频| 少妇高潮的动态图| 欧美性感艳星| 久久久欧美国产精品| 又粗又硬又长又爽又黄的视频| 日本三级黄在线观看| 亚洲精品成人久久久久久| 亚洲成人一二三区av| 亚洲精品视频女| 精品一区二区三区视频在线| 日本免费在线观看一区| 最后的刺客免费高清国语| 久久久久国产网址| 观看美女的网站| 日日摸夜夜添夜夜添av毛片| 久久精品综合一区二区三区| 国产探花在线观看一区二区| 亚洲一级一片aⅴ在线观看| 久久精品国产鲁丝片午夜精品| 精品国产乱码久久久久久小说| 久久精品久久久久久久性| 日日啪夜夜撸| 少妇人妻精品综合一区二区| 欧美精品一区二区大全| 精品人妻熟女av久视频| 纵有疾风起免费观看全集完整版| 亚洲欧美一区二区三区国产| av播播在线观看一区| 国产乱来视频区| 国产黄频视频在线观看| 国产中年淑女户外野战色| 国产永久视频网站| 亚洲欧美中文字幕日韩二区| 亚洲欧美精品自产自拍| 春色校园在线视频观看| 亚洲电影在线观看av| 午夜老司机福利剧场| 高清视频免费观看一区二区| 激情 狠狠 欧美| 国产成人午夜福利电影在线观看| 久久久国产一区二区| 国产大屁股一区二区在线视频| 美女主播在线视频| 女的被弄到高潮叫床怎么办| 欧美日韩精品成人综合77777| 久热久热在线精品观看| 亚洲天堂av无毛| 中文字幕制服av| 国产一区二区在线观看日韩| 亚洲国产成人一精品久久久| 少妇被粗大猛烈的视频| 狠狠精品人妻久久久久久综合| 看非洲黑人一级黄片| 久久久久久久国产电影| 视频区图区小说| 久久久久久久久久久丰满| 亚洲国产精品专区欧美| 97在线人人人人妻| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 欧美97在线视频| 国产av不卡久久| 久久这里有精品视频免费| 久久久久久久午夜电影| 国产又色又爽无遮挡免| 成人漫画全彩无遮挡| 欧美人与善性xxx| 久久精品久久精品一区二区三区| 岛国毛片在线播放| 天天躁日日操中文字幕| 日韩在线高清观看一区二区三区| 欧美性猛交╳xxx乱大交人| 亚洲人成网站在线观看播放| 噜噜噜噜噜久久久久久91| 伦精品一区二区三区|