• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Time Highlighted Multi-Interest Network for Sequential Recommendation

    2023-10-26 13:15:00JiayiMaTianhaoSunandXiaodongZhang
    Computers Materials&Continua 2023年9期

    Jiayi Ma,Tianhao Sunand Xiaodong Zhang

    College of Computer Science,Chongqing University,Chongqing,400044,China

    ABSTRACT Sequential recommendation based on a multi-interest framework aims to analyze different aspects of interest based on historical interactions and generate predictions of a user’s potential interest in a list of items.Most existing methods only focus on what are the multiple interests behind interactions but neglect the evolution of user interests over time.To explore the impact of temporal dynamics on interest extraction,this paper explicitly models the timestamp with a multi-interest network and proposes a time-highlighted network to learn user preferences,which considers not only the interests at different moments but also the possible trends of interest over time.More specifically,the time intervals between historical interactions and prediction moments are first mapped to vectors.Meanwhile,a time-attentive aggregation layer is designed to capture the trends of items in the sequence over time,where the time intervals are seen as additional information to distinguish the importance of different neighbors.Then,the learned items’transition trends are aggregated with the items themselves by a gated unit.Finally,a self-attention network is deployed to capture multiple interests with the obtained temporal information vectors.Extensive experiments are carried out based on three real-world datasets and the results convincingly establish the superiority of the proposed method over other state-of-the-art baselines in terms of model performance.

    KEYWORDS Recommender system;temporal dynamics;multi-interest network;trends;attention mechanism

    1 Introduction

    Recommender systems explore users’potential interests based on historical interactions and then analyze the correlation between interests and items to provide personalized recommendations.Collaborative filtering [1] is a classic algorithm of recommender systems,and it works based on the fact that similar users may share similar preferences and similar items may be liked by users.Many traditional methods [2–4] are designed based on it to alleviate the problem of data sparsity.With the advent of deep learning,neural network-based approaches[5]are proposed for their capacity for representation.For example,Guo et al.[6] combined deep learning and factorization machines to learn both high-order and low-order feature interactions.Wang et al.[7]captured richer embedding representations based on high-order connectivity in the user-item graph.However,the above methods learn fixed user embeddings,which cannot capture dynamic demands as user preferences always change over time.

    Sequential recommendation treats historical interactions as a chronological sequence that contains the evolution of user preferences to guarantee more accurate,customized,and dynamic recommendations.The early research on the sequential recommendation predominantly relied on Markov Chains (MC) [8],which made the assumption that the dependency of each interaction lies in its preceding sequence.Recent methods [9,10] have achieved satisfactory performance by converting information into low-dimensional embeddings and using neural networks to learn user representations.For more effective and interpretable recommendations,some methods[11–13]utilized auxiliary relations in knowledge graph to capture more connections between items to enrich item embedding,rather than being limited to a sequence.In addition,the idea of self-supervised[14]which adaptively adjusts parameters by the difference between the current optimal solution and the global optimal solution has been introduced into recommendation systems,several models [15,16] extracted user embeddings from multiple perspectives and make the distance between them closer to obtain more accurate representation of users and improve model performance.

    To better match the real-world recommendation scenarios,multi-interest-based models [17,18]have been proposed to extract multiple interests that represent different aspects of user preferences.Following the idea,Tan et al.[19]constructed intent prototypes and assign different weights to obtain various interests.Chen et al.[20]explored the periodicity and interactivity of user behavior sequences to enrich item embedding and utilized an attention network to extract multiple interests.Nevertheless,these methods have the following two problems: (1) they only answer the question“What are the interests of users”without distinguishing the importance of different interests in predicted time;(2)the transition trends of items in sequences that can simulate the potential interest of users over time are not fully mined.This work argues that temporal information is a key factor in extracting user interests,and transition trends can reflect possible points of interest.As shown in Fig.1,the user interacted with three categories of products in the last week: laptops,clothes,and desks.Previous models tend to treat each interest equally,resulting in content that users are not interested in still being recommended.By contrast,this work considers timestamps when extracting interests,pays more attention to recent interests,and recommends items that the user may be like based on the trends.Besides,Fig.2a shows the distribution of time intervals between two adjacent items on Amazon Books,it can be seen that the timespan in the historical sequences might be large.Traditional approaches treat user interests(items marked with black,blue,and green series)equally,but the time gap between the interactions related to the black series and the last interaction exceeds 100 days,the user(whose most recent interests are blue and green series)is unlikely to be interested in it,suggesting the possibility of utilizing temporal information.Fig.2b plots the behavior analysis of a user’s last thirty interactions,revealing that interactions tend to be grouped within a short period of time,while the items that are interacted within a short period of time often have correlations.So,the transformation of items in the sequence can reflect the dynamic changes of points of interest to a certain extent.

    Figure 1:An example of sequential recommendation using temporal information.Two more recent interests“clothes&desks”and possible trends of them are mainly considered when predicting the top four items

    Figure 2:Analysis of temporal information about user interactions on Amazon Books.(a) Is the distribution of time intervals between two consecutive interactions (the long tail of the distribution is not included).(b)Is the behavior analysis of a user’s last thirty interactions.The X-axis denotes the time since the first interaction,while the Y-axis denotes the interaction count.Items with different categories are represented by circles of different colors

    Aiming at the problems that the existing multi-interest methods cannot capture the temporal dynamics of user preferences,and do not make full use of interaction history to simulate the changing trends of user interests.This paper proposes a novel method calledtime highlightedmultiinterest network for sequentialrecommendation(TimiRec),which assigns different weights to multiple interests based on the prediction moment,and aggregates the updated neighbor item representations as the transition trends of the current item.The main contributions of this work are summarized as follows:

    ? Based on the prediction moment,a linear time interval function is designed to generate time interval information as the key factor to extract multiple interests from the user’s behavior sequence.

    ? A time-attentive aggregation layer is introduced to aggregate neighbor items,in which an attention network is used to update the representations of neighbor items by considering the time intervals between the item and its neighbors,thereby capturing the changing trends of points of interest.And a gated unit is deployed to adaptively fuse the initial items embeddings and the captured trends.

    ? The effectiveness of the proposed algorithm is evaluated by comparing it with various baseline methods on three real-world datasets,and the results convincingly establish the superior performance of the proposed model over state-of-the-art baselines.

    2 Related Work

    2.1 Sequential Recommendation

    Sequential recommender systems treat users’historical behaviors in chronological order and model the dependencies between items to provide more accurate recommendations.The model based on Markov Chains[8,21]is one of the most classical models.Nevertheless,a significant shortcoming of MC-based methods is their limited consideration of long-term dependencies,since they only rely on the most recent interactions.The field of sequential recommendation has embraced the advancements in deep learning,incorporating Recurrent Neural Network (RNN) and its variants,such as Long Short-Term Memory(LSTM)and Gated Recurrent Unit(GRU).Zhou et al.[22]used GRU to increase the accuracy of prediction by supervising the hidden state.However,RNN-based methods that utilize the current state and previous state as input still have several problems,such as difficulty in parallelization and learning long-term dependencies.To tackle these problems,inspired by Transformer[23],Kang et al.[10]stacked self-attention layers to capture item relevance.Moreover,recent studies focused on incorporating interaction timestamps into the sequential modeling process.For instance,Li et al.[24]added relative time interval and position information into item embeddings.Following a normal distribution,Wang et al.[11]introduced two distinct temporal kernel functions to explicitly model the evolution of user preferences over time in terms of“complement”and“substitute”relations.Jiang et al.[25]designed a time weighting function to enhance the influence of the time effect of evaluation.

    2.2 Attention Mechanism

    Attention mechanism is a technique that considers the importance of each item in the input sequence to the output,recognizing that not all items in the sequence are equally important.Chen et al.[26] introduced attention mechanism into recommender systems as an additional component earlier.Xiao et al.[27] combined attention networks and Factorization Machine (FM) to improve the performance and interpretability of the model.Wang et al.[28] learned an attentive transactional context embedding which paid more attention to relevant items.And Cai et al.[29]measured the importance of different friends in social networks based on attention mechanism.More Recently,Vaswani et al.[23]proposed a sequence-to-sequence method named Transformer with a pure attention mechanism,which surpasses Convolutional Neural Network (CNN)/RNN-based approaches and achieves state-of-the-art performance.Unlike sequentially propagating sequence information,Transformer introduces the concept of query,key,and value to capture the relationship between items in the sequence,and then updates each item based on the similarity.This enables the model to simultaneously focus on all relevant parts of the sequence,which allows for better long-term dependencies modeling and improving the overall performance of the model.

    3 Methodology

    In this section,we begin by formulating the task of sequential recommendation and subsequently present the approach to map time intervals to corresponding vectors and construct the neighbor-aware graph,as well as the details of the proposed framework.

    3.1 Problem Formulation

    3.2 Linear Time Interval

    To capture temporal dynamics,a simple way is to use a time decay function [11,30],but its disadvantages are that the fitting capacity is limited and the importance weights are not normalized.Another way[24,31]is to map timestamps to vectors,although it can improve model performance,its calculation of relative time intervals is overly dependent on the minimum value.

    Besides,note that in Fig.2b,interactions tend to be grouped in a short time,indicating the importance of ensuring the small range of time intervals are treated equally.Taking advantage of the above two approaches,this paper designs a linear time interval function to model the effect of temporal information.Specifically,for a given anonymous time sequenceT=[t1,t2,...,t|T|],the linear time interval between interaction at timetand recommendation moment is defined as follows:

    wheretrrepresents the timestamp of the target item,andαis the coefficient of the linear function.By adjusting the value ofα,ensure that the time interval is the same within the range ofα.And then the linear time interval sequence is transformed intoI=[i1,i2,...,i|T|]with a clip operation to avoid sparse relation encoding:

    wheremis the threshold.

    3.3 Neighbor-Aware Graph Construction

    Generally,adjacent interactions in sequence are often related to similar interests.To capture the changing trends of items and enrich the representation of itemxi,we attentively aggregate the embedding of its neighbors(i.e.,δ-neighbor set) in graphG,which is generated based on the pairwise item transitions in interaction records of all users.Since the time interval between paired items can indicate their correlation,utilizing it to distinguish the importance of different neighbors is necessary.Note that the same pairwise item transition may occur at multiple different time intervals,so choose the largest one to cover all possible situations and ensure the comprehensiveness of the model.

    3.4 TimiRec Framework

    Fig.3 provides an overview of our proposed framework,TimiRec.Each part of the model will be described in detail next.

    Figure 3:The architecture of TimiRec.The linear function is applied to time interval information.A neighbor-aware construct and a time-attentive aggregation layer are deployed to sample top-H collaborative neighbors and model the trends of items over time,respectively.Finally,the multi-interest extraction layer is introduced to generate diverse preferences from the output of the gated fusion unit

    3.4.1 Embedding Layer

    The interaction sequence [x1,x2,...,x|X|] is converted into a fixed-length sequenceX=[x1,x2,...,xL],whereLrepresents the maximum length.If the sequence is longer thanL,only the most recentLitems are kept,and zero pad the sequence with zero on the left of the sequence if it is shorter thanL.The time interval sequence[i1,i2,...,i|T|]is also transformed intoI=[i1,i2,...,iL]to maintain the corresponding time interval of each interaction.

    EX∈R|V|?dis the learnable embedding matrix for all items,wheredrepresents the latent dimension.Then,the embedding of the behavior sequenceX∈RL?dis obtained by a lookup operation.Similar to the behavior sequence,another embedding matrixEI∈Rm?dis created for linear time interval sequenceI,wheremrepresents the maximum number of time intervals.After retrieval,the time interval sequence embeddingI∈RL?dis obtained.

    3.4.2 Time-Attentive Neighbor Relation Aggregation

    3.4.3 Multi-Interest Extraction Layer

    Considering the impact of temporal dynamics,the temporal embeddingI∈RL?dis used as the query of attention mechanism to selectively aggregate behavior sequences:

    whereW3∈Rd?4dandW4∈R4d?Kare learnable parameters.A∈RL?Kestimates the contribution of each item in the sequence to the multiple interests,S∈RK?dindicates multiple interests andKis the number of interests.

    Figure 4:Illustration of the time-attentive aggregation layer,which aggregates the representation of neighbor items by considering time interval information

    3.4.4 Model Training

    After obtaining the interest embeddings from the multi-interest extraction layer,for the target itemv,the argmax operator is used to select the corresponding interest vector fromKcandidate interest representations as user embedding:

    When provided with a training sample(u,v)containing the user embeddingsand item embeddingev,the probability of useruwill interact with itemvis calculated as follows:

    Since the sum operator in Eq.(11)is computationally time-consuming,a sampled softmax method is introduced to minimize the following objective function:

    3.4.5 Model Testing

    Different from the training phase,different interests representing different aspects of user preferences can independently provide top-Nitems in the testing phase.To get the final top-Nprediction results fromK?Ncandidate items,a straightforward approach is to select those items with the greatest similarity as the final prediction based on the inner product of candidate items and user interests,and it’s defined as:

    whereevis the embedding of a candidate item,andskindicates thek-th interest.

    4 Experiments

    In this section,to validate the effectiveness of the proposed framework,extensive experiments with other state-of-the-art baseline methods are carried out on three real-world datasets.

    4.1 Experimental Settings

    4.1.1 Datasets

    TimiRec is evaluated on three public datasets of diverse domains and sizes,Table 1 presents the statistics information of these datasets.

    Table 1:The dataset statistics

    ? Amazon1http://jmcauley.ucsd.edu/data/amazon/.: A commonly used review dataset that comprises various sub-datasets,and the following two specific sub-datasets are used:Books and Beauty.

    ? MMTD2http://www.cp.jku.at/datasets/MMTD/.:Million Musical Tweets Dataset(MMTD)[32]is a dataset of listening events collected from Twitter.

    To ensure data quality,interactions involving users and items with less than 5 occurrences are filtered out,and all users are split into training/validation/test sets in a ratio of 8:1:1.For model training,the complete sequences of interactions from the training users are utilized.More specifically,for a user sequenceXu=,each training sampleuses the firsttbehaviors to predict the(t+1)-thinteraction,wheret=4,5,...,|Xu|-1.The number of neighborsHis 20,and the distanceδbetween the item and neighbors is set to 2.Each training sample is truncated to 20.For model testing,the first 80% of interactions of the user sequence from validation and test users are used as input of the trained model to infer user embeddings,and metrics are calculated by predicting the remaining 20% of interactions.

    4.1.2 Evaluation Metrics

    To evaluate the performance of the proposed TimiRec,three widely adopted evaluation criteria for top-N recommendation are used in our experiments,i.e.,Recall,Normalized Discounted Cumulative Gain (NDCG),and Hit Ratio (HR).Among them,Recall@N represents the proportion of ground truth items included in the recommended N candidates,NDCG@N is a position-aware metric that assigns higher scores to ground truth items appearing at higher positions in the recommendation list,and HR@N focuses on determining whether the ground-truth item is present among the recommended items.N is set to 20 and 50 in our experiments.

    4.1.3 Baselines

    Comparative evaluations are conducted with TimiRec and the following methods:

    ?YouTube DNN[33]:it is a deep learning model designed for an industrial recommendation that applies deep neural networks to YouTube video recommendation(YouTube DNN).

    ?GRU4Rec[9]:it first applies GRU for the sequential recommendation.

    ?MIND[17]:it is a multi-interest model that incorporates a capsule network to extract multiple user interests.

    ?ComiRec-DR[18]:it follows MIND that extracts multiple interests using dynamic routing,and considers both diversity and accuracy of recommendation with a controllable factor.

    ?ComiRec-SA[18]: another variant of ComiRec that utilizes self-attention to extract diverse interests.

    ?PIMI[20]:a state-of-the-art model based on ComiRec-SA,periodicity and interactivity of user behavior sequence are explored to collect features before extracting multiple interests.

    4.1.4 Details

    TimiRec is implemented with TensorFlow.The embedding dimension is set to 64.For Books,MMTD,and Beauty,the reciprocal of the coefficient 1/αare set to 1 day,1 min,and 1 day,the number of time interval thresholdsm/nare set to 32/8,128/16,and 64/16,respectively,and the number of interest embedding is 4.The number of samples for sampled softmax loss is set to 10.The maximum number of training iterations is set to 1 million.A widely used optimizer Adam [34] is adopted for optimization with a learning ratelr=0.001.

    4.2 Overall Performance

    Table 2 shows a comprehensive summary of the performance of various methods across different datasets.Compared with GRU4Rec and YouTube DNN,which represent users as a single vector,multi-interest methods MIND,ComiRec,and PIMI achieve significant improvement,which implies that extracting multiple interests is more consistent with real-world scenarios.By incorporating the concepts of periodicity and interactivity in user behavior sequence,PIMI obtains notable enhancements compared to ComiRec.Our proposed method,TimiRec,consistently outperforms other baseline methods by effectively capturing the temporal dynamics of user interests.

    4.3 Ablation Study

    Table 3 presents the comparison results under the evaluation metrics of Recall@20,NDCG@20,and HR@20,where TimiRec-time and TimiRec-neigh denote TimiRec without temporal information and neighbor aggregate unit respectively.It can be found that removing any part harms results,suggesting that both of them are important to capture sequential information.Besides,lacking temporal information will lead to bigger performance loss.This is because the time interval information directly affects the weight of the current item when extracting interests,while neighbors are aggregated into items to enrich the item representation.

    Table 2:Performance comparison of TimiRec and other baselines(%).In each row,the best performance is bolded,and the second best is underlined

    4.4 Time Interval Function

    To verify the effectiveness of the designed linear time interval function,we removed the neighboraggregation layer(i.e.,TimiRec-neigh in Section 4.3),and modify the Eq.(1)as:

    Eq.(14)represents the logarithmic function,β is a coefficient,which assumes the effect of time on user interests gradually slows down as the time interval increases.Eq.(15)is the processing method in the paper proposed by Li et al.[24],min(iu)means the minimum time interval of useru.Because each training sample relies on the firsttbehaviors to predict the(t+1)-th interaction,leading to the value ofmin(iu)may change,so the relation between items will also change.Table 4 illustrates the performance when the time function changes.Experimental results demonstrate the great improvement brought by the proposed method.

    Table 4:The performance of the time interval function(%)

    4.5 Study on Hyper-Parameters

    In order to gain a deeper understanding of how different hyper-parameters impact the performance of the model,the effect of time interval threshold m and n,the coefficient of time interval functionα,and the number of interestsKare studied.

    As shown in Tables 5 and 6,the values of time interval thresholdsmandnare varied to investigate the effect of modeling temporal dynamics.The parametermcontrols the maximum time scope that directly affects user interests,andnensures the quality of aggregated neighbor items.The values ofmandnare selected from{16,32,64,128}and{2,4,8,16},respectively.Experimental results on Amazon Books show that model achieves the best performance whenm/nis set to 32/8.An excessively large time interval threshold may result in sparse encoding,conversely,setting a time interval threshold that is too small may lead to insufficient learning.

    Table 5:Effect of threshold m(%)

    Fig.5 illustrates the comparison of model performance on Amazon Books across different values of the coefficients α.The X-axis indicates 1/α,which is chosen from{0.25 day,0.5 days,1 day,1.5 days,2 days}.As is observed,TimiRec has the best performance when 1/α is 1 day.Combined with the best setting of 32 form,it can be further inferred that a person reads a book in about one month.

    Figure 5:Performance comparison for the number of coefficients α on Amazon Books

    Fig.6 presents the Metrics@20 and Metrics@50 results,demonstrating the effect of the interest numberKon Amazon Books.TimiRec obtains the best performance whenKis 4.Increasing the number of interests does not always improve the model effect,which is in line with real-world recommendation scenarios,where users usually do not have too many or too few interests.

    Figure 6:Effect of the number of interest K on Amazon Books

    4.6 Training Efficiency

    As shown in Fig.7,the metrics of Recall@20 are tested on Amazon Books during the training process for the proposed model and two other state-of-the-art methods to show training efficiency compared to the proposed model.It can be observed that the evaluation metric of Recall@20 has roughly the same trend with iteration on three models.In terms of the average time per iteration,TimiRec takes an average of 0.050 s,which is 1.79 times larger than ComiRec-SA,which attributes to the aggregation of neighbor information and the preprocessing of time interval information.But compared with PIMI with an average iteration time of 0.100 s,the training efficiency of our model has been greatly improved,this is because the computation of the stacked three-layer self-attention network in the interactivity module is very time-consuming.

    Figure 7:Training efficiency on Amazon Books

    4.7 Case Study

    The attention weights among multiple interests and items in the input sequence are visualized,which demonstrates the advantage of the proposed method by comparison with ComiRec-SA.Fig.8 illustrates the heatmap of the attention weights(corresponding to the value ofAin Eq.(9))associated with a user randomly selected from Amazon Books.Compared with ComiRec-SA which is not aware of time,it can be seen that for items under the same interest,TimiRec will assign higher weights to the more recently interacted items.And for items of the same category with smaller timespans,TimiRec will assign similar weights(Interest 2).

    Figure 8:Heatmap of attention weights among multiple interests and input items. I represents the linear time interval sequence corresponding to the input sequence of a user from Amazon Books

    5 Conclusion

    This work proposes a novel framework named TimiRec,which utilizes temporal information to extract multiple user interests.Specifically,multiple interests of users are generated by highlighting the time intervals in the multi-interest extraction layer,and combined with the neighbor-aware aggregation unit to capture possible trends of points of interest.The effectiveness and efficiency of the proposed model have been empirically verified through experiments conducted on three real-world datasets.In future work,we will combine temporal information and knowledge graph to build bridges between items and further explore their relationships to capture the possible trend of interests more comprehensively.

    Acknowledgement:The resources and computing environment are provided by Chongqing University,Chongqing,China.We are thankful for their support.

    Funding Statement:This work is supported in part by the National Natural Science Foundation of China under Grant 61702060.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design,analysis and interpretation of result: Jiayi Ma;draft manuscript preparation: Jiayi Ma,Tianhao Sun;data collection: Jiayi Ma,Xiaodong Zhang;All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The authors have shared the link to the data in the paper.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    一本精品99久久精品77| 亚洲成av人片在线播放无| 欧美一级毛片孕妇| 国产成年人精品一区二区| 亚洲美女视频黄频| 国产高清视频在线观看网站| 天堂√8在线中文| 男女做爰动态图高潮gif福利片| 美女大奶头视频| 国产精品免费一区二区三区在线| 嫩草影院入口| 国产精品永久免费网站| 在线观看舔阴道视频| 香蕉久久夜色| 成人三级做爰电影| 欧美丝袜亚洲另类 | 欧美黄色片欧美黄色片| ponron亚洲| 国产精品乱码一区二三区的特点| 一区二区三区国产精品乱码| 精品久久蜜臀av无| 两性夫妻黄色片| 成年人黄色毛片网站| 美女cb高潮喷水在线观看 | 色综合婷婷激情| 黑人操中国人逼视频| 亚洲最大成人中文| 久久精品亚洲精品国产色婷小说| 日本熟妇午夜| 我的老师免费观看完整版| 亚洲片人在线观看| 国产成人影院久久av| 国产v大片淫在线免费观看| 中文资源天堂在线| 亚洲国产精品sss在线观看| 成人特级av手机在线观看| 香蕉国产在线看| 国产在线精品亚洲第一网站| 国产亚洲精品久久久久久毛片| 国产高清激情床上av| 黄频高清免费视频| 狂野欧美白嫩少妇大欣赏| 免费看日本二区| 人妻丰满熟妇av一区二区三区| 免费看日本二区| 午夜精品一区二区三区免费看| 国产乱人伦免费视频| 亚洲中文字幕日韩| 日韩中文字幕欧美一区二区| 国产三级在线视频| 免费高清视频大片| 一夜夜www| or卡值多少钱| 精品免费久久久久久久清纯| 极品教师在线免费播放| 狂野欧美激情性xxxx| 搡老熟女国产l中国老女人| 国产乱人视频| 国产精品免费一区二区三区在线| 国产黄a三级三级三级人| 最新中文字幕久久久久 | 欧美午夜高清在线| 最近在线观看免费完整版| 黄色丝袜av网址大全| av福利片在线观看| 国产精品一及| 久久精品91无色码中文字幕| 啦啦啦观看免费观看视频高清| 亚洲 欧美 日韩 在线 免费| 99在线人妻在线中文字幕| 好男人在线观看高清免费视频| 日韩大尺度精品在线看网址| 国产精品影院久久| 成人午夜高清在线视频| 国产精品久久电影中文字幕| 久久香蕉精品热| 香蕉av资源在线| 十八禁人妻一区二区| 丰满人妻熟妇乱又伦精品不卡| 不卡av一区二区三区| 身体一侧抽搐| 久久99热这里只有精品18| 一进一出抽搐gif免费好疼| 少妇人妻一区二区三区视频| 天天一区二区日本电影三级| ponron亚洲| 香蕉av资源在线| 国产成人精品久久二区二区91| 亚洲成a人片在线一区二区| 欧美日韩国产亚洲二区| 国内揄拍国产精品人妻在线| 两性夫妻黄色片| 精品国内亚洲2022精品成人| 人人妻人人澡欧美一区二区| 日本a在线网址| 美女 人体艺术 gogo| 亚洲午夜理论影院| av中文乱码字幕在线| 久久久久久久午夜电影| 国产精品 国内视频| e午夜精品久久久久久久| 一级毛片高清免费大全| 国产爱豆传媒在线观看| 亚洲国产欧洲综合997久久,| 俺也久久电影网| av黄色大香蕉| 网址你懂的国产日韩在线| 国产精品亚洲一级av第二区| 国产精品影院久久| 成年女人永久免费观看视频| 国产精品久久视频播放| 91麻豆av在线| 国产精品亚洲美女久久久| 女人高潮潮喷娇喘18禁视频| 露出奶头的视频| 麻豆国产av国片精品| 久久久久精品国产欧美久久久| 国产伦一二天堂av在线观看| 成人特级黄色片久久久久久久| 国产又色又爽无遮挡免费看| 精品久久久久久久毛片微露脸| 男人的好看免费观看在线视频| 黄色片一级片一级黄色片| 免费在线观看日本一区| 小说图片视频综合网站| 成人三级黄色视频| 丰满人妻一区二区三区视频av | 婷婷亚洲欧美| 在线十欧美十亚洲十日本专区| 亚洲 欧美 日韩 在线 免费| 欧美午夜高清在线| 99久久国产精品久久久| 国语自产精品视频在线第100页| 国产91精品成人一区二区三区| 1024手机看黄色片| 亚洲av片天天在线观看| 九九久久精品国产亚洲av麻豆 | 国产精品国产高清国产av| 国产精品av视频在线免费观看| 十八禁人妻一区二区| 国产又黄又爽又无遮挡在线| 不卡av一区二区三区| 久久香蕉精品热| 日韩免费av在线播放| 国产伦精品一区二区三区视频9 | 国产三级在线视频| 人妻夜夜爽99麻豆av| 中亚洲国语对白在线视频| 免费大片18禁| 欧美黄色淫秽网站| 国产视频一区二区在线看| 19禁男女啪啪无遮挡网站| 国产亚洲欧美98| 婷婷精品国产亚洲av| 亚洲精品美女久久久久99蜜臀| 黄色片一级片一级黄色片| 免费在线观看视频国产中文字幕亚洲| 国产精品免费一区二区三区在线| 性欧美人与动物交配| 亚洲色图av天堂| 亚洲av免费在线观看| 国产精品美女特级片免费视频播放器 | 国产伦精品一区二区三区四那| 日本与韩国留学比较| 国产精品 国内视频| 亚洲欧美日韩高清在线视频| 亚洲国产精品sss在线观看| 丝袜人妻中文字幕| 欧美中文综合在线视频| 国产精品国产高清国产av| 亚洲成人精品中文字幕电影| 99国产综合亚洲精品| 国产久久久一区二区三区| 麻豆久久精品国产亚洲av| 成在线人永久免费视频| 精品久久久久久久久久免费视频| www日本黄色视频网| a在线观看视频网站| 真人做人爱边吃奶动态| 1024香蕉在线观看| 成年女人毛片免费观看观看9| 在线a可以看的网站| 99热只有精品国产| 十八禁人妻一区二区| e午夜精品久久久久久久| 一级毛片高清免费大全| 国产欧美日韩精品亚洲av| 日韩av在线大香蕉| 国产激情久久老熟女| 成人av一区二区三区在线看| 一个人观看的视频www高清免费观看 | 国产爱豆传媒在线观看| 欧美成人一区二区免费高清观看 | 成人午夜高清在线视频| 看片在线看免费视频| 97人妻精品一区二区三区麻豆| 桃红色精品国产亚洲av| 亚洲最大成人中文| 九九热线精品视视频播放| 日韩欧美在线二视频| 午夜视频精品福利| 欧美zozozo另类| 免费在线观看日本一区| 免费av毛片视频| 给我免费播放毛片高清在线观看| 精品一区二区三区av网在线观看| 亚洲专区国产一区二区| 一a级毛片在线观看| 免费搜索国产男女视频| 级片在线观看| 欧美日韩国产亚洲二区| 老汉色∧v一级毛片| 美女高潮喷水抽搐中文字幕| 岛国在线免费视频观看| 午夜激情福利司机影院| 99久久精品一区二区三区| 国产精品1区2区在线观看.| 中文字幕人成人乱码亚洲影| 黄频高清免费视频| 中出人妻视频一区二区| 国产毛片a区久久久久| avwww免费| 欧美黑人巨大hd| 成人18禁在线播放| 一级黄色大片毛片| 国产精品香港三级国产av潘金莲| 狠狠狠狠99中文字幕| 99热这里只有精品一区 | 女人高潮潮喷娇喘18禁视频| 一本一本综合久久| 岛国在线观看网站| av在线天堂中文字幕| 国产又黄又爽又无遮挡在线| 久久亚洲精品不卡| 性欧美人与动物交配| 999精品在线视频| av视频在线观看入口| 窝窝影院91人妻| 久久这里只有精品中国| www.www免费av| 国产高清videossex| 一个人看的www免费观看视频| 亚洲激情在线av| 夜夜夜夜夜久久久久| 亚洲中文字幕一区二区三区有码在线看 | 亚洲在线自拍视频| 丰满的人妻完整版| 婷婷丁香在线五月| 午夜福利18| 亚洲精品久久国产高清桃花| 最近最新中文字幕大全电影3| 激情在线观看视频在线高清| 美女午夜性视频免费| 免费人成视频x8x8入口观看| 国产单亲对白刺激| 757午夜福利合集在线观看| 久久久水蜜桃国产精品网| 99热这里只有精品一区 | tocl精华| 小说图片视频综合网站| 制服丝袜大香蕉在线| 看免费av毛片| 丰满人妻一区二区三区视频av | 制服人妻中文乱码| 午夜a级毛片| 亚洲人成伊人成综合网2020| 国产爱豆传媒在线观看| www.www免费av| 日本熟妇午夜| 综合色av麻豆| 夜夜躁狠狠躁天天躁| 国产精品香港三级国产av潘金莲| 少妇的丰满在线观看| 久久精品人妻少妇| 午夜精品久久久久久毛片777| 变态另类成人亚洲欧美熟女| 90打野战视频偷拍视频| 99久国产av精品| 日本熟妇午夜| 男女做爰动态图高潮gif福利片| 欧美成人一区二区免费高清观看 | 日日干狠狠操夜夜爽| 国产激情偷乱视频一区二区| 亚洲18禁久久av| 日韩欧美在线二视频| 亚洲欧美日韩无卡精品| 日本黄色片子视频| 最好的美女福利视频网| 很黄的视频免费| 人妻夜夜爽99麻豆av| 国产黄色小视频在线观看| 国产成人欧美在线观看| 免费在线观看视频国产中文字幕亚洲| 国内揄拍国产精品人妻在线| 最近最新免费中文字幕在线| 精品久久久久久久久久免费视频| 91麻豆精品激情在线观看国产| 欧美成人免费av一区二区三区| 又粗又爽又猛毛片免费看| 成年人黄色毛片网站| av天堂中文字幕网| 国产精品 国内视频| 国产精品亚洲美女久久久| 欧美一区二区国产精品久久精品| 国产综合懂色| 久久天堂一区二区三区四区| 天堂影院成人在线观看| 黄色女人牲交| 伦理电影免费视频| 国产精品久久久久久精品电影| 精品久久蜜臀av无| 少妇人妻一区二区三区视频| 少妇的丰满在线观看| 国产精品影院久久| 免费在线观看亚洲国产| 国产极品精品免费视频能看的| 哪里可以看免费的av片| 香蕉国产在线看| 在线观看舔阴道视频| 国产一区二区在线av高清观看| 久久国产精品人妻蜜桃| 亚洲av免费在线观看| 黄色丝袜av网址大全| 男人舔女人下体高潮全视频| 欧美极品一区二区三区四区| 欧美在线黄色| 国内精品久久久久久久电影| 手机成人av网站| 亚洲精品美女久久av网站| 欧美成人免费av一区二区三区| 欧美一区二区精品小视频在线| 国产伦人伦偷精品视频| 亚洲天堂国产精品一区在线| 精品人妻1区二区| 俺也久久电影网| 欧美精品啪啪一区二区三区| 88av欧美| e午夜精品久久久久久久| 国产成人av激情在线播放| 亚洲中文字幕一区二区三区有码在线看 | 亚洲欧美一区二区三区黑人| xxx96com| 精品久久久久久,| 亚洲人与动物交配视频| 日本成人三级电影网站| 欧美日韩乱码在线| 国产激情欧美一区二区| 啦啦啦免费观看视频1| 在线观看日韩欧美| 网址你懂的国产日韩在线| 97超视频在线观看视频| 99热6这里只有精品| 一个人看视频在线观看www免费 | 国产精品一区二区三区四区久久| 亚洲在线观看片| 宅男免费午夜| 99热这里只有是精品50| 九色成人免费人妻av| 99热精品在线国产| 国产私拍福利视频在线观看| 国产精品一区二区免费欧美| 哪里可以看免费的av片| 一级毛片高清免费大全| 国产精品av久久久久免费| 国产亚洲精品综合一区在线观看| 极品教师在线免费播放| 亚洲欧美日韩东京热| 亚洲欧美一区二区三区黑人| 精品福利观看| 麻豆av在线久日| 亚洲专区国产一区二区| 国产精华一区二区三区| 1000部很黄的大片| 欧美日韩一级在线毛片| 欧美+亚洲+日韩+国产| 午夜激情福利司机影院| 精品久久久久久,| 亚洲精品456在线播放app | 99热这里只有精品一区 | 国产单亲对白刺激| 男人舔女人下体高潮全视频| 国产高清videossex| 国产欧美日韩一区二区三| 精品久久久久久久久久久久久| 国产精品久久久久久精品电影| 中文在线观看免费www的网站| 黄片小视频在线播放| 国产又黄又爽又无遮挡在线| 国产欧美日韩精品亚洲av| 国产激情欧美一区二区| 国产精品精品国产色婷婷| 亚洲无线观看免费| 51午夜福利影视在线观看| 不卡av一区二区三区| 亚洲欧美日韩卡通动漫| 亚洲无线观看免费| 亚洲av成人av| 亚洲avbb在线观看| 18禁观看日本| 男人的好看免费观看在线视频| 亚洲专区字幕在线| 亚洲国产欧洲综合997久久,| 欧美黄色淫秽网站| 老汉色∧v一级毛片| 老鸭窝网址在线观看| 欧美日韩精品网址| 高清毛片免费观看视频网站| 一区二区三区高清视频在线| 国产视频内射| 精华霜和精华液先用哪个| 熟女少妇亚洲综合色aaa.| 久久精品夜夜夜夜夜久久蜜豆| 亚洲国产精品久久男人天堂| 男女做爰动态图高潮gif福利片| 免费看日本二区| 伊人久久大香线蕉亚洲五| 母亲3免费完整高清在线观看| 国产激情偷乱视频一区二区| 久久中文字幕人妻熟女| 老熟妇乱子伦视频在线观看| 国产成人影院久久av| 国产99白浆流出| 熟女人妻精品中文字幕| 人人妻人人澡欧美一区二区| 色av中文字幕| 99热只有精品国产| 亚洲中文日韩欧美视频| 日日干狠狠操夜夜爽| 亚洲专区国产一区二区| av女优亚洲男人天堂 | 亚洲avbb在线观看| 亚洲av免费在线观看| 国产av麻豆久久久久久久| 听说在线观看完整版免费高清| 中文字幕av在线有码专区| 亚洲中文av在线| 两个人看的免费小视频| 男女做爰动态图高潮gif福利片| 在线看三级毛片| 午夜视频精品福利| 毛片女人毛片| 此物有八面人人有两片| 色综合欧美亚洲国产小说| 他把我摸到了高潮在线观看| 国产探花在线观看一区二区| 国产精品爽爽va在线观看网站| 又黄又粗又硬又大视频| 一边摸一边抽搐一进一小说| 两个人的视频大全免费| 国产精品香港三级国产av潘金莲| 亚洲欧美精品综合久久99| 99久久精品热视频| 99精品在免费线老司机午夜| 免费大片18禁| 变态另类成人亚洲欧美熟女| 在线观看日韩欧美| 国产精品久久久av美女十八| 国产人伦9x9x在线观看| 欧美一区二区国产精品久久精品| 在线十欧美十亚洲十日本专区| 在线观看午夜福利视频| 亚洲 国产 在线| 久久草成人影院| 色综合亚洲欧美另类图片| 免费在线观看日本一区| 性色av乱码一区二区三区2| 一个人看的www免费观看视频| 给我免费播放毛片高清在线观看| 日本五十路高清| 国产视频内射| 国产成人系列免费观看| 美女高潮的动态| 村上凉子中文字幕在线| 两人在一起打扑克的视频| 欧美又色又爽又黄视频| 久99久视频精品免费| h日本视频在线播放| а√天堂www在线а√下载| 一级毛片精品| 亚洲av成人av| 天堂网av新在线| 亚洲av成人av| 久久久精品大字幕| 黄片小视频在线播放| 国产成人啪精品午夜网站| 啦啦啦免费观看视频1| 欧洲精品卡2卡3卡4卡5卡区| 哪里可以看免费的av片| 亚洲va日本ⅴa欧美va伊人久久| 中国美女看黄片| 国产乱人视频| 蜜桃久久精品国产亚洲av| 免费高清视频大片| 一本综合久久免费| 高清毛片免费观看视频网站| av中文乱码字幕在线| 色尼玛亚洲综合影院| 亚洲精品粉嫩美女一区| 成人高潮视频无遮挡免费网站| 国产亚洲av嫩草精品影院| 此物有八面人人有两片| 国产精品 欧美亚洲| 国产午夜精品久久久久久| 一个人看视频在线观看www免费 | 午夜视频精品福利| 精品日产1卡2卡| 亚洲18禁久久av| 亚洲成av人片在线播放无| a级毛片在线看网站| 18禁观看日本| 日韩欧美在线二视频| 国产一区二区在线av高清观看| 成人高潮视频无遮挡免费网站| 欧美成狂野欧美在线观看| 久久久国产欧美日韩av| 男女那种视频在线观看| 久久中文字幕人妻熟女| 一区二区三区激情视频| 亚洲一区高清亚洲精品| 亚洲精品色激情综合| 国产成+人综合+亚洲专区| 性色avwww在线观看| 亚洲欧美精品综合一区二区三区| 18禁黄网站禁片午夜丰满| 色噜噜av男人的天堂激情| xxx96com| 一夜夜www| 色哟哟哟哟哟哟| 91麻豆av在线| 国产黄a三级三级三级人| 岛国在线观看网站| 精品熟女少妇八av免费久了| 国产精品香港三级国产av潘金莲| 99久久精品一区二区三区| 99久久无色码亚洲精品果冻| 又粗又爽又猛毛片免费看| svipshipincom国产片| 久久久久久大精品| 黑人巨大精品欧美一区二区mp4| 亚洲人成伊人成综合网2020| 91麻豆精品激情在线观看国产| 黑人操中国人逼视频| 国产成人影院久久av| 国产野战对白在线观看| 男女那种视频在线观看| 51午夜福利影视在线观看| 两个人的视频大全免费| 精品久久久久久成人av| 久久久久久国产a免费观看| 亚洲av五月六月丁香网| 午夜激情欧美在线| 国产私拍福利视频在线观看| 无限看片的www在线观看| 久久久色成人| 最新中文字幕久久久久 | 97超视频在线观看视频| 一个人看视频在线观看www免费 | 波多野结衣巨乳人妻| 中亚洲国语对白在线视频| 99国产极品粉嫩在线观看| 黄色丝袜av网址大全| 亚洲九九香蕉| 亚洲精品在线美女| 成年女人毛片免费观看观看9| 天堂√8在线中文| 亚洲色图av天堂| 精品一区二区三区四区五区乱码| 国产精品,欧美在线| 成人永久免费在线观看视频| 久久午夜亚洲精品久久| 男女下面进入的视频免费午夜| 真人做人爱边吃奶动态| 99riav亚洲国产免费| 欧美日本亚洲视频在线播放| 91老司机精品| 69av精品久久久久久| 久久精品人妻少妇| 国产又黄又爽又无遮挡在线| 性色av乱码一区二区三区2| 国产精品综合久久久久久久免费| 亚洲精品乱码久久久v下载方式 | 久久久国产成人精品二区| 亚洲18禁久久av| 欧美不卡视频在线免费观看| 在线免费观看不下载黄p国产 | 看黄色毛片网站| 在线观看午夜福利视频| 亚洲国产精品999在线| 看免费av毛片| 午夜福利在线观看吧| 日韩欧美精品v在线| 看免费av毛片| 午夜福利在线观看吧| 天天躁狠狠躁夜夜躁狠狠躁| 99久久精品一区二区三区| 丝袜人妻中文字幕| 巨乳人妻的诱惑在线观看| 亚洲av中文字字幕乱码综合| 久久久久九九精品影院| 18禁黄网站禁片免费观看直播| 99国产精品99久久久久| a在线观看视频网站| 亚洲成人中文字幕在线播放| 看片在线看免费视频| 黄片大片在线免费观看| 日韩 欧美 亚洲 中文字幕| 精品久久久久久久人妻蜜臀av| 男人的好看免费观看在线视频| 99re在线观看精品视频| 亚洲成av人片在线播放无| 一本久久中文字幕| 亚洲真实伦在线观看| 伦理电影免费视频| 舔av片在线| 日本成人三级电影网站|