• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Self-supervised graph learning with target-adaptive masking for session-based recommendation

    2023-02-06 09:43:42YitongWANGFeiCAIZhiqiangPANChengyuSONG

    Yitong WANG,Fei CAI,Zhiqiang PAN,Chengyu SONG

    Science and Technology on Information Systems Engineering Laboratory,National University of Defense Technology,Changsha 410073,China

    Abstract:Session-based recommendation aims to predict the next item based on a user’s limited interactions within a short period.Existing approaches use mainly recurrent neural networks(RNNs)or graph neural networks(GNNs)to model the sequential patterns or the transition relationships between items.However,such models either ignore the over-smoothing issue of GNNs,or directly use cross-entropy loss with a softmax layer for model optimization,which easily results in the over-fitting problem.To tackle the above issues,we propose a self-supervised graph learning with target-adaptive masking(SGL-TM)method.Specifically,we first construct a global graph based on all involved sessions and subsequently capture the self-supervised signals from the global connections between items,which helps supervise the model in generating accurate representations of items in the ongoing session.After that,we calculate the main supervised loss by comparing the ground truth with the predicted scores of items adjusted by our designed target-adaptive masking module.Finally,we combine the main supervised component with the auxiliary self-supervision module to obtain the final loss for optimizing the model parameters.Extensive experimental results from two benchmark datasets,Gowalla and Diginetica,indicate that SGL-TM can outperform state-of-the-art baselines in terms of Recall@20 and MRR@20,especially in short sessions.

    Key words:Session-based recommendation;Self-supervised learning;Graph neural networks;Target-adaptive masking

    1 Introduction

    As an effective way to tackle the information overload issue,recommender systems(RSs)can help people make accurate choices when facing abundant products or services(Singhal et al.,2017;Wang XN and Tan,2020).Traditional approaches like collaborative filtering mainly pay attention to the long-term preference of users while neglecting their current interaction patterns(Wang X et al.,2019).This may lead to inaccurate recommendations in some realworld situations,where a user’s long-term interactions are not available(Wang SJ et al.,2022).Thus,session-based recommendation has been proposed to generate recommendations by modeling the user’s recent limited interactions(Hidasi et al.,2016;Wu et al.,2019).

    Existing methods apply mainly Markov chains(MCs)to extract the sequential patterns between adjacent interactions for session-based recommendation(Rendle et al.,2010).Moreover,considering the advantage of recurrent neural networks(RNNs)in sequential data modeling,Hidasi et al.(2016)first used the gate recurrent unit(GRU)to model the session sequences and proposed session-parallel mini-batches to modify the basic GRU.In addition,Li J et al.(2017)applied the attention mechanism to determine the importance of items,which was then used to generate the user’s main intent.Recently,benefiting from the capability of graph neural networks(GNNs)in modeling complex transition relations between items,GNN-based methods have been widely investigated and have achieved remarkable achievements on session-based recommendation(Chen TW and Wong,2020;Qiu et al.,2020).For instance,Wang ZY et al.(2020)introduced the global context enhanced GNN(GCE-GNN)to aggregate the item representations from both the session graph and the global graph through a soft attention mechanism.

    However,MC-and RNN-based methods fail to take the complicated transition patterns between items into account(Xia et al.,2021b).Furthermore,despite the fact that GNN-based methods have achieved outstanding success,the following limitations still remain.On one hand,deep GNNs may lead to an over-smoothing issue,in which the learned representations of adjacent items in the session are highly similar,making the item features indistinguishable(Chen M et al.,2020).On the other hand,the training objective of existing models is to minimize the cross-entropy loss.In the training process,the target item score will continually increase while other items,including the neighbors of the target item,are treated as negative samples with decreasing scores.However,items adjacent to each other in the graph should be similar;i.e.,it is unreasonable that the scores of the neighbors of the target item are continuously decreased.Thus,it leads to an over-fitting problem and model performance degradation.

    To tackle the aforementioned issues,in this study,we propose the self-supervised graph learning with target-adaptive masking(SGL-TM)method for session-based recommendation.Specifically,we first construct the global connections between the items of all involved sessions,and then obtain the neighbor and non-neighbor information of each item as the self-supervised signals for enhancing item representation.Subsequently,given a current session,we apply the graph attention network(GAT)to generate the representations of items in the session.After that,we combine the long-and short-term interests of the current session to obtain the user preference.Then,for each session,we employ the designed target-adaptive masking module to remove the neighbors of the target item for calculating the main supervised loss,which is then combined with the auxiliary self-supervised loss for model optimization.We implement the comprehensive experiments on two real-world datasets,Gowalla and Diginetica.The results demonstrate that our SGL-TM outperforms state-of-the-art baselines in terms of both Recall@20 and MRR@20.

    Generally,the main contributions of this paper can be summarized as follows:

    1.We propose a neoteric self-supervised framework for session-based recommendation,which can solve the over-smoothing problem using the global connections between items in graph learning.

    2.We design a target-adaptive masking module to tackle the over-fitting issue resulting from the cross-entropy loss with a softmax layer in the model optimization.

    3.The results of the experiments implemented on two public,real-world datasets,Gowalla and Diginetica,indicate that our proposed SGL-TM outperforms the competitive baselines in terms of Recall@20 and MRR@20.

    2 Related works

    In this section,we introduce the related works from two aspects,session-based recommendation and self-supervised recommendation.

    2.1 Session-based recommendation

    Conventional approaches for session-based recommendation typically use Markov chains to capture the temporal patterns between the user’s adjacent interactions.For instance,Rendle et al.(2010)combined matrix factorization(MF)with Markov chains to simultaneously capture the sequential patterns and the user’s long-term preference in the session.Moreover,Shani et al.(2005)adopted Markov decision processes to use the transition dependencies between items for generating recommendations.

    Recently,neural models such as convolutional neural networks(CNNs)and RNNs have been widely used in session-based recommendation.For instance,Yuan et al.(2019)adopted CNNs to learn item representations from both long-term and current item dependencies,where the interactions in a session were regarded as an image and the convolution operation was applied to generate the item embeddings.Moreover,Hidasi et al.(2016)introduced GRU4REC,which first employs RNNs to model the whole session sequence and introduces a session-parallel training mechanism for model optimization.Furthermore,Tan et al.(2016)proposed multiple optimization techniques for enhancing GRU4REC,such as data augmentation and generalized distillation,to make item prediction accurate.In addition,to emphasize the user’s main intent in the current session,Li J et al.(2017)introduced the neural attentive recommendation model based on a hybrid encoder to concurrently take the sequential behavior and the user’s main intent into consideration.Besides,Liu et al.(2018)proposed a short-term memory priority model for capturing both the long-and short-term interests of the user.

    Considering the powerful ability of GNNs in modeling connections between items,GNNs have been introduced into session-based recommendation and achieved considerable performance.Specifically,the current session is first transformed to a session graph,where GNNs are applied to propagate information between items for learning accurate item representations.For example,Wu et al.(2019)introduced session-based recommendation with GNNs(SR-GNN),which took the lead in adopting GNNs to learn item embeddings,where incoming and outgoing matrices were adopted to reveal the transition pattern between items.On the basis of SR-GNN,Xu et al.(2019)further used a self-attention mechanism to extract the long-term patterns between items in the session.Furthermore,Qiu et al.(2020)employed FGNN+to connect different sessions using a global graph for capturing cross-session information to generate accurate recommendations.Moreover,Wang ZY et al.(2020)proposed to compute the information flow from both the global and session levels by relying on the constructed global graph and session graph,respectively.In addition,Chen TW and Wong(2020)proposed a shortcut graph attention layer with a lossless encoding scheme to tackle the information loss issue of GNN applications in session-based recommendation.

    However,the CNN-and RNN-based approaches fail to take the complicated transition patterns between items into account,limiting the recommendation performance.Moreover,although the GNNbased methods have achieved outstanding achievements in session-based recommendation,they generally face a serious over-smoothing issue.That is,GNNs lead to significant similarities among the item representations learned in a session,which results in inaccurate recommendations.

    2.2 Self-supervised recommendation

    Self-supervised learning(SSL)(Hjelm et al.,2019)is a neoteric paradigm of machine learning that mines the pseudo labels from data to supervise models and uses the learned representations for various downstream tasks.SSL has achieved great accomplishments in natural language processing(Kong et al.,2020;Zhang et al.,2020),computer vision(Afouras et al.,2020;Chen T et al.,2020),etc.

    Recently,SSL has also been used in recommender systems to generate accurate recommendations.For instance,Xie et al.(2022)designed contrastive learning for sequential recommendation(CL4SRec),in which three data enhancement methods were proposed to extract self-supervised signals from the user’s behavior to generate the user representation.Xin et al.(2020)used reinforcement learning to extend existing sequential recommendation models,aiming to distinguish different types of user-item interactions.Moreover,Yao et al.(2021)solved the issue of data sparsity in large-scale item recommendations by applying a new data augmentation method and contrastive learning to enhance item representations.Zhou et al.(2020)introduced S3-Rec(self-supervised learning for sequential recommendation),which learns the correlations among sequences,subsequences,items,and attributes based on mutual information maximization(MIM)to obtain self-supervised signals and to strengthen the item representations through pre-training for recommendation.Furthermore,Xia et al.(2021b)designed a novel dual channel hypergraph convolutional network(DHCN)to obtain the complex high-order information between items,in which SSL was used to maximize the mutual information between the session representations generated through two channels.In addition,Xia et al.(2021a)built graph encoders to simultaneously leverage the internal and external connectivity information of sessions,which was then used to tackle the problem of data sparsity by contrastive learning.

    Although the aforementioned self-supervised approaches have made significant improvements in recommender systems,they consider merely the supervised signals from a session-item perspective or a session-session perspective,while generally ignoring the connections between items in different sessions,which can provide additional supervisory information.In this study,we use the global-level connections between items in the global graph to enhance the item representation learning on the basis of the main supervised learning.

    3 Approach

    In this section,we first formalize the task of session-based recommendation.Subsequently,we introduce the proposed SGL-TM in detail,which consists mainly of three components(user preference generation,self-supervised learning,and model optimization).

    Session-based recommendation aims to recommend the next item with which a user will interact based on the limited interactions in an ongoing session.DenoteV={v1,v2,...,v|V|}the item set,which contains all items(|V|represents the number of items inV).Assume thatU={S1,S2,...,S|U|}indicates all sessions,where|U|represents the number of sessions.DenoteSi={v1,v2,...,vn}theithsession inU,which includesnitems in chronological order,andvt(t=1,2,...,n)represents the item with which interaction occurs at timestamptinSi.Given a current sessionSi,we first generate the user preference,which is applied to obtain the prediction scores of all items.Eventually,items ranked in the topKpositions will form the recommended list for the user.

    Fig.1 presents the framework of our proposed SGL-TM model for session-based recommendation.Given a current session,we first learn the item embeddings by the information propagation of GAT.Then we obtain the user’s long-term and current interests in the session,and concatenate them to generate the final session representation.Specifically,we establish a global graph to reveal the transition relations between items among different sessions according to all sessions in the training set.Then,on one hand,we obtain the self-supervised signals from the global-level connections between items to calculate the self-supervised loss.On the other hand,we calculate the prediction scores of all items in the candidate set,which are then adjusted by the designed target-adaptive masking module to obtain the cross-entropy loss as the main supervised loss.Finally,the main supervised loss and self-supervised loss are combined to generate the final loss for model optimization.

    Fig.1 Framework of self-supervised graph learning with target-adaptive masking(SGL-TM)(GAT:graph attention network)

    3.1 User preference generation

    3.1.1 Item representation learning

    For a sessionSi,we first establish a directed local graph to reflect the transition relationships between items inSi.Specifically,Sican be denoted aswhereanddenote the node set and edge set in the local graph,respectively.Furthermore,={x1,x2,...,xm}comprises all the unique items inSi,wherem≤nbecause items may be repeatedly interacted in the session.Each edgeeij∈indicates that the user interacts withxibeforexjin the session.

    After constructing the local graph,we conduct information propagation onto learn the item representations.Here we adopt the GAT(Veli?ckovi? et al.,2018),considering that compared to gated graph neural networks(GG-NNs)(Li YJ et al.,2016)and graph convolutional networks(GCNs)(Kipf and Welling,2017),GAT can dynamically determine the different importance of its neighbors to each node in the local graph.Specifically,we first initialize the embedding vectorxiof each nodexiin?Gias follows:

    in which“Embedding”indicates the embedding layer,x0i∈Rddenotes the initialized item embeddings ofxi,anddrepresents the dimension of embeddings.

    To obtain the pairwise transition relationships between items in the ongoing session,we adopt GAT to propagate neighbor information for each node.For each nodexiat layerl,the importance degree of different neighbor nodes is determined by employing a self-attention mechanism,where the importance between nodesxiandxjis denoted by an attention coefficient calculated as follows:

    whereW0∈R2dandW1,W2∈Rd×drepresent the trainable parameters,[;]denotes the concatenation operation,andσ1indicates the Leaky ReLU function.

    Next,the attention coefficients are normalized via using a softmax layer:

    in whichNviindicates the neighbor set ofviin the local graph?Gi.

    Finally,a linear combination of the corresponding neighbors is used to update the node vector ofviusing the attention coefficients:

    whereσandW3∈Rd×ddenote the sigmoid function and learnable parameters,respectively.

    After the information propagation on multilayer GATs,we obtain the item representationsxiin the local graph?Gi.Finally,the item sequence is recovered from the local graph to gain the representations of the chronological items within the current session as{z1,z2,...,zn}.

    3.1.2 Session representation generation

    After learning the item representations in the current session,following previous works(Xu et al.,2019;Qiu et al.,2020),we take the user’s long-term and current interests into account to generate the final session representation as the user preference.

    Due to that the last interacted item in the session can reflect the current interest of the user(Qiu et al.,2019;Pan et al.,2020),we directly use the embeddings of the last item as the current preferenceuc∈Rd,i.e.,uc=zn.In addition,the long-term preference of the user is obtained by aggregating the item representations using an attention mechanism:

    whereulindicates the long-term preference of the user,andW4∈RdandW5,W6∈Rd×ddenote the trainable parameters.

    Eventually,we attain the session representation by combining the long-term and current interests of the user as follows:

    whereW7∈Rd×ddenotes the trainable parameters.

    3.2 Self-supervised learning

    Existing session-based recommendation methods generally ignore the over-smoothing issue of GNNs,resulting in the indistinguishability of the learned item representations.Thus,in this subsection,we employ self-supervised learning to mine the informative signals about item transitions among different sessions from an item-item perspective.This enhances item representation learning to prevent the over-smoothing issue,which includes global graph construction and self-supervision.

    3.2.1 Global graph construction

    For the sake of obtaining the pairwise transition relationship between items across different sessions,we construct a global session graph,where the self-supervised signals generated from the global connections are taken into consideration.Assume thatrepresents the global graph involving the global connections,in whichreflects the unique items in all sessions,namelyV,andindicates all the item transitions in the training set.Each edgeeij∈denotes that itemsviandvjare adjacently interacted within a session.

    Moreover,for popular items,there may exist a large number of neighbor items.Here,to avoid the bias caused by accidental clicks,we employ max sampling to chooseMmost associated items as the ultimate neighbors for each node in the global graph,where the relevance is determined by the edge weights,which indicates the level of similarity between two nodes.By max sampling,loose connections between nodes are filtered out to make sure that the final neighbors are highly related to the target node.

    3.2.2 Self-supervision

    After constructing the global graph,we can obtain the self-supervised signals through the connections between items among different sessions.Specifically,given an itemviin a session,the set of its neighbors in the global graph is denoted asNvi.Forvi,the embeddings learned by GNNs should be more similar to its neighbor nodevj(vj∈Nvi)than a non-neighbor node(∈VNvi),where“”represents the set subtraction operation.Based on the intuition,we employ the Jensen-Shannon divergence(JSD)(Hjelm et al.,2019)as the contrastive loss,which aims to minimize the distance betweenviand its neighborsNvi,and maximize the distance betweenviand its unconnected itemsVNvi.The self-supervised loss can be calculated as

    whereziis the item representation ofvioutput by GNNs,andvjandare the item vectors of nodesvj(vj∈Nvi)and(∈VNvi)generated by the embedding layer respectively.Moreover,Tindicates the number of neighbors or non-neighbors of,which is sampled from the neighborsNviin the global graph using max sampling,or from the unconnected itemsVNviby random sampling,respectively.In addition,fis a linear mapping function calculated as

    f:Rd×Rd→R is the similarity calculation function,which uses the embeddings of two items as inputs and subsequently scores the similarity between them.

    This means that each item in the current session can obtain information from its neighbor items in the global graph by contrastive learning,which can be used as additional information to enhance item representation learning to solve the over-smoothing issue caused by using GAT in the current session.

    3.3 Model optimization

    3.3.1 Target-adaptive masking

    After obtaining the session representationu,the final recommendation probability of each candidate item can be generated by multiplyinguwith the embeddings of the candidates as follows:

    In addition,because directly adopting the crossentropy loss with a softmax layer can cause overfitting(that is,similar to the unconnected items,the scores of the neighbors of the target item are continuously decreased),we design a target-adaptive masking component.Specifically,given a global graphGg={Vg,Eg},for the target itemvtarget,we first sample its neighbors withNlargest edge weights in the global graph as the final neighbors ofvtarget,denoted asNvtarget.Then we mask the sampledNitems from the original candidate setVas the updated candidate set:

    Eventually,the cross-entropy is applied as the main supervised loss as follows:

    in whichyidenotes the one-hot vector of the ground truth.

    In this way,the scores of the target item’s neighbors will not be decreased in the training procedure,thus alleviating the over-fitting issue to a certain extent.

    3.3.2 Multi-task learning

    After generating the main supervised loss in Eq.(11)and the auxiliary self-supervised loss in Eq.(7),we obtain the total loss by combining the main and self-supervisions losses as follows:

    in whichλis the trade-offparameter controlling the magnitude of self-supervised learning.Finally,we update the item embeddings and learn the trainable parameters using the back propagation algorithm(Rumelhart et al.,1986).

    We detail the main process of SGL-TM training in Algorithm 1.For all sessions in the training set,we establish a global graph to show the connections of items among different sessions(line 1).Then,in the current session,item representations are generated(lines 4 and 5),that is,local graph construction and information propagation by GAT.We generate the current interest of the ongoing session(line 6)and long-term interest(line 7),and combine them to obtain the session representation(line 8).We perform max sampling on the neighbors and random sampling on the non-neighbors ofvi(line 9),and calculate the similarity according to their item embeddings to generate the self-supervised loss(line 10).The prediction score is calculated(line 11)and the target-adaptive masking on the candidate set is performed(line 12).We calculate the main supervised loss by a cross-entropy loss(line 13).Finally,we obtain the total loss and employ back propagation to optimize the model parameters.

    Algorithm 1 Training procedure of SGL-TM Input:set U=■S1,S2,...,S|U|■;λ:the trade-offparameter Output:I:the item representations in V;ψ:the trainable parameters in SGL-TM 1:?G={?V,?E}←global graph construct(U)2:for each training iteration do 3: for Si in U do 4: ?Gi={?Vi,?Ei}←local graph construct(Si)5: zi=information propagation(vi)based on Eqs.(1)-(4)6: uc=current interest(zn)7: ul=long-term interest(zi)based on Eq.(5)8: u=session representations(uc,ul)based on Eq.(6)9: Nvi,VNvi=sample(?G)10: Lself=JSD(zi,vj,ˉvj)based on Eqs.(7)and(8)11: ?yi=prediction(u,vi)based on Eq.(9)12: VUpdate=global graph-aware masking(V,Nvtarget)based on Eq.(10)13: Lmain=cross-entropy loss(?yi)based on Eq.(11)14: Model optimization loss:L=Lmain+λLself 15: end for 16: Use back propagation for the optimization of trainable parameters 17:end for 18:return I and ψ

    4 Experiments

    4.1 Research questions

    We guide our experiments by addressing the following five research questions:

    1.Can our SGL-TM model be superior to stateof-the-art baselines on real-world datasets?

    2.How significant is the contribution of selfsupervised learning and target-adaptive masking modules to the performance of SGL-TM?

    3.How does SGL-TM perform with various magnitudes of self-supervised signals incorporated?

    4.How does SGL-TM perform at different session lengths?

    5.What are the efforts of the important hyper-parameters,i.e.,the number of neighbor and non-neighbor pairs and the number of neighbors masked,on the performance of SGL-TM?

    4.2 Datasets

    We employ two real-world datasets that are extensively employed in session-based recommendations,i.e.,Gowalla(https://snap.stanford.edu/data/loc-gowalla.html)and Diginetica(http://cikm2016.cs.iupui.edu/cikm-cup),to conduct the experiments.Gowalla is a check-in dataset that contains the location information shared by users when they checked in between February 2009 and October 2010(Davidson et al.,2010).Following Tang and Wang(2018),we retain 30 000 of the most popular locations for the experiments.Diginetica collects users’behavior information from an e-commerce website,which is released by CIKM Cup 2016(Cheng et al.,2017).Here we use the transaction data,which are appropriate for session-based recommendation.Moreover,following Pan et al.(2022),items that occur less than 5 times and sessions with length 1 are removed from both datasets.Detailed statistics of the two datasets after prepossessing are presented in Table 1.

    Table 1 Dataset statistics

    4.3 Baselines

    We compare the performance of our model against the following representative baselines:

    FPMC(Rendle et al.,2010)is a compound method based on personalized Markov chains,which is applied in capturing the sequential patterns between adjacent items.

    NARM(Li J et al.,2017)employs an RNN and an attention mechanism to model the sequential signals in the session and captures the main intent of the user.

    NextItNet(Yuan et al.,2019)adopts a CNN to model the session sequence by learning high-level representations from item dependencies.

    FGNN(Qiu et al.,2019)adopts a Readout function and a weighted attention graph layer to learn the representations of sessions and items.

    SR-GNN(Wu et al.,2019)converts each session into a graph and generates item representations using GG-NNs.

    GCE-GNN(Wang ZY et al.,2020)employs both global-and local-level pairwise transitions between items to obtain the user preference.

    S2-DHCN(Xia et al.,2021b)introduces two hypergraphs to extract the beyond-pairwise relations of items and adopts self-supervised learning to enhance item representation.

    4.4 Evaluation metrics

    We employ Recall@Kand MRR@Kto assess the performance of recommendation models,which are always used in previous works(Qiu et al.,2019;Choi et al.,2021).

    Recall@K:It is an indicator for measuring whether the next-click item appears in the topKpositions of the recommended list,which is formulated as follows:

    in whichN1denotes the number of all test sequences andnhitindicates the number of next-click items in the topKpositions of the recommended list.

    MRR@K:This is a metric that takes the rank of a target item in the recommended list into consideration.It is set to 0 when the correct recommended item is not returned in the topKpositions of the recommended list;otherwise,it is generated as follows:

    in whichN2andvcorrectindicate the number of all test sessions(|Stest|)and the target item of a session,respectively.

    4.5 Experimental setup

    We divide Gowalla and Diginetica into the training and test sets for training and evaluation,respectively.For the Gowalla dataset,we employ the sessions that occurred in the last week for testing.For the Diginetica dataset,the last 20% of the sessions are separated as the test set.

    Moreover,following Chen TW and Wong(2020),we use the data augmentation strategy for both datasets.Adam(Kingma and Ba,2015)is adopted as our optimizer to train the model.We set the dimension of the item embeddings and the mini-batch size to 128 and 512,respectively.Moreover,the learning rate is initialized at 0.001 and attenuated by 0.5 every three epochs.In addition,the recommendation numberKis set to 20 for evaluation.Furthermore,we use grid search to find the best performance of SGL-TM on both datasets by ranging the dropout and layer numbers in{0,0.1,0.25}and{1,2,3},respectively.Similarly,we employ grid search to determine the number of neighbor and nonneighbor pairs in the self-supervised learning module(i.e.,M)and the number of neighbors masked in the target-adaptive masking component(i.e.,N)by rangingMandNboth in{1,2,3,4}.

    5 Results and discussion

    5.1 Overall performance

    We compare SGL-TM with state-of-the-art recommendation baselines in terms of Recall@20 and MRR@20 on both datasets.The experimental results are displayed in Table 2,from which we can draw the following observations:

    Table 2 Model performance

    First,compared with the neural models,the traditional recommendation methods like FPMC,based on Markov chains,have completely lost the superior-ity.This is due to the fact that MC-based methods pay more attention to the pairwise sequential signals between adjacent items,and ignore other sequential information in the whole session.Moreover,it is obvious that the performance of NARM on the two datasets is superior to that of NextItNet,because the session sequence in the session-based recommendation is generally generated within a short time and thus is likely to be time-dependent.Compared with the CNN used in NextItNet,RNN applied in NARM is better at modeling such time-dependent sequences.In addition,among the baselines,the GNN-based methods achieve the best performance,which suggests that GNNs can precisely model the complicated transition patterns between items in the session.Moreover,GCE-GNN andS2-DHCN outperform SR-GNN in most cases on both datasets,which proves that capturing both local and global levels of item information can help users make accurate predictions.However,GCE-GNN andS2-DHCN lose the competition to SR-GNN in terms of MRR@20 on the Gowalla dataset,which may be a result of the interference of irrelevant items in other sessions making them unable to accurately identify the user’s intention.

    Generally,our method SGL-TM is superior to the competitive baselines in terms of both Recall@20 and MRR@20 on both datasets,revealing its effectiveness for the session-based recommendation task.The reasons can be summarized as follows:The self-supervised learning module can employ abundant information from other sessions,which can help the model generate accurate item representations by avoiding the over-smoothing issue caused by GNNs.In addition,the target-adaptive masking module can effectively avoid the phenomenon that the scores of items adjacent to the target item are decreasing in a manner comparable to those of other unrelated items in the training process.Moreover,SGL-TM gains the improvements over the best baselines,namely,SR-GNN andS2-DHCN,by 0.58% in terms of Recall@20 and 4.29%in terms of MRR@20 on Gowalla,respectively;while the corresponding improvements are 0.84% and 2.62% on the Diginetica dataset.It can be found that the improvement of our SGL-TM on MRR@20 is more obvious than that on Recall@20 on both datasets,which suggests that compared to hitting the target item in the recommended list,our SGL-TM is more efficient at ranking it in the correct location.

    5.2 Ablation study

    In this subsection,we develop three variants of SGL-TM(i.e.,Base,Base-TM,and Base-SSL)to explore the contribution of each component in our proposal by comparing the variants to SGL-TM on the Gowalla and Diginetica datasets.Specifically,the variant Base merely adopts GAT to model the session sequence without self-supervised learning or the target-adaptive masking module.Base-TM and Base-SSL remove the self-supervised learning module and the target-adaptive masking module from SGL-TM,respectively.The results of the variants and SGL-TM are presented in Table 3.

    According to Table 3,it can be found that both self-supervised learning and target-adaptive masking modules contribute to the improvement of model performance.Moreover,removing the self-supervised learning module results in a larger degradation of the model performance than removing the target-adaptive masking.This may be because,compared to the over-fitting problem,the over-smoothing issue of the model is more prominent and has a greater effect on the recommendation accuracy.In addition,comparing Base-TM and SGL-TM,we find that after getting rid of the selfsupervised learning module,the model performance obviously drops,which shows that the global-level connections between items reflected in different sessions can provide additional self-supervisions in addition to the supervised signals in the session-item perspective for item representation learning.Moreover,compared with SGL-TM,the model performance of Base-TM is reduced by 1.79% and 3.75% in terms of Recall@20 and MRR@20 on Diginetica,while the performance dropping rates are 1.24%and 3.91%on the Gowalla dataset.It is evident that the contribution of the self-supervised learning module on ranking the target items at the correct location is larger than that on hitting them in the recommended list.

    Table 3 Ablation study

    5.3 Impact of the magnitude of selfsupervision

    We introduce a trade-offparameterλin Eq.(12)in our proposal to control the magnitude of incorporating the self-supervised signals.Specifically,we test the performance of SGL-TM by tuning the parameterλin{0,0.0001,0.001,0.01,0.1,0.2,0.5,1.0,2.0,3.0}to investigate the impact of the self-supervised learning module.The results are presented in Fig.2.

    Fig.2 Impact of the magnitude of self-supervision

    Based on Fig.2,the introduction of selfsupervised signals with an appropriateλeffectively improves the performance of recommender systems.This is due to the fact that the self-supervised learning module alleviates the over-smoothing problem and introduces the global-level information from other sessions to generate accurate item representations.In detail,for the Diginetica dataset,asλgoes up,the performance of SGL-TM in terms of both metrics,first exhibits an increasing trend and then decreasing,where the best performance is achieved whenλis 0.5.This is because whenλis small,the self-supervised signals are insufficient to solve the over-smoothing issue,while a largeλmay cause the over-fitting problem and thus lead to relatively low model performance.For the Gowalla dataset,whenλis within 1,asλincreases,the performance of SGL-TM first increases and then gently fluctuates in terms of both Recall@20 and MRR@20.Whenλis large enough on the Gowalla dataset(e.g.,2 or 3),the performance of SGL-TM begins to decline.Whenλis within 1,the different trends of the model performance on Diginetica and Gowalla may be made clear by the fact that the average number of clicks per item on Gowalla is larger than that on Diginetica.This indicates that the degree of items in the Gowalla dataset is correspondingly higher,so the Gowalla dataset is more likely to suffer from a serious over-smoothing problem.Thus,more selfsupervised signals are required to tackle the problem on Gowalla(Chen M et al.,2020;Nie et al.,2020,2021).

    5.4 Impact of session length

    Considering that sessions have various lengths in real-world situations,it is necessary to assess the performance of SGL-TM as well as the baselines when dealing with sessions with various lengths.Specifically,we separate the sessions in Gowalla and Diginetica into three groups with different lengths:“Short,”containing sessions of less than 6 items;“Long,”including sessions of more than 10 items;and the others denoted as“Medium.”Subsequently,we examine the performance of SGL-TM and the competitive baselines,namely,SR-GNN,GCE-GNN,andS2-DHCN,on the three groups of sessions on both datasets.The outcomes are exhibited in Fig.3.

    Fig.3 Impact of the session length on Gowalla(a)and Diginetica(b).References to color refer to the online version of this figure

    According to Fig.3,SGL-TM outperforms the baselines in terms of both Recall@20 and MRR@20 at different session lengths,demonstrating the effectiveness of our proposal on dealing with sessions of various lengths.Especially on the Gowalla dataset,the performance of our proposal in terms of both metrics improves when the session length increases.However,the performance of SGL-TM shows an opposite trend with the session length increasing on Diginetica.We analyze the phenomenon that is caused by the difference between the characteristics of the datasets.Specifically,for Gowalla under the check-in scenario,users generally pay attention to similar places.Thus,with the number of interactive items increasing,more information can be provided to determine the user’s purpose.Differently,for the Diginetica dataset,as the number of interactive items goes up,the user preference may become difficult to capture considering that the intent changes rapidly in the e-commerce scenario.In addition,it is observed that SGL-TM gains the most obvious improvements over the competitive baselines in terms of both metrics on the Short sessions among the three groups on both datasets,and the performance gap between SGL-TM and the baselines gradually decreases as the session length increases,which reveals that our proposal can effectively capture the user intent from limited historical interactions.

    5.5 Hyper-parameter analysis

    We examine the influence of some important hyper-parameters(includingMandN,which are introduced in the self-supervised learning and targetadaptive masking components,respectively)on the performance of SGL-TM.

    5.5.1 Impact of hyper-parameterM

    To illustrate the influence of the number of neighbor and non-neighbor pairs in the selfsupervised learning module(i.e.,M)on the recommendation accuracy,we searchMin{1,2,3,4}.The results of SGL-TM on Gowalla and Diginetica are presented in Fig.4.

    On the Diginetica dataset,from Fig.4,we find that the peak performance of SGL-TM is achieved whenMis 2.With the number of item pairs increasing,the performance of SGL-TM first increases and then decreases.This could be explained by adding a proper number of item pairs because the self-supervised signals can improve the model performance;however,introducing too many item pairs will result in self-supervision redundancy,which leads the model to over-fitting.Interestingly,on Gowalla,the performance of SGL-TM continually improves as the number of item pairs increases.The phenomenon is consistent with that mentioned in Section 5.3 and can be explained by the fact that the Gowalla dataset needs more self-supervised signals than Diginetica.

    Fig.4 Impact of M on both datasets

    5.5.2 Impact of hyper-parameterN

    We study the influence of the number of neighbors masked in the target-adaptive masking component,i.e.,N.Specifically,we tuneNin{1,2,3,4},and the results on Gowalla and Diginetica are presented in Fig.5.We can observe that on both the Gowalla and Diginetica datasets,SGL-TM achieves the peak performance whenNis 4.AsNincreases,the performance of SGL-TM on both Gowalla and Diginetica in terms of both metrics generally increases,which indicates that as the number of masked neighbors of the target item increases,our target-adaptive masking module can more effectively tackle the over-fitting problem.

    Fig.5 Impact of N on both datasets

    6 Conclusions and future work

    In this paper,we have proposed a self-supervised graph learning with target-adaptive masking(SGLTM)method for session-based recommendation.Specifically,we have employed self-supervised learning to tackle the over-smoothing issue of GNNs by introducing the global-level transition relations between items.Moreover,we have designed a targetadaptive masking module to effectively overcome the over-fitting problem of a cross-entropy loss with a softmax layer.The results of the experiments implemented on two real-world datasets,namely,Gowalla and Diginetica,have verified the effectiveness of our SGL-TM in terms of Recall@20 and MRR@20.

    As for future work,we plan to investigate the similarity among sessions to capture more selfsupervised signals to generate accurate recommendations.In addition,we would like to tackle the cold-start issue in recommender systems by mining the user-item interaction information in the session graph from multiple perspectives(Zheng et al.,2021).

    Contributors

    Yitong WANG designed the research and drafted the paper.Chengyu SONG made major revisions to the paper.Fei CAI and Zhiqiang PAN further modified and finalized the paper.

    Compliance with ethics guidelines

    Yitong WANG,Fei CAI,Zhiqiang PAN,and Chengyu SONG declare that they have no conflict of interest.

    Data availability

    The data that support the findings of this study are available from the corresponding author upon reasonable request.

    精品国产乱码久久久久久小说| 中文字幕人妻丝袜一区二区| 国产精品 国内视频| 操美女的视频在线观看| 精品少妇内射三级| 涩涩av久久男人的天堂| 中文字幕另类日韩欧美亚洲嫩草| 久久久久国产一级毛片高清牌| 亚洲中文日韩欧美视频| 久久久久久久国产电影| 国产深夜福利视频在线观看| 国产在线免费精品| 少妇人妻久久综合中文| 我要看黄色一级片免费的| 老熟女久久久| 免费观看人在逋| 国产成人免费观看mmmm| 在线十欧美十亚洲十日本专区| 91成人精品电影| 久久女婷五月综合色啪小说| 日韩制服骚丝袜av| 中文字幕av电影在线播放| 在线观看www视频免费| kizo精华| 免费高清在线观看视频在线观看| 精品一区二区三区av网在线观看 | 在线十欧美十亚洲十日本专区| 男女午夜视频在线观看| 在线看a的网站| 嫩草影视91久久| 国产成人系列免费观看| 亚洲,欧美精品.| 国产亚洲欧美在线一区二区| 久久久久久久久久久久大奶| 后天国语完整版免费观看| 国产欧美日韩综合在线一区二区| e午夜精品久久久久久久| 久久久精品94久久精品| 精品少妇久久久久久888优播| 在线 av 中文字幕| 久热这里只有精品99| 欧美精品人与动牲交sv欧美| 久久综合国产亚洲精品| 成年av动漫网址| 五月开心婷婷网| 99热国产这里只有精品6| 亚洲色图综合在线观看| 韩国精品一区二区三区| 亚洲欧美一区二区三区黑人| 欧美另类一区| 丝袜在线中文字幕| 国产一卡二卡三卡精品| videosex国产| 国产精品 国内视频| 精品人妻一区二区三区麻豆| 999久久久国产精品视频| 国产野战对白在线观看| 一区二区日韩欧美中文字幕| 国产伦理片在线播放av一区| 99精国产麻豆久久婷婷| 亚洲天堂av无毛| 高清视频免费观看一区二区| 亚洲精品美女久久av网站| 国产三级黄色录像| 天天添夜夜摸| 国产伦人伦偷精品视频| 久久精品亚洲av国产电影网| 亚洲国产欧美在线一区| 91精品国产国语对白视频| 亚洲精品成人av观看孕妇| 国产成人系列免费观看| 91成年电影在线观看| 欧美日韩成人在线一区二区| 老司机午夜福利在线观看视频 | videosex国产| cao死你这个sao货| 亚洲人成电影免费在线| 国产欧美日韩一区二区精品| 成人三级做爰电影| 97人妻天天添夜夜摸| 欧美变态另类bdsm刘玥| 亚洲人成电影免费在线| 在线 av 中文字幕| 另类亚洲欧美激情| 黄色片一级片一级黄色片| 大片免费播放器 马上看| 女人高潮潮喷娇喘18禁视频| cao死你这个sao货| 青青草视频在线视频观看| 老司机靠b影院| 国产一区二区激情短视频 | 国产成人系列免费观看| 欧美日韩成人在线一区二区| 99久久综合免费| 国产亚洲精品第一综合不卡| 麻豆av在线久日| www.熟女人妻精品国产| 国产在视频线精品| 亚洲精品久久久久久婷婷小说| 久久精品成人免费网站| 国产一区二区三区av在线| 成人黄色视频免费在线看| 青春草亚洲视频在线观看| 亚洲精品中文字幕在线视频| 91av网站免费观看| 精品第一国产精品| 国产一级毛片在线| 日本a在线网址| 亚洲欧洲日产国产| 999久久久国产精品视频| 国产精品自产拍在线观看55亚洲 | 纵有疾风起免费观看全集完整版| 亚洲精品久久午夜乱码| 好男人电影高清在线观看| 日韩欧美免费精品| 久久热在线av| videos熟女内射| 天天操日日干夜夜撸| 亚洲欧洲精品一区二区精品久久久| 妹子高潮喷水视频| 肉色欧美久久久久久久蜜桃| 青春草亚洲视频在线观看| 飞空精品影院首页| 美女脱内裤让男人舔精品视频| 成人三级做爰电影| 久久久久久久大尺度免费视频| 日韩视频一区二区在线观看| 色综合欧美亚洲国产小说| 又紧又爽又黄一区二区| 亚洲av电影在线进入| 91麻豆av在线| 美女视频免费永久观看网站| 日本a在线网址| 男女国产视频网站| 99国产极品粉嫩在线观看| 男女免费视频国产| 精品国内亚洲2022精品成人 | 中文字幕精品免费在线观看视频| 深夜精品福利| 在线av久久热| 超碰97精品在线观看| 亚洲人成77777在线视频| 欧美一级毛片孕妇| 欧美 日韩 精品 国产| 国产精品av久久久久免费| 法律面前人人平等表现在哪些方面 | 91精品三级在线观看| 久久精品久久久久久噜噜老黄| 亚洲欧美清纯卡通| 亚洲,欧美精品.| 18禁黄网站禁片午夜丰满| 国产亚洲精品一区二区www | avwww免费| 亚洲美女黄色视频免费看| 免费av中文字幕在线| 一级毛片女人18水好多| 亚洲黑人精品在线| 亚洲欧美成人综合另类久久久| 色视频在线一区二区三区| 老司机深夜福利视频在线观看 | 国产成人精品在线电影| 嫁个100分男人电影在线观看| 自拍欧美九色日韩亚洲蝌蚪91| 少妇猛男粗大的猛烈进出视频| 老司机靠b影院| 十八禁网站网址无遮挡| 午夜久久久在线观看| 老司机靠b影院| 肉色欧美久久久久久久蜜桃| 久久久久久久精品精品| 欧美97在线视频| 两性夫妻黄色片| 伦理电影免费视频| 9色porny在线观看| 满18在线观看网站| 免费看十八禁软件| 亚洲第一欧美日韩一区二区三区 | 99热国产这里只有精品6| √禁漫天堂资源中文www| 韩国高清视频一区二区三区| 亚洲精品中文字幕一二三四区 | 老司机靠b影院| 国产亚洲精品久久久久5区| 精品一区在线观看国产| 亚洲av成人不卡在线观看播放网 | 成人av一区二区三区在线看 | 性色av一级| 精品国产一区二区三区久久久樱花| 熟女少妇亚洲综合色aaa.| 亚洲国产欧美日韩在线播放| 亚洲av日韩在线播放| 淫妇啪啪啪对白视频 | 视频区图区小说| a级毛片黄视频| 国产一区二区三区综合在线观看| www.熟女人妻精品国产| 国产成人系列免费观看| 视频区图区小说| 水蜜桃什么品种好| 肉色欧美久久久久久久蜜桃| 91精品国产国语对白视频| 一区福利在线观看| 在线av久久热| 亚洲全国av大片| 亚洲国产欧美一区二区综合| 又黄又粗又硬又大视频| 最近最新免费中文字幕在线| 亚洲色图综合在线观看| 国产欧美日韩一区二区三 | 少妇粗大呻吟视频| 18在线观看网站| 女人爽到高潮嗷嗷叫在线视频| 在线永久观看黄色视频| 少妇的丰满在线观看| 蜜桃国产av成人99| 国产黄色免费在线视频| 成年人午夜在线观看视频| 亚洲精品国产精品久久久不卡| 久久久久久亚洲精品国产蜜桃av| 一本综合久久免费| 午夜福利视频精品| 国产片内射在线| 国产免费视频播放在线视频| 午夜精品国产一区二区电影| 久久久国产精品麻豆| 中文字幕人妻熟女乱码| 91老司机精品| 老司机影院成人| av线在线观看网站| 大香蕉久久网| 中文欧美无线码| 精品一区二区三区av网在线观看 | 精品一区二区三区av网在线观看 | 国产精品久久久久成人av| 新久久久久国产一级毛片| 桃红色精品国产亚洲av| 欧美日韩av久久| 久久中文看片网| 69精品国产乱码久久久| 美女国产高潮福利片在线看| 国产亚洲欧美在线一区二区| 国产高清videossex| 新久久久久国产一级毛片| 国产精品成人在线| 久久久久国产精品人妻一区二区| 最近最新中文字幕大全免费视频| 国产亚洲精品一区二区www | www.999成人在线观看| 2018国产大陆天天弄谢| 亚洲国产毛片av蜜桃av| 日本wwww免费看| 老司机在亚洲福利影院| 精品国产乱码久久久久久男人| 久久久久久久精品精品| 亚洲国产精品999| 人妻久久中文字幕网| 韩国精品一区二区三区| 99热国产这里只有精品6| 亚洲欧美日韩另类电影网站| 久久久久久久久久久久大奶| 久久久久网色| 久久中文字幕一级| 国产免费一区二区三区四区乱码| 国产一区二区激情短视频 | 精品久久久久久久毛片微露脸 | 少妇被粗大的猛进出69影院| 在线av久久热| 男女午夜视频在线观看| 亚洲国产av影院在线观看| 欧美日韩黄片免| 一级片免费观看大全| 精品久久久精品久久久| 久久国产精品大桥未久av| 我的亚洲天堂| 考比视频在线观看| 午夜成年电影在线免费观看| 日韩精品免费视频一区二区三区| 亚洲综合色网址| 亚洲精品一区蜜桃| 精品国产一区二区久久| 好男人电影高清在线观看| 青草久久国产| 一区二区三区激情视频| 伊人亚洲综合成人网| 欧美日韩国产mv在线观看视频| 老汉色av国产亚洲站长工具| 欧美久久黑人一区二区| 日韩电影二区| 国产97色在线日韩免费| 新久久久久国产一级毛片| 午夜精品久久久久久毛片777| 九色亚洲精品在线播放| 黄片播放在线免费| 少妇粗大呻吟视频| 成年人黄色毛片网站| a 毛片基地| 精品免费久久久久久久清纯 | 精品国产乱子伦一区二区三区 | 男女国产视频网站| 精品国产超薄肉色丝袜足j| 考比视频在线观看| 国产福利在线免费观看视频| 国产97色在线日韩免费| 精品福利观看| 国产免费av片在线观看野外av| 亚洲色图 男人天堂 中文字幕| 久久精品国产综合久久久| 久久精品国产亚洲av香蕉五月 | 国产高清视频在线播放一区 | 亚洲精品国产色婷婷电影| 黑人巨大精品欧美一区二区mp4| 男女床上黄色一级片免费看| 色婷婷av一区二区三区视频| 亚洲自偷自拍图片 自拍| 另类亚洲欧美激情| 精品熟女少妇八av免费久了| 亚洲欧美日韩高清在线视频 | 国产精品 国内视频| 91大片在线观看| 老司机靠b影院| 中文字幕另类日韩欧美亚洲嫩草| 男女国产视频网站| 亚洲全国av大片| 最近最新免费中文字幕在线| 国产精品免费视频内射| 亚洲精品日韩在线中文字幕| 日韩一区二区三区影片| 久久精品亚洲av国产电影网| 婷婷色av中文字幕| 日本av免费视频播放| 黄网站色视频无遮挡免费观看| 日韩人妻精品一区2区三区| 亚洲一区中文字幕在线| 高清黄色对白视频在线免费看| 午夜福利影视在线免费观看| 国产伦人伦偷精品视频| 中文字幕精品免费在线观看视频| 日本a在线网址| 中文字幕精品免费在线观看视频| 国产免费福利视频在线观看| 一边摸一边抽搐一进一出视频| 亚洲av日韩在线播放| 老司机深夜福利视频在线观看 | 亚洲欧美日韩另类电影网站| 欧美亚洲 丝袜 人妻 在线| 又紧又爽又黄一区二区| 男女边摸边吃奶| 国产精品1区2区在线观看. | 亚洲成av片中文字幕在线观看| 成人18禁高潮啪啪吃奶动态图| 蜜桃在线观看..| 美女大奶头黄色视频| 热99国产精品久久久久久7| 美女脱内裤让男人舔精品视频| 精品卡一卡二卡四卡免费| 成人亚洲精品一区在线观看| 天天躁日日躁夜夜躁夜夜| videosex国产| 国产亚洲精品一区二区www | 国产男女超爽视频在线观看| 日韩免费高清中文字幕av| 国产av精品麻豆| 欧美乱码精品一区二区三区| 日本a在线网址| 国产精品熟女久久久久浪| 亚洲七黄色美女视频| 国产精品熟女久久久久浪| 日本a在线网址| 国产精品熟女久久久久浪| 亚洲欧美精品自产自拍| 岛国在线观看网站| 99国产精品一区二区蜜桃av | 又黄又粗又硬又大视频| 纵有疾风起免费观看全集完整版| 久久国产精品影院| 欧美精品一区二区免费开放| 午夜精品国产一区二区电影| 日本撒尿小便嘘嘘汇集6| av在线app专区| 日韩大片免费观看网站| 男女无遮挡免费网站观看| 国产亚洲午夜精品一区二区久久| 国产一卡二卡三卡精品| 一级毛片女人18水好多| 国产精品一区二区在线观看99| 91精品伊人久久大香线蕉| 一个人免费在线观看的高清视频 | 99国产精品免费福利视频| 亚洲精品久久午夜乱码| 人人妻人人澡人人爽人人夜夜| 国产成人免费无遮挡视频| 色94色欧美一区二区| 9191精品国产免费久久| 亚洲国产精品一区二区三区在线| 一二三四在线观看免费中文在| 丝袜人妻中文字幕| 老司机靠b影院| 一进一出抽搐动态| 新久久久久国产一级毛片| 国产成人欧美在线观看 | 99国产极品粉嫩在线观看| 女人精品久久久久毛片| 久久精品熟女亚洲av麻豆精品| 午夜久久久在线观看| 他把我摸到了高潮在线观看 | 999精品在线视频| 欧美日韩一级在线毛片| 91成人精品电影| 亚洲国产欧美在线一区| 热99久久久久精品小说推荐| 亚洲精品久久午夜乱码| 黑丝袜美女国产一区| 老司机深夜福利视频在线观看 | 国产成人a∨麻豆精品| 国产亚洲一区二区精品| 亚洲精品国产av蜜桃| 91av网站免费观看| 久久人人爽av亚洲精品天堂| 天天影视国产精品| 亚洲美女黄色视频免费看| 人妻人人澡人人爽人人| 久久亚洲国产成人精品v| 国产欧美日韩综合在线一区二区| 国产精品99久久99久久久不卡| 天天添夜夜摸| 五月天丁香电影| 亚洲av欧美aⅴ国产| 日韩一区二区三区影片| 国产欧美日韩综合在线一区二区| 丝袜喷水一区| 欧美少妇被猛烈插入视频| 97精品久久久久久久久久精品| 午夜福利在线观看吧| 国产日韩欧美视频二区| 99精品久久久久人妻精品| 高清欧美精品videossex| 精品国产乱码久久久久久男人| 欧美性长视频在线观看| 久久精品成人免费网站| 欧美日韩亚洲高清精品| 成人国产av品久久久| 男女无遮挡免费网站观看| 精品一区在线观看国产| 黄色视频不卡| 老司机深夜福利视频在线观看 | 黄色 视频免费看| 国产精品秋霞免费鲁丝片| 涩涩av久久男人的天堂| 久久久精品94久久精品| 岛国毛片在线播放| 午夜日韩欧美国产| 日韩熟女老妇一区二区性免费视频| 91老司机精品| 美女中出高潮动态图| 视频区图区小说| 777久久人妻少妇嫩草av网站| 亚洲中文av在线| www.熟女人妻精品国产| 国产亚洲精品第一综合不卡| 一区在线观看完整版| av免费在线观看网站| 满18在线观看网站| 欧美人与性动交α欧美软件| 亚洲av电影在线进入| 啦啦啦视频在线资源免费观看| 人人妻人人爽人人添夜夜欢视频| 丰满人妻熟妇乱又伦精品不卡| 亚洲国产看品久久| 男女高潮啪啪啪动态图| 欧美激情极品国产一区二区三区| 不卡av一区二区三区| 汤姆久久久久久久影院中文字幕| 成人国产一区最新在线观看| 国产麻豆69| 黄色 视频免费看| 欧美日韩中文字幕国产精品一区二区三区 | 日韩人妻精品一区2区三区| 两性午夜刺激爽爽歪歪视频在线观看 | 国产精品自产拍在线观看55亚洲 | 无限看片的www在线观看| 日韩人妻精品一区2区三区| 不卡一级毛片| 日本撒尿小便嘘嘘汇集6| 欧美黑人精品巨大| 一级毛片精品| 欧美在线一区亚洲| 老汉色∧v一级毛片| 热99re8久久精品国产| 老熟女久久久| 日韩大码丰满熟妇| 欧美日韩福利视频一区二区| 老司机在亚洲福利影院| 免费观看av网站的网址| 欧美日韩国产mv在线观看视频| av有码第一页| 中国国产av一级| 日本91视频免费播放| 免费在线观看黄色视频的| 蜜桃国产av成人99| 国产色视频综合| 麻豆乱淫一区二区| 欧美亚洲 丝袜 人妻 在线| 侵犯人妻中文字幕一二三四区| 纵有疾风起免费观看全集完整版| 老司机亚洲免费影院| 国产成人影院久久av| 久久精品国产a三级三级三级| 国产精品免费大片| 国产三级黄色录像| 欧美日韩av久久| av福利片在线| 欧美在线黄色| 乱人伦中国视频| 国产亚洲午夜精品一区二区久久| 国产在线免费精品| 下体分泌物呈黄色| 国产国语露脸激情在线看| 国产又爽黄色视频| 国产99久久九九免费精品| 亚洲av日韩在线播放| 动漫黄色视频在线观看| 日韩制服骚丝袜av| 日韩 欧美 亚洲 中文字幕| 黑人欧美特级aaaaaa片| 免费观看人在逋| 国产精品免费视频内射| 亚洲一卡2卡3卡4卡5卡精品中文| 亚洲欧美色中文字幕在线| 黄色a级毛片大全视频| 免费观看av网站的网址| 亚洲精品中文字幕在线视频| 中文精品一卡2卡3卡4更新| 精品人妻熟女毛片av久久网站| 热99国产精品久久久久久7| 如日韩欧美国产精品一区二区三区| 天天躁狠狠躁夜夜躁狠狠躁| 高潮久久久久久久久久久不卡| 亚洲欧美精品自产自拍| 另类精品久久| 中文精品一卡2卡3卡4更新| 日韩,欧美,国产一区二区三区| 国产老妇伦熟女老妇高清| 别揉我奶头~嗯~啊~动态视频 | 成人国产av品久久久| 欧美xxⅹ黑人| 成人av一区二区三区在线看 | 男女床上黄色一级片免费看| 99国产精品一区二区三区| 国产熟女午夜一区二区三区| 男女无遮挡免费网站观看| 国产成人免费无遮挡视频| 亚洲国产欧美一区二区综合| av线在线观看网站| 久久午夜综合久久蜜桃| 中文字幕人妻丝袜制服| 热re99久久精品国产66热6| 在线av久久热| 男人添女人高潮全过程视频| 日日夜夜操网爽| 久久免费观看电影| 男女午夜视频在线观看| 国产成人av激情在线播放| 高清黄色对白视频在线免费看| 国产欧美日韩一区二区三区在线| 黑人猛操日本美女一级片| 90打野战视频偷拍视频| 建设人人有责人人尽责人人享有的| 一本—道久久a久久精品蜜桃钙片| 在线精品无人区一区二区三| 欧美激情久久久久久爽电影 | 9191精品国产免费久久| 一区二区三区精品91| 免费少妇av软件| 久久国产亚洲av麻豆专区| 久久精品成人免费网站| 免费黄频网站在线观看国产| 看免费av毛片| 一级,二级,三级黄色视频| 涩涩av久久男人的天堂| 日本av手机在线免费观看| 啪啪无遮挡十八禁网站| 好男人电影高清在线观看| av天堂久久9| 久久人妻熟女aⅴ| 三上悠亚av全集在线观看| 亚洲欧洲精品一区二区精品久久久| 91av网站免费观看| 在线亚洲精品国产二区图片欧美| www.精华液| 99国产综合亚洲精品| 美女视频免费永久观看网站| 午夜成年电影在线免费观看| 国产精品av久久久久免费| 飞空精品影院首页| www.熟女人妻精品国产| 狠狠婷婷综合久久久久久88av| 国产一级毛片在线| 欧美少妇被猛烈插入视频| 中文字幕人妻丝袜制服| 亚洲精品国产av蜜桃| 青春草亚洲视频在线观看| 久久99一区二区三区| av线在线观看网站| 国产成人av激情在线播放| 久久久久久久久久久久大奶| videos熟女内射| 欧美精品av麻豆av| 18禁国产床啪视频网站| 色婷婷久久久亚洲欧美| 视频区图区小说| 一本一本久久a久久精品综合妖精| 曰老女人黄片| 国产97色在线日韩免费| 国产精品九九99| 久久久精品区二区三区|