• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    CALTM:A Context-Aware Long-Term Time-Series Forecasting Model

    2024-01-20 13:02:30CanghongJinJiapengChenShuyuWuHaoWuShuopingWangandJingYing

    Canghong Jin ,Jiapeng Chen ,Shuyu Wu ,Hao Wu ,Shuoping Wang and Jing Ying

    1School of Computer and Computing Science,Hangzhou City University,Hangzhou,310015,China

    2School of Computer Science and Engineering,Macau University of Science and Technology,Macau,999078,China

    3College of Computer Science and Technology,Zhejiang University,Hangzhou,310058,China

    ABSTRACT Time series data plays a crucial role in intelligent transportation systems.Traffic flow forecasting represents a precise estimation of future traffic flow within a specific region and time interval.Existing approaches,including sequence periodic,regression,and deep learning models,have shown promising results in short-term series forecasting.However,forecasting scenarios specifically focused on holiday traffic flow present unique challenges,such as distinct traffic patterns during vacations and the increased demand for long-term forecastings.Consequently,the effectiveness of existing methods diminishes in such scenarios.Therefore,we propose a novel longterm forecasting model based on scene matching and embedding fusion representation to forecast long-term holiday traffic flow.Our model comprises three components:the similar scene matching module,responsible for extracting Similar Scene Features;the long-short term representation fusion module,which integrates scenario embeddings;and a simple fully connected layer at the head for making the final forecasting.Experimental results on real datasets demonstrate that our model outperforms other methods,particularly in medium and long-term forecasting scenarios.

    KEYWORDS Traffic volume forecasting;scene matching;multi module fusion

    1 Introduction

    With remarkable economic advancements and rising living standards in recent years,there has been a notable surge in the preference for automobile travel,engendering a surge in highway traffic volume during holidays and leading to heightened congestion.At present,the handling of highway congestion problems usually relies on appropriate traffic control measures,such as vehicle restrictions,closure of the highway gate.And these common control measures usually need to be formulated in conjunction with the forecast results of traffic volume in order to achieve better results.Therefore,accurately forecasting traffic volume during holidays and implementing effective control measures are crucial for managing highway congestion and ensuring smooth operations.However,the complexity of transportation networks and the dynamic nature of traffic volume pose significant challenges in real-world scenarios.

    Existing traffic volume forecasting methods encompass short-term and trend forecasting,with the former being mainstream.Short-term forecasting techniques can be categorized into mainstream time series forecasting and deep learning-based forecasting.Mainstream time series forecasting methods often rely on statistical models,such as the Autoregressive Integrated Moving Average model(ARIMA)[1],Kalman filter utilizing seasonal traffic volume data[2],SVM combined with a denoising algorithm [3],and state vector combined with K-nearest neighbors (KNN) method [4].In recent years,deep learning models have gained prominence in traffic volume forecasting owing to their generalization capabilities and adeptness in capturing subtle changes.Deep Belief Networks (DBN)with a probabilistic directed graph model [5] leverage a multi-layer node graph structure to capture time-space correlations and extract effective features from traffic data.Recurrent Neural Network(RNN)structures[6],including Long Short-Term Memory(LSTM)[7],Gated Recurrent Unit(GRU)[7,8],and other sequence-based deep learning models,learn temporal dependencies through contextual analysis of sequences.Additionally,the Temporal Convolutional Network (TCN) [9] captures the spatiotemporal evolution of traffic volume through convolutional neural networks.However,these methods struggle to fully analyze the correlation of traffic volume across different roads and extract dynamic characteristics,leading to suboptimal forecasting accuracy.

    Recent research has improved forecasting accuracy by incorporating spatial factors,such as Graph Convolutional Networks(GCN)[10]model,which aggregates neighboring nodes’information through graph embedding techniques,then uses convolution to extract features that capture dynamic traffic data and spatiotemporal variation dependency.T-GCN[11],AST-GCN-LSTM model[12,13]incorporate GCN and GRU for capturing spatial and temporal dependencies,respectively,improving the traditional combination of RNN and GCN for spatiotemporal traffic volume forecasting.The Spatio-Temporal Graph Convolutional Network (ST-GCN) model with a spatial-graph attention mechanism [14] is used for traffic volume forecasting.Finally,a knowledge-graph fusion cell (KFCell) [15] that combines knowledge and volume features were added to the spatiotemporal graph convolution network to improve performance in different forecasting ranges.HetGAT [16] utilizes a weighted graph attention network(GAT)to encode the input temporal features and to decode the output velocity sequence through the express network structure.Although the above methods use the traffic network topology combined with the deep learning model to obtain good traffic volume forecasting results,they only focus on short-term forecasting scenarios and only consider the recent traffic volume information.This paper focuses on holiday traffic volume forecasting,which requires considering the characteristics of different holidays and recent traffic conditions.Moreover,as holiday traffic forecasting belongs to trend forecasting,it cannot rely on short-term historical data,and effective features need to be mined from long-term historical data.In rare trend forecasting methods,Li et al.[17]decomposed the original traffic volume data into three periods:long,medium,and short,and processed them separately.They integrated the predicted results of the three deep learning models to make long-term forecastings of traffic volume.However,the time series decomposition of this method requires the datasets to have strong periodicity and rely on short-term historical traffic volume data.Obviously,more than the generality of this method is needed,it is unsuitable for holiday traffic datasets with significant fluctuations,and it cannot mine adequate information from complex and changeable historical data sets.

    With the goal of addressing the above issues,we propose a holiday traffic volume forecasting model called CALTM(Context-Aware Long-Term Time-Series Forecasting Model)based on long and short-term scene representation fusion.The model consists of three parts:the similar scene matching module (S2M2),the long-short term representation fusion module,and the head of a simple full connection layer.Among them,the similar scene matching module is applicable to all time series problems with rich context features,and is a plug-and-play module that can find series with similar scenes to the series to be predicted from the historical data set.In order to comprehensively consider various influencing factors of holiday traffic volume,the model finds historical traffic volume with similar scenes through the similar scene matching module,then combines recent traffic volume data,similar historical traffic volume data and“time,weather,epidemic control,and tolls”conditions using the long-short term representation fusion module for holiday forecasting.The model uses a multi-head attention mechanism and context feature encoder to fuse multiple related features and fully utilize the information in the text to improve the accuracy of the model’s forecastings.Extensive experiments on a real etc dataset demonstrate that CALTM sets a new state-of-the-art performance.Especially,compared with N-BEATS[18],CALTM reduces the MSE 27.2%,12.0%,26.3%,and 7.7%on the four datasets H7,W7,H3,and W3,respectively.

    In conclusion,our contributions are summarized as follows:

    · We propose a novel plug-in module,S2M2,that can be applied to all time series problems with rich contextual features to identify similar historical time series.It is a multi-dimensional method and performs effective embedding,which can adapt to the effective encoding method of special scenes such as weekends and holidays.

    · For the purpose of addressing the challenge of insufficient long-term sequence forecast capability in recurrent neural networks,we propose CALTM,a holiday traffic volume forecasting model framework that combines scene and sequential context.

    · We use the volume data of expressway toll stations in Zhejiang Province to verify the validity of traffic volume forecast on statutory holidays(Qingming,May Day,Dragon Boat Festival,etc.)and ordinary weekends.The experimental results demonstrate that our model has improved compared with the mainstream statistical model,recurrent neural network model,convolutional network for mining time series features.

    2 Problem Formulation

    Definition 1:A collection of toll/section flow sequences.Among them,section refers to the branch roads that converge on the main highway.Letseqdenote a toll/section flow sequence,representing the highway traffic volume sequence<v1,v2,...,vStep>at a given time stepStep.LetSrepresent the set of all toll/section flow sequences for a given time instanceT.Thus,we haveseq∈S.

    Definition 2: Representation of contextual traffic factors.The contextual features such as “time,weather,epidemic situation,and toll or free access of the highway” in the dataset are mapped to vector representations via embeddings,and concatenated to obtain a context representation through a Contextual Feature Encoder(MLP).The context representation is denoted byhc(Formal representation:tuple).

    Definition 3: Collection of similar toll/section flow sequences.Given the collection of toll/section flow sequencesS,the top-k sequences are selected to form a collection of similar sequences denoted bySsim=seq1,seq2,...,seqkusing a similarity sequence retrieval algorithm applied on the sequences inS.

    Definition 4:Traffic representation with attention interaction.Given a sequence of observed traffic volumeseqobservedto be predicted and the collection of similar toll/section flow sequencesSsim,an attention interaction module is used to fuse them together and obtain a traffic representation with attention interaction denoted byhsim.Among them,theseqobservedis the traffic flows fromTahead+kdays toTahead+1 before the holiday to be predicted.AndTaheaddenotes we will predict holidays afterTaheaddays.

    Definition 5:The input of the CALTM modelincludes four parts:toll similar traffic flow sequenceseqsim,toll,cross-sectional similar traffic flow sequenceseqsim,section,observed traffic flow sequenceseqobserved,and four contextual features:time,wea,epi,andfee.Among the four contextual features,timerepresents the date type,wearepresents weather,epirepresents epidemic control andfeerepresents whether the highway is free.

    Problem Definition:Given a collection of toll/section flow sequencesSand the context representationhc,the objective is to traverse the collectionSfromTahead+kbackwards in time to obtain similar toll/section flow sequencesseqsim,tollandseqsim,sec.These sequences,along withhcand the observed traffic volume sequenceseqobserved,are then used as input to the model for predicting the holiday traffic volume sequenceseqtat a given time instanceTtwith time stepStept.

    3 Methodology

    In this section,we present our Context-Aware Long-Term Time-Series Forecasting Model termed asCALTM,as shown in Fig.1.The proposed framework contains three components,i.e.,the similar scene matching module for Similar Scene Mining,the long-short term representation fusion module for mixing scenario cues,and the head of a simple fully connection layer for flow forecasting.In the following sections,we first describe the specific details of the similar scene matching module,then elaborate on the concrete construction of long-short term representation fusion,which consists of three main components:the recurrent neural network encoder,the attention interaction,and the contextual representation.

    Figure 1:The overall view of CALTM framework

    3.1 Similar Scene Matching Module(S2M2)

    The similar scene matching module(S2M2),depicted in the left part of Fig.1,is the core design to mine the time series context.It aims to identify historical time series with similar characteristics to a given time series.Since holiday traffic forecasting falls under long-term forecasting,it is necessary to retrieve sequences of equal length that are similar to the external information of the sequence being forecasted.For main/tollgate sequences,contrastive learning is performed hierarchically on an enhanced context view using TS2Vec,increasing the distance between historical time series with different attributes.Furthermore,using the k-Shape clustering algorithm to cluster the encoded embeddings is beneficial for us to obtain the historical traffic flow sequence of the scene similar to the sequence to be forecasted according to the similarity of attributes.Finally,by searching for the category with the highest attribute similarity,the time series closest to the center point of that category are obtained as similar sequences.Aiming to handle traffic sequences lacking context features,specific rules are defined to identify similar sequences.In this paper,for the section volume sequences,the sequences of similar dates in previous years of the same section are considered as similar sequences.It is worth noting that our general-purpose method,S2M2,can be applied to all time series problems with rich contextual features.The whole algorithm is summarized with pseudocode in Algorithm 1.We give a detailed explanation as follows.

    3.1.1ClusterSeriesScenarios

    The toll traffic sequence data is clustered according to the three attributes of holidays (hol),weather (wea),and epidemic situation (epi).Determine the optimal number of clusters n using the elbow method,and obtain clustering resultsSflows={flows1,flows2,...,flowsn},where thenrepresents the number of clusters,flowsi=,andxrepresents the total number of traffic sequences in each cluster.

    3.1.2EncodeSeriesviaTS2Vec

    TS2Vec [19] improves the robustness of features in the input time series collection through timestamp masking and random cropping.Then,instance-wise and temporal contrastive losses are combined to capture contextual representations of time series.The two losses complement each other,with instance contrast learning user-specific characteristics and temporal contrast mining dynamic trends over time.The specific encoding process of TS2Vec is shown in Fig.2.And the main role of TS2Vec in this article is to expand the differences between traffic flow sequences with different temporal features,in order to achieve better clustering results in the future.Therefore,the role of TS2Vec can be understood as expanding the distance between positive and negative samples,in order to better extract useful features.

    Figure 2:Comparative learning framework

    In order to learn discriminative representations over time,TS2Vec employs a Temporal Contrastive Loss.This loss considers the representations of the same timestamp from two views of the input time series as positives,while treating those at different timestamps from the same time series as negatives.Letidenotes the index of the input time series sample andtdenote the timestamp.The representations for the same timestamp t but from two augmentations ofxiare denoted asri,tandThe temporal contrastive loss for thei-th time series at timestamptis then formulated as

    whereΩis the set of timestamps within the overlap of the two subseries,and L is the indicator function.

    The instance-wise contrastive loss indexed with(i,t)can be formulated as

    whereBdenotes the batch size.We use representations of other time series at timestamptin the same batch as negative samples.

    The two types of losses complement each other.For instance,when dealing with a collection of electricity consumption data that pertains to different users,the instance contrast loss can capture the user-specific patterns,while the temporal contrast loss focuses on capturing the evolving trends over time.The complete loss function is defined as follows:

    whereNis the number of time series,andTis the length of each time series,thereforNTis the sum of the lengths of all time series.

    3.1.3ExtractDiscriminativeFeatureofSeries

    Suppose the attributes for holidays,weather,and episodes during a given time period are represented asholt,weat,epit,andflowsihave their respective centroids asholi,weai,epii.Calculate the distance between each category’s centroid andSflowsusing the formula for Euclidean distance to find the closest category,which is denoted as.The distance is defined by the following formula:

    3.1.4MeasureSimilarityofSeries

    3.1.5DefinetheRankingRule

    For cross-sectional flow rate,due to the lack of contextual features in the etc dataset of G25-G50 expressway (only time is available),and the time period is from June 2020 to June 2021,which corresponds to historical periods of holidays and festivals,a matching rule was used to identify similar sequences.If the observed date falls on a weekend,the cross-sectional flow rate sequence datasetDsecis filtered to retain the data for weekends and select the k most recent data.If the observed date is a legal holiday,historical holiday data from the same period is selected.If there are less than k data available,the most recent legal holiday data is used as a supplement to obtain a set of similar sequencesSsec=

    3.2 Long-Short Term Representation Fusion Module

    The long-short term representation fusion module,depicted on the right side of Fig.1,which contains three components:the recurrent neural network encoder,the attention interaction,and the contextual representation.The input of the recurrent neural network encoder is mainly composed of three types of traffic volume sequences:the toll station volume sequence,section volume sequence,and recent holiday volume sequence to be forecasted.The volume sequence before the forecasted holiday is used as the query,and the encoding results of k toll station sequences provided by similar scene matching modules are used as key and value.The attention module employsMultiheadAttention(Q,K,V)to obtain the toll station attention interaction representation.Similarly,the volume sequence before the forecasted holiday is used as the query,and the encoding results of k sections are used as key and value.MultiheadAttention(Q,K,V) is applied to obtain the section attention interaction representation.Then the contextual feature encoding module includes categorical variables such as time,weather,pandemic situation,and whether the highway is toll-free which represent contextual features.They are mapped to vector representations through an embedding layer and concatenated before being input to the contextual feature encoder(MLP)to obtain the contextual representation.Finally,the output representation of the attention interaction module and the contextual representation are concatenated and fed to an MLP for feature fusion,then through a linear layer,the medium and long-term holiday traffic forecasting results are obtained.

    3.2.1RecurrentNeuralNetworkEncoder

    This module is designed to encode three types of flow sequences,including K similar traffic sequencesStollon the main line,recent holiday traffic sequencesto be forecasted,and K similar traffic sequencesSsecat cross sections,in order to obtain their respective representations.

    whereseqirepresents the input at each time step,hi-1represents the output value of the previous unit,andhican be regarded as a mixture of the unit state and hidden state in GRU,which records historical information.The scaling factorricontrols how much information fromhi-1will be retained.contains both historical information and new input at the current time step,whilezicontrols the proportion ofhi-1andseqithat flow intohi.σis a sigmoid function used to transform data into the range between 0 and 1,and it is also a crucial element in selectively preserving information.

    3.2.2AttentionInteractionModule

    This module utilizes a multi-head attention mechanism to obtain the attention interaction representation of the main-road similar flow sequence and cross-section similar flow sequence.(hobserved,Etoll)and(hobserved,Esec)are respectively used as inputs for two attention interaction modules.

    After mapping the above three sequences to vector representations through embedding,the flow representation is obtained by using multi-head attention mechanism to fuse the vector of recent holiday traffic with the other two vectors through attention,and then concatenating them.

    The attention interaction module of the model adopts multi-head attention mechanism (Multi-Head Attention).With the aim of construct multi-head attention,self-attention mechanism is firstly built,which guides the model to learn different semantic information from different perspectives.Given inputseq,Q(query),K(key)andV(value)are obtained by multiplying seq with three coefficient matrices.Qis the query sequence,whileK-Vrepresents key-value pairs.dkis the dimension of each key feature,which is used to prevent gradient vanishing.The similarity betweenQandKis first calculated through dot product operation,then weighted scaling is performed based ondk,followed by softmax normalization to obtain the weight matrix ofK,which is finally multiplied byVto obtain the feature matrix.The specific formula for self-attention is as follows:

    Multi-head attention is an improvement to the attention mechanism,which can provide encoding representation information from different subspaces in the output of the attention layer,thereby enhancing the model’s expressive power.It calculates each independent self-attention mechanism,concatenates and integrates the outputs of each self-attention mechanism,and obtains the final multihead attention score.

    By using the traffic flow sequence prior to the forecasted holiday as the query and k sets of toll station encoding results as key and value,the toll station attention interaction representation can be obtained through multi-head attention mechanism.

    By using the traffic flow sequence prior to the forecasted holiday as the query and k sets of crosssection encoding results as key and value,the cross-section attention interaction representation can be obtained through multi-head attention mechanism.

    3.2.3ContextualRepresentation

    Treating the contextual features of the dataset such as ‘time,weather,epidemic situation,and toll-free status’as categorical variables.After mapping them to vector representations using an embedding layer and concatenating them,the contextual features are inputted into a context feature encoder (MLP) to obtain the contextual representation.Each feature is transformed into a vector representation through an embedding layer,resulting in:

    wherefeattime,featwea,featepi,featfeerepresent the four contextual features of time,weather,epidemic situation,and whether the highway is free,respectively.

    The concatenated features are then input into an MLP to extract higher-level representations:

    At last,the representation obtained from the attention interaction module and contextual representation are concatenated and fed into an MLP to forecast the traffic volume for the next 24 h during the holiday.

    4 Experiments

    4.1 Implementation Details

    4.1.1Datasets

    We conduct experiments on a real-world etc dataset of G25-G50 expressway,which provides actual traffic measurement data for the Zhejiang sections of Changshen Expressway and Shanghai-Chongqing Expressway.The dataset encompasses various dimensions of data,including toll station flow,section flow,epidemic information,high-speed free status,weather conditions,and time.It covers a period of 272 days,from January 1st to September 30th,2022,and comprises data collected every 5 minutes,resulting in approximately 6.53 million data points from 24 toll stations.It is worth noting that our experimental forecasting object is the traffic flow of the expressway toll station(usually the road intersection).The data on traffic flow is collected and organized by the nearest gantry before and after the toll station.The interaction between roads was not involved in the experimental process,so the forecasting of the model is not closely related to the traffic topology relationship.

    Based on the datasets mentioned above,a series of data preprocessing steps was performed.Firstly,the relevant section ID can be obtained based on the collector ID of the main line and toll station.Then,the historical traffic flow data of the section ID during holidays can be retrieved and added to the inverted table.After that,we constructed a traffic map based on the connectivity between the main line,toll stations,and sections.And based on this graph,obtain the section ID to be forecasted and use them to retrieve holiday data in the inverted table.Weather conditions,epidemic situations,and highspeed charges were then appended to the respective data points as additional information.Following this,the traffic data from the main line and toll stations were processed.Initially,data aggregation was performed based on the main line and toll station at an hourly granularity.Subsequently,weekend and holiday data were selected as the ground truth for forecasting.Finally,the observation sequence was extracted based on the time interval T before the forecasting cycle and the observation window W and we obtained the processed dataset.

    In this study,a total of 179,263 section data and 65,968 toll station data were extracted and analyzed.In order to evaluate the performance of the model on legal holidays and ordinary weekends,we divided the processed dataset mentioned above into a holiday dataset(H)and a weekend dataset(W),which are used for forecasting traffic volumes on holidays and regular weekends.Furthermore,we select the first three days’sequence(H3/W3)and the previous week’s sequence(H7/W7)of the target holiday sequenceseqtargetfor similar sequence retrieval,and the number of days in advance is denoted asTahead.Following the division of the dataset in[20-22],we randomly select 70%as the training set,10%as the validation set,and 20%as the test set for each dataset.

    4.1.2EvaluationCriteria

    In the experiment,the performance of different methods was evaluated using the Mean Squared Error (MSE) metric.The MSE measures the deviation between forecasted values and actual values and is formulated as follows:

    whereYirepresents the actual holiday traffic volume,?Yirepresents the forecasted holiday traffic volume,andndenotes the total number of data points.In general,a smaller mean squared error indicates that the forecasted holiday traffic volume is closer to the actual holiday traffic volume,indicating better performance of the model.

    4.1.3Baselines

    We compare the performance our CALTM with the following twelve mainstream time series models.Among them,LSTM,GRU,and N-BEATS are neural network models,while Native,AutoARIMA,and others are based on statistical methods.

    · SESOpt [23]:A statistical model with smoothing coefficient optimization for time series forecasting.

    · SeasESOpt[24]:Seasonal modeling and optimization are added on the basis of SESOpt,and seasonal forecasting is performed using weighted average of time series data.

    · Naive[25]:A statistical model based on simple assumptions that uses an average of historical data or values at a previous point in time as a forecasting for the next point in time.

    · Seasonal WindowAverage[25]:The time series is modeled by the combination of autoregression,difference and moving average.

    · HistoricAverage [25]:A time series forecasting method based on the historical data average,which uses the historical data average as a forecast value for future time points.

    · Randomwalk WithDrift[25]:Based on the random walk model,a fixed drift term is introduced to describe the stochastic trend more accurately and to forecast the future value.

    · AutoARIMA[26]:A time series model that describes the trend,seasonality,and randomness of time series data through differences over time and an autoregressive moving average model.

    · LSTM[27]:A variant of a recurrent neural network(RNN)for processing time series data with a memory unit and three gated units to model long-term dependencies.

    · GRU [28]:A gated recurrent neural network for processing and forecasting time series data.Compared with standard RNNS,GRUs use fewer parameters and allow information to flow more directly through a gating mechanism,which better avoids the problem of disappearing gradients and has certain computational efficiency advantages while retaining long-term memory.

    · TCN [29]:A time series forecasting model based on convolutional neural networks (CNNS)captures the dynamics and trends in time series by stacking multiple convolutional layers and residual connections.

    · N-BEATS (Neural basis expansion analysis for interpretable time series forecasting) [18]:An emerging neural network model based on fully connected layers.Its core idea is to concatenate multiple blocks,with each block learning a part of the sequence information.The input to the next block removes the information that has already been learned by the previous block,only fitting the information that has not been learned by the previous block,similar to the idea of GBDT.Finally,the estimated results of each block are added to obtain the final forecasting result.

    4.1.4ParameterSettings

    Due to the different data volumes of the holiday dataset and the weekend dataset,we set the training batchsize to 64 and 128,respectively.But other parameters are the same for both datasets.We set the learning rate to 1e-2,and train the model using the AdamW solver for 400 epochs.The GRU encoding layer in our model sets units to 32 and we set numheads to 2 and keydim to 16 in MultiHeadAttention.For the parameters of the baseline,except for modifying the input sequence length to 120 and the output sequence length to 24,we use their respective default parameters.And all experiments are conducted using 2 RTX 3090 GPUs.

    4.2 Comparison with the Mainstream Time Series Forecasting Methods

    The results of different methods for traffic flow forecasting are listed in Table 1,depicts the traffic flow forecasting of the model at a toll station for 24 h during Qingming Festival and weekends(H7,W7,H3 and W3).In each column,the best result is highlighted in boldface and the second best is underlined.From the statistics,we draw the following observations:

    · Our proposed CALTM consistently outperforms all baselines on both four datasets.For example,compared with the second best method N-BEATS,CALTM reduces the MSE by 0.0046,0.0047,0.0041,and 0.0024 on the four holiday and weekend datasets (H7,W7,H3,and W3).Besides,CALTM has more obvious advantages in the datasets of legal holidays(H7,H3),because legal holidays contain more context features than weekends.

    · Traditional time series forecasting methods such as SeasESOpt are based on statistical models,which are limited to short-term forecasting and difficult to deal with nonlinear relationships,so they perform poorly on holiday and weekend datasets.

    · The performance of traditional deep learning models LSTM,GRU,and TCN used for time series forecasting has been greatly improved compared with statistical models.However,they are generally suitable for short-term forecasting problems.It is difficult to capture the trend characteristics of traffic flow and do not consider the context characteristics of holidays,so the overall performance is far from N-BEATS and CALTM.

    · N-BEATS achieves good forecasting results by decomposing the time series layer by layer and mining more trend features.However,N-BEATS only forecasts based on the time series itself,cannot consider the context characteristics of holidays,and cannot mine the historical traffic flow series with similar scenes to enhance the forecasting effect,which has limitations.

    Table 1: Performance comparison on two scenario datasets w.r.t MSE

    4.3 Exploration Studies

    4.3.1AnalysisonWeekdays,WeekendsandPublicHolidays

    We counted the 24-h average flow of 18 high-speed toll stations on different dates,nine of which are shown in Fig.3.There are three types of dates:weekdays,weekends,and holidays.It can be seen from Fig.3 that there are certain differences in the traffic flow of different types of days,while the difference between the three types of traffic flow at all stations between 0:00 and 6:00 is not significant.In other time periods,the flow differences between these three types of dates are also reflected in different stations.

    Figure 3: (Continued)

    Figure 3:Excerpts from the 24-h average time series of different date types at 18 stations

    Except for Lincheng Station,the holiday traffic of the other eight stations is lower than that of weekdays and weekends to a certain extent,which may be related to the fact that these areas are not popular tourist cities.The three traffic flows at each station are mainly concentrated in the time period from 6:00 to 18:00,and then begin to gradually decrease.Except Renhefu Station,there are two peak periods in the eight stations when the traffic is mainly concentrated,one peak between 6:00 and 9:00,and the other peak between 12:00 and 18:00.Overall,there is a trend of first growth and then decline,and then growth in decline.In general,Renhefu station has a peak period between 6:00 and 18:00,after which it begins to decline.

    4.3.2AnalysisonWeather

    To explore the influence of weather on traffic flow,we plot the 24-h average traffic flows under different weather conditions,nine of which are shown in Fig.4.The 272-day real etc data of highways include sunny days,heavy rains,hazes,snowy days,and typhoons.Concretely,there were 179 sunny days and 93 days of heavy rain,haze,snow,typhoon,etc.Among the 93 days of special weather,the number of days with haze,snow,or typhoon is less than 10 days,so rainy days,haze,snow,and typhoon are collectively classified as other weather.Fig.4 depicts 24-h traffic flow at different stations in sunny and other weather.It can be seen that the traffic flow on sunny days is not significantly different from that on other weather days,so weather factors have little influence on the traffic flow.

    Figure 4:Excepts from 24-h average time series of different weather at 18 stations

    4.3.3AblationExperiments

    To verify the effectiveness of the proposed CALTM,we constructed ablation experiments on both holiday and weekend datasets.Fig.5 presents the results of an ablation study on the model’s S2M2 module,represented by MSE figure.Comparative experiments were conducted to investigate the impact of incorporating retrieved similar sequences into the model.Four advanced observation days(Tahead=1,Tahead=3,Tahead=5,andTahead=7)were considered for both the holiday datasets(H3 and H7) and weekend datasets (W3 and W7).The experimental results demonstrate that the CALTM model with the S2M2 module consistently achieved significantly lower MSE value than the model without this retrieval mechanism,particularly on the holiday datasets.The most substantial improvement was observed when usingTahead=5,where the MSE value decreased by 78.67%.On the weekend dataset,the most significant improvement was observed atTahead=7,with a reduction in MSE value of 43.46%.These findings indicate that similarity sequence retrieval effectively enhances feature extraction and improves the model’s overall performance.

    Fig.6 showcases the traffic forecast results of the CALTM model and other comparative models across four datasets(H7,H3,W7,and W3)for two toll stations.Each dataset corresponds to specific conditions,denoted by the number of days ahead (e.g.,H7 represents the holiday dataset withTahead=7).Notably,the CALTM model demonstrates the most prominent forecasting results at the Nanzhuang Dou station,exhibiting a highly accurate and well-fitting forecasting curve.The N-BEATS model also performs well,while the LSTM and the statistical-based SeasESOpt models exhibit poor performance,characterized by unstable forecasting curves.Similar patterns were observed across other toll stations as well.By observing the traffic flow on different datasets within the same tollbooth,we discovered that the holiday datasets generally outperformed the weekend datasets,particularly when considering similar datasets with smallerTaheadvalues.Additionally,when comparing different toll stations within the same dataset,we found that stations with higher traffic volumes exhibited better overall forecasting performance.Obviously,Nanzhuangdou station,with the higher traffic volume,consistently achieved the more accurate and fitting forecasting results than Huzhounanlushan station.

    Figure 6: Visual analysis of forecasting results of different methods at two toll stations,including CALTM,N-BEATS,LSTM,SeasESOpt and groundtruth

    5 Conclusion

    In this paper,we propose a novel plug-in module,namely S2M2,that can be applied to all time series problems with rich contextual features to identify similar historical time series.Furthermore,We propose CALTM,a holiday traffic flow forecasting model framework that combines scene and sequential context to address the challenge of insufficient long-term sequence forecast capability in recurrent neural networks.We then evaluate the proposed model on a real-world etc dataset of G25-G50 expressway,which provides actual traffic measurement data for the Zhejiang sections of Changshen Expressway and Shanghai-Chongqing Expressway.The experimental results demonstrate that our proposed model substantially reduces the forecasting MSE compared with the mainstream methods.Additionally,we have collaborated with the local transportation design institute to deploy the CALTM model for practical application.

    There are some ideas that we would have liked to add during the representation and the forecasting functions to improve the accuracy,such as the topology of traffic networks,time-series graph neural networks,and correlation of traffic flow.Moreover,if the scene is special and cannot extract enough similar scenes for forecasting,some few-shot learning approaches could be utilized to train a model on a small number of matched samples.

    Acknowledgement:The authors would like to thank the support of advanced computing resources provided by the Supercomputing Center of Hangzhou City University.

    Funding Statement:This research was funded by the Natural Science Foundation of Zhejiang Province of China under Grant (No.LY21F020003),Zhejiang Science and Technology Plan Project (No.2021C02060),and the Scientific Research Foundation of Hangzhou City University(No.X-202206).

    Author Contributions:The authors confirm contribution to the paper as follows:study conception and design:C.Jin,J.Chen,J.Ying;data collection:J.Chen,S.Wu;analysis and interpretation of results:J.Chen,S.Wu,H.Wu;draft manuscript preparation:C.Jin,J.Chen,S.Wu,H.Wu,S.Wang,J.Ying.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data obtained and/or analyzed during the current study are available from the corresponding author on reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    免费高清视频大片| 免费在线观看视频国产中文字幕亚洲| 99re在线观看精品视频| 国产伦一二天堂av在线观看| 午夜免费激情av| 国产一区二区三区在线臀色熟女| 日韩av在线大香蕉| 精品久久久久久成人av| 看片在线看免费视频| 日韩中文字幕欧美一区二区| 日韩欧美精品v在线| 亚洲熟女毛片儿| 中文字幕最新亚洲高清| 亚洲av免费在线观看| 国产极品精品免费视频能看的| 淫妇啪啪啪对白视频| 精品一区二区三区av网在线观看| 成年版毛片免费区| www日本在线高清视频| 蜜桃久久精品国产亚洲av| 婷婷精品国产亚洲av在线| 久久九九热精品免费| 久久中文字幕一级| 国产视频内射| 精品久久久久久久久久久久久| 久久久久性生活片| 久久久久久国产a免费观看| 特级一级黄色大片| 真人一进一出gif抽搐免费| 国产精品女同一区二区软件 | 国产精品美女特级片免费视频播放器 | 欧美一级a爱片免费观看看| 精品久久久久久久毛片微露脸| 嫩草影视91久久| 天天躁狠狠躁夜夜躁狠狠躁| 亚洲无线在线观看| 久久天堂一区二区三区四区| 熟妇人妻久久中文字幕3abv| 欧美黄色淫秽网站| 欧美黄色片欧美黄色片| 午夜视频精品福利| 国产午夜福利久久久久久| 午夜两性在线视频| 免费在线观看视频国产中文字幕亚洲| 高清毛片免费观看视频网站| 午夜a级毛片| 一区二区三区激情视频| 黄色日韩在线| 精品乱码久久久久久99久播| 99热只有精品国产| 成人三级黄色视频| 看黄色毛片网站| 国产精品久久电影中文字幕| 久久性视频一级片| 欧美日韩乱码在线| 亚洲欧美日韩卡通动漫| 91在线精品国自产拍蜜月 | 中文字幕人成人乱码亚洲影| 国产精品99久久久久久久久| 老司机午夜十八禁免费视频| 老鸭窝网址在线观看| 偷拍熟女少妇极品色| 国产私拍福利视频在线观看| 久久天堂一区二区三区四区| 久久久水蜜桃国产精品网| 老汉色av国产亚洲站长工具| 一区福利在线观看| 色综合婷婷激情| 狂野欧美激情性xxxx| 99热这里只有精品一区 | 两个人视频免费观看高清| 亚洲黑人精品在线| 亚洲第一电影网av| 亚洲成av人片在线播放无| 在线十欧美十亚洲十日本专区| 亚洲五月天丁香| 国产一区在线观看成人免费| 精品一区二区三区视频在线观看免费| 一区二区三区国产精品乱码| 婷婷精品国产亚洲av| 亚洲 欧美 日韩 在线 免费| 三级男女做爰猛烈吃奶摸视频| xxxwww97欧美| 中文字幕人妻丝袜一区二区| 97人妻精品一区二区三区麻豆| 亚洲美女黄片视频| 一个人观看的视频www高清免费观看 | 国产精品日韩av在线免费观看| 亚洲自拍偷在线| 色综合婷婷激情| 麻豆成人午夜福利视频| 精品免费久久久久久久清纯| 欧美成人免费av一区二区三区| 国产精品一区二区三区四区免费观看 | 波多野结衣巨乳人妻| 给我免费播放毛片高清在线观看| 免费看美女性在线毛片视频| 国产精品久久久久久人妻精品电影| 91麻豆av在线| 夜夜躁狠狠躁天天躁| 搡老熟女国产l中国老女人| www.熟女人妻精品国产| 欧美日韩福利视频一区二区| 日本 欧美在线| 国产精品精品国产色婷婷| 午夜福利视频1000在线观看| 精品久久久久久成人av| 午夜久久久久精精品| 一本一本综合久久| 亚洲精品中文字幕一二三四区| 日韩三级视频一区二区三区| 香蕉av资源在线| 欧美zozozo另类| 老司机午夜十八禁免费视频| 亚洲第一电影网av| 久久久久久久久免费视频了| 桃色一区二区三区在线观看| 亚洲精华国产精华精| 91av网站免费观看| 18禁裸乳无遮挡免费网站照片| 在线看三级毛片| 级片在线观看| 国内久久婷婷六月综合欲色啪| 美女高潮的动态| 亚洲成人久久爱视频| 别揉我奶头~嗯~啊~动态视频| xxxwww97欧美| 免费av毛片视频| 我要搜黄色片| 两性夫妻黄色片| 在线观看日韩欧美| 性色avwww在线观看| 男女之事视频高清在线观看| 欧美在线一区亚洲| 99久久无色码亚洲精品果冻| 国产成人aa在线观看| 国产伦精品一区二区三区四那| 国产精品1区2区在线观看.| 欧美绝顶高潮抽搐喷水| 成年免费大片在线观看| av在线天堂中文字幕| 国产亚洲欧美在线一区二区| 精品久久久久久久毛片微露脸| 首页视频小说图片口味搜索| 亚洲欧美日韩东京热| 亚洲精品在线美女| 一区二区三区激情视频| 国产精品99久久99久久久不卡| 国产高潮美女av| 久久精品人妻少妇| av福利片在线观看| 老司机深夜福利视频在线观看| 亚洲色图av天堂| 黑人巨大精品欧美一区二区mp4| 久久久久亚洲av毛片大全| 一个人免费在线观看电影 | 色在线成人网| 美女黄网站色视频| 亚洲av日韩精品久久久久久密| www日本黄色视频网| 精品国产超薄肉色丝袜足j| 久久精品91蜜桃| 亚洲专区中文字幕在线| 国产亚洲欧美在线一区二区| 国产亚洲欧美98| 超碰成人久久| 久久久久亚洲av毛片大全| avwww免费| 亚洲 国产 在线| 国产乱人视频| av中文乱码字幕在线| 香蕉av资源在线| 亚洲精品中文字幕一二三四区| 天堂影院成人在线观看| 亚洲熟女毛片儿| 亚洲美女视频黄频| 国产一区二区在线av高清观看| 黄片小视频在线播放| 国产午夜精品久久久久久| 国产精品女同一区二区软件 | 成熟少妇高潮喷水视频| 免费看光身美女| 国产真人三级小视频在线观看| 精品久久久久久,| 国产精品1区2区在线观看.| 婷婷亚洲欧美| 精品不卡国产一区二区三区| 亚洲av成人不卡在线观看播放网| 搡老岳熟女国产| 露出奶头的视频| tocl精华| 国内精品久久久久久久电影| 国产成人影院久久av| 久久久久九九精品影院| 女警被强在线播放| 亚洲av电影不卡..在线观看| 一区二区三区激情视频| 热99在线观看视频| 日韩三级视频一区二区三区| 18禁美女被吸乳视频| 搡老妇女老女人老熟妇| 国产精品九九99| 十八禁人妻一区二区| 欧美丝袜亚洲另类 | 精品国产亚洲在线| 亚洲欧美一区二区三区黑人| 99热精品在线国产| 每晚都被弄得嗷嗷叫到高潮| 久久精品国产清高在天天线| 日韩三级视频一区二区三区| a级毛片a级免费在线| 亚洲精品中文字幕一二三四区| 国产精华一区二区三区| 欧美日韩亚洲国产一区二区在线观看| 亚洲av中文字字幕乱码综合| 午夜福利18| 一个人观看的视频www高清免费观看 | 香蕉丝袜av| 欧美一区二区国产精品久久精品| 亚洲九九香蕉| 亚洲欧美日韩高清专用| 日韩欧美国产在线观看| 欧美日本视频| 国产乱人视频| 制服丝袜大香蕉在线| 久久伊人香网站| 免费无遮挡裸体视频| av中文乱码字幕在线| 午夜a级毛片| 国产精品免费一区二区三区在线| 深夜精品福利| 最近最新中文字幕大全免费视频| 国产精华一区二区三区| 久久久国产成人精品二区| 欧美中文日本在线观看视频| 我的老师免费观看完整版| 成在线人永久免费视频| 国产成人一区二区三区免费视频网站| 人妻丰满熟妇av一区二区三区| 亚洲一区二区三区不卡视频| 一个人看视频在线观看www免费 | 久久热在线av| 欧美日韩瑟瑟在线播放| 波多野结衣高清无吗| 99在线人妻在线中文字幕| 欧美中文日本在线观看视频| 此物有八面人人有两片| 精品一区二区三区视频在线 | av福利片在线观看| 免费看a级黄色片| 午夜福利欧美成人| 丰满人妻一区二区三区视频av | www日本在线高清视频| 亚洲精品在线观看二区| 一本一本综合久久| 黑人操中国人逼视频| 国产野战对白在线观看| 麻豆国产97在线/欧美| 日韩有码中文字幕| 波多野结衣巨乳人妻| av天堂中文字幕网| 黑人巨大精品欧美一区二区mp4| 日韩欧美 国产精品| 国产真人三级小视频在线观看| 成人特级av手机在线观看| 毛片女人毛片| 色尼玛亚洲综合影院| 在线观看免费视频日本深夜| 99久久无色码亚洲精品果冻| 亚洲中文日韩欧美视频| 色在线成人网| 国产精品av久久久久免费| 天天躁日日操中文字幕| 男插女下体视频免费在线播放| 精品久久久久久久末码| 欧美另类亚洲清纯唯美| 国产精品99久久久久久久久| 男人舔女人的私密视频| 亚洲中文av在线| 亚洲av电影不卡..在线观看| 国产精品久久久av美女十八| 国产日本99.免费观看| 国产精品一区二区三区四区免费观看 | 天天躁狠狠躁夜夜躁狠狠躁| 国产精品,欧美在线| 这个男人来自地球电影免费观看| 成年免费大片在线观看| 亚洲欧美精品综合久久99| 天天一区二区日本电影三级| 中文字幕熟女人妻在线| 特大巨黑吊av在线直播| 精品国产超薄肉色丝袜足j| 国产精品自产拍在线观看55亚洲| 99热6这里只有精品| 国产精品一区二区免费欧美| 国产伦精品一区二区三区四那| 两性午夜刺激爽爽歪歪视频在线观看| 激情在线观看视频在线高清| 午夜影院日韩av| 久久久国产精品麻豆| 婷婷丁香在线五月| 色在线成人网| 免费看十八禁软件| 人妻丰满熟妇av一区二区三区| 99视频精品全部免费 在线 | 我要搜黄色片| 一本一本综合久久| www.精华液| 欧美日韩瑟瑟在线播放| 男女做爰动态图高潮gif福利片| 中文字幕精品亚洲无线码一区| 亚洲国产日韩欧美精品在线观看 | 黑人操中国人逼视频| 欧美乱色亚洲激情| 日韩免费av在线播放| 国产成年人精品一区二区| www.www免费av| 成人精品一区二区免费| 国产男靠女视频免费网站| 中文字幕熟女人妻在线| 99在线视频只有这里精品首页| 亚洲人与动物交配视频| 99久久精品国产亚洲精品| 免费在线观看日本一区| 国产精品99久久99久久久不卡| 美女黄网站色视频| 制服人妻中文乱码| 精品人妻1区二区| 狠狠狠狠99中文字幕| 成年免费大片在线观看| 天堂影院成人在线观看| 国产激情欧美一区二区| 成年版毛片免费区| 亚洲成人精品中文字幕电影| 天堂√8在线中文| 成年免费大片在线观看| av福利片在线观看| 色综合亚洲欧美另类图片| 制服丝袜大香蕉在线| a级毛片在线看网站| 国产精品美女特级片免费视频播放器 | 99久国产av精品| 又黄又粗又硬又大视频| 亚洲精品粉嫩美女一区| 成年免费大片在线观看| 好看av亚洲va欧美ⅴa在| 男人舔奶头视频| 在线观看一区二区三区| 亚洲熟妇熟女久久| 欧美色视频一区免费| 蜜桃久久精品国产亚洲av| 757午夜福利合集在线观看| 国产精品av久久久久免费| 天天躁日日操中文字幕| 国产高清videossex| 999精品在线视频| 91麻豆精品激情在线观看国产| 欧美成人性av电影在线观看| 成年人黄色毛片网站| 狂野欧美激情性xxxx| 不卡av一区二区三区| 日韩成人在线观看一区二区三区| 久久香蕉国产精品| 成年免费大片在线观看| 国产精品一区二区三区四区免费观看 | 国产私拍福利视频在线观看| 香蕉丝袜av| 国产激情久久老熟女| 日韩欧美国产一区二区入口| 国产97色在线日韩免费| 美女免费视频网站| 蜜桃久久精品国产亚洲av| 一区福利在线观看| 色尼玛亚洲综合影院| 人人妻,人人澡人人爽秒播| 日韩高清综合在线| 国产精品永久免费网站| 亚洲国产高清在线一区二区三| 亚洲国产欧美人成| 老司机午夜福利在线观看视频| 9191精品国产免费久久| 不卡一级毛片| 久久婷婷人人爽人人干人人爱| 最近最新免费中文字幕在线| 精品一区二区三区av网在线观看| 麻豆久久精品国产亚洲av| 亚洲av免费在线观看| 毛片女人毛片| 在线观看66精品国产| 性色avwww在线观看| 日本黄色视频三级网站网址| 国产真实乱freesex| 精品电影一区二区在线| 成年免费大片在线观看| 亚洲国产看品久久| 欧美丝袜亚洲另类 | 免费在线观看日本一区| 一区二区三区激情视频| 日本精品一区二区三区蜜桃| 好男人在线观看高清免费视频| 中文资源天堂在线| 一区福利在线观看| 久久久国产成人免费| 国产精品一区二区三区四区免费观看 | 最近最新中文字幕大全电影3| av在线蜜桃| 视频区欧美日本亚洲| 久9热在线精品视频| 国产av麻豆久久久久久久| 国产野战对白在线观看| 色播亚洲综合网| 国产成人av激情在线播放| 精品久久久久久,| 首页视频小说图片口味搜索| 97碰自拍视频| 99久久无色码亚洲精品果冻| 一级毛片高清免费大全| 亚洲欧美日韩无卡精品| 国产精品日韩av在线免费观看| 在线看三级毛片| 韩国av一区二区三区四区| 少妇丰满av| 99国产精品一区二区蜜桃av| 国产成人av激情在线播放| 五月玫瑰六月丁香| 国产久久久一区二区三区| av天堂在线播放| 国产精品久久久久久人妻精品电影| 久久人妻av系列| 成年女人看的毛片在线观看| 99热只有精品国产| 精品一区二区三区四区五区乱码| 网址你懂的国产日韩在线| 女生性感内裤真人,穿戴方法视频| 国产激情久久老熟女| 国产免费av片在线观看野外av| 欧美3d第一页| 欧美+亚洲+日韩+国产| 精品久久久久久久末码| 精品久久久久久久久久久久久| 免费无遮挡裸体视频| 草草在线视频免费看| 日韩有码中文字幕| 国产精品99久久99久久久不卡| 一个人免费在线观看的高清视频| 岛国在线免费视频观看| 大型黄色视频在线免费观看| 亚洲av熟女| 免费一级毛片在线播放高清视频| 久久99热这里只有精品18| 99热6这里只有精品| 不卡av一区二区三区| 露出奶头的视频| 国产激情欧美一区二区| 国产一区二区在线av高清观看| 99久久精品一区二区三区| 亚洲精品美女久久久久99蜜臀| 黄色丝袜av网址大全| 少妇人妻一区二区三区视频| 性色avwww在线观看| 国产熟女xx| 黄片小视频在线播放| 少妇的丰满在线观看| 大型黄色视频在线免费观看| 99riav亚洲国产免费| 日本熟妇午夜| 听说在线观看完整版免费高清| 国产1区2区3区精品| 一二三四社区在线视频社区8| 男人舔奶头视频| 国产精品日韩av在线免费观看| 日韩欧美国产在线观看| 亚洲精品在线观看二区| 毛片女人毛片| 亚洲国产精品合色在线| 久久久久久久久久黄片| 老司机午夜福利在线观看视频| 12—13女人毛片做爰片一| 又大又爽又粗| 国产免费男女视频| 成年女人永久免费观看视频| 亚洲av成人精品一区久久| 男女做爰动态图高潮gif福利片| e午夜精品久久久久久久| 不卡一级毛片| 亚洲av日韩精品久久久久久密| 久久人妻av系列| 亚洲国产精品sss在线观看| 久99久视频精品免费| 日本成人三级电影网站| 两个人视频免费观看高清| 一个人看视频在线观看www免费 | 91av网站免费观看| 国产综合懂色| 一区福利在线观看| 午夜激情福利司机影院| 国产黄片美女视频| 听说在线观看完整版免费高清| 一本精品99久久精品77| 美女午夜性视频免费| 色视频www国产| av中文乱码字幕在线| 亚洲人成电影免费在线| av在线蜜桃| 19禁男女啪啪无遮挡网站| 日本一本二区三区精品| 1024香蕉在线观看| 老司机午夜福利在线观看视频| 成人国产一区最新在线观看| 久久久久九九精品影院| 日韩精品中文字幕看吧| 制服丝袜大香蕉在线| 国产亚洲精品av在线| 九色成人免费人妻av| 最新美女视频免费是黄的| 男女做爰动态图高潮gif福利片| 色av中文字幕| www日本黄色视频网| 成熟少妇高潮喷水视频| 国产成人系列免费观看| 国产成人av教育| 日本黄大片高清| 国产精品 国内视频| 国产午夜精品论理片| 麻豆国产97在线/欧美| 国产精品一区二区免费欧美| 99精品欧美一区二区三区四区| 亚洲av成人精品一区久久| 国产私拍福利视频在线观看| 欧美中文综合在线视频| 久久亚洲精品不卡| 国产综合懂色| 99国产综合亚洲精品| 亚洲熟女毛片儿| 夜夜夜夜夜久久久久| 色吧在线观看| 亚洲一区二区三区不卡视频| 黄频高清免费视频| 在线十欧美十亚洲十日本专区| 两个人视频免费观看高清| 中文资源天堂在线| 亚洲精品美女久久av网站| 国产aⅴ精品一区二区三区波| 亚洲片人在线观看| 中出人妻视频一区二区| 在线视频色国产色| xxxwww97欧美| 美女大奶头视频| 人妻久久中文字幕网| 巨乳人妻的诱惑在线观看| 久久草成人影院| 五月伊人婷婷丁香| 中文在线观看免费www的网站| 90打野战视频偷拍视频| 国产精品国产高清国产av| 国产精品久久视频播放| 精品国产亚洲在线| 亚洲五月婷婷丁香| 日本黄色视频三级网站网址| 黄片大片在线免费观看| 久久国产精品影院| 午夜福利在线观看吧| 免费看光身美女| 日日干狠狠操夜夜爽| 老司机深夜福利视频在线观看| 日日干狠狠操夜夜爽| 少妇熟女aⅴ在线视频| 黄频高清免费视频| 成年版毛片免费区| 日本一二三区视频观看| 久久久久久人人人人人| 日本三级黄在线观看| 亚洲精品美女久久av网站| 在线观看舔阴道视频| av视频在线观看入口| 国产黄片美女视频| 中国美女看黄片| 日本成人三级电影网站| 在线观看午夜福利视频| 在线a可以看的网站| 亚洲精华国产精华精| 99视频精品全部免费 在线 | 一区二区三区国产精品乱码| 免费观看人在逋| 成人永久免费在线观看视频| 少妇裸体淫交视频免费看高清| 日本熟妇午夜| 亚洲国产欧美一区二区综合| 欧美又色又爽又黄视频| 黑人欧美特级aaaaaa片| 成人午夜高清在线视频| 日韩精品青青久久久久久| 天堂影院成人在线观看| 51午夜福利影视在线观看| 黄色 视频免费看| 精品99又大又爽又粗少妇毛片 | 香蕉丝袜av| 熟妇人妻久久中文字幕3abv| 舔av片在线| 国产精品99久久久久久久久| 欧美成人性av电影在线观看| 手机成人av网站| 日本一二三区视频观看| 两个人的视频大全免费| x7x7x7水蜜桃| 搡老岳熟女国产| www.熟女人妻精品国产| 成在线人永久免费视频| 人妻夜夜爽99麻豆av| 国内久久婷婷六月综合欲色啪| 午夜亚洲福利在线播放| 悠悠久久av| www国产在线视频色| 久久天堂一区二区三区四区| 午夜精品一区二区三区免费看|