• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An Enhanced Multiview Transformer for Population Density Estimation Using Cellular Mobility Data in Smart City

    2024-05-25 14:39:00YuZhouBosongLinSiqiHuandDandanYu
    Computers Materials&Continua 2024年4期

    Yu Zhou ,Bosong Lin ,Siqi Hu and Dandan Yu

    1College of Computer Science and Software Engineering,Shenzhen University,Shenzhen,518060,China

    2College of Management,Shenzhen University,Shenzhen,518060,China

    3Information Center,The First Affiliated Hospital of Dalian Medical University,Dalian,116011,China

    ABSTRACT This paper addresses the problem of predicting population density leveraging cellular station data.As wireless communication devices are commonly used,cellular station data has become integral for estimating population figures and studying their movement,thereby implying significant contributions to urban planning.However,existing research grapples with issues pertinent to preprocessing base station data and the modeling of population prediction.To address this,we propose methodologies for preprocessing cellular station data to eliminate any irregular or redundant data.The preprocessing reveals a distinct cyclical characteristic and high-frequency variation in population shift.Further,we devise a multi-view enhancement model grounded on the Transformer(MVformer),targeting the improvement of the accuracy of extended time-series population predictions.Comparative experiments,conducted on the above-mentioned population dataset using four alternate Transformer-based models,indicate that our proposed MVformer model enhances prediction accuracy by approximately 30%for both univariate and multivariate time-series prediction assignments.The performance of this model in tasks pertaining to population prediction exhibits commendable results.

    KEYWORDS Population density estimation;smart city;transformer;multiview learning

    1 Introduction

    In recent years,the prevalence of wireless communication devices,such as smartphones,has significantly increased.As of 2019,there were 3.5 billion smartphone users globally,accounting for 42.63% of the world’s population [1].The rapid expansion of mobile communication services has led to the accumulation of massive amounts of system logs and paperless office data by mobile communication operators.We collectively refer to this data as cellular station data,which includes temporal and spatial user information and can accurately reflect the number of people at a specific time and location[2–6].By utilizing this information,we can estimate changes in the number of people in a particular area and study urban population movement and approximate activity trajectories.

    Population density is a critical measure for urban traffic management[7].It can be analyzed using base station signaling data to predict future changes.These predictions offer valuable information for urban planning [8],traffic management,industry applications,and social security [9].For instance,subway station locations can be planned based on population density distribution and movement[10].Traffic load predictions can inform government traffic management [11].Forecasting crowd movements can assist venue managers,such as those at banks and shopping centers,to prepare in advance.This preparation can reduce customer waiting times and increase service levels and management efficiency[12].In the context of social security,predictions can be made about possible abnormal events using base station data,population density distribution,and crowd movement information [13].These predictions allow for preemptive measures to be taken [14].In conclusion,predicting population density over time enables us to identify and address potential social and security risks early.It enhances urban operation and management efficiency and contributes to social and economic development.

    Currently,many traditional neural network models are employed for short-term time-series prediction on conventional datasets [11,15–19].Concurrently,Transformer-based models have been introduced to capture long-term dependencies in traditional time-series forecasting and have shown promising results[20–23].

    To enhance the rationality and effectiveness of base station and population data,we need to preprocess the raw cellular station data[24].However,we found that the preprocessing methods and steps in existing research are incomplete and lack systematic descriptions.On the other hand,current population prediction methods mostly consider short-term time series predictions and cannot predict long-term trends in population changes.In addition,some existing long-term prediction models fail to fully capture the high-frequency change characteristics of population numbers,leading to insufficient optimization of local high-frequency changes in population numbers.These models continue to fall short in capturing characteristics of high-frequency changes in time-series data predictions based on variations in population size.

    To address these problems,we first propose a complete base station data preprocessing method,including five steps: removing abnormal invalid data,duplicate data,ping-pong data,abnormal drift data,and organizing and counting data according to certain time slices.Then,we chose the Transformer-based model for population density prediction to solve the problem of long-term time series prediction of population density.In addition,we designed a multi-view module embedded in the time-domain or frequency-domain based Transformer model to capture the local high-frequency change characteristics of population numbers,thereby improving the prediction accuracy of the Transformer baseline model.We propose a multi-view mechanism to enhance the capturing rate of high-frequency variation patterns by combining both time-domain and frequency-domain signals in the data.

    In summary,the main contributions of this paper are as follows:

    1.We propose a comprehensive cellular data station data preprocessing method that processes temporal and spatial information,providing a reliable data foundation for subsequent population prediction.

    2.We design a multi-view enhancement model based on the Transformer,dubbed MVformer.By introducing the time-domain based Transformer structure and frequency-domain enhancement techniques,we effectively improve the accuracy of long-term time series population prediction.

    3.Through experimental verification,our model demonstrates good performance in population prediction tasks,proving the model’s effectiveness and feasibility.

    These contributions have a driving effect on the study of base station data in population prediction and related fields,providing strong support and decision-making basis for urban planning,traffic management,industry applications,and social security,among other fields.At the same time,the methods and experimental results of this paper provide references for future related research.

    2 Related Work

    This study primarily aims to predict urban population density,particularly forecasting future trends of these changes.The handling of temporal information is crucial in our prediction model.This aligns with the growing interest in temporal prediction in the machine learning and artificial intelligence fields.Significant research over the past few years has been devoted to improving temporal prediction models.

    2.1 Traditional Machine Learning Methods

    Traditional models often employ statistical methods.One such method is the Autoregressive Integrated Moving Average (ARIMA) model.Ho et al.[25] proposed this model for the reliable temporal prediction of repairable systems.Another method is the autoregressive(AR)models,which have been applied to time series forecasting.Bergmeir et al.[26]demonstrated this application using Kfold cross-validation.Additionally,Roy et al.[27]proposed the use of the Least Absolute Shrinkage and Selection Operator (LASSO) in linear regression.They used this method for predicting stock trends.

    However,these traditional methods cannot capture complex temporal patterns effectively.Therefore,researchers have started exploring the use of Neural Network models for time series pattern learning and prediction.

    2.2 Recurrent Neural Networks(RNN)Based Methods

    In recent years,with the advancement of deep learning technologies,more and more researchers have applied them to temporal prediction.For example,Siami-Namini et al.[15]found that compared with ARIMA,Long Short-Term Memory (LSTM) networks based on Recurrent Neural Networks(RNN) can significantly reduce the error rate and improve the prediction accuracy in temporal prediction tasks.Chung et al.[16]used LSTM units and Gated Recurrent Units(GRU)in RNNs to complete tasks such as speech signal modeling and other temporal signal construction tasks,achieving good results.Zhang et al.[17]based their work on the GRU neural network and used Fourier Cycle Decomposition(FCD)to decompose the time series into trend,cycle,and residual sub-sequences for training and forecasting in the temporal prediction model.

    2.3 Convolutional Neural Networks(CNN)Based Methods

    In short-term temporal prediction,besides RNN models,Convolutional Neural Networks(CNN)have also been widely applied.Koprinska et al.[18] used CNN in energy time series prediction tasks,yielding good results.Bai et al.[19]conducted a systematic evaluation of generic convolutional and recursive architectures for sequence modeling and found that simple convolutional architectures outperform typical recurrent networks on different tasks and datasets.They proposed a generic Temporal Convolutional Network (TCN).Yadav et al.[28] proposed a deep learning model that combines a multi-layer LSTM structure with multiple parallel CNN modules.This model aims to capture long-term dependencies in time series data and extract both local and global features for improving the accuracy of predicting future trends and patterns in time series data.

    2.4 Graph Neural Networks(GNN)Based Methods

    Jiang et al.[29,30] highlighted that Graph Neural Networks (GNNs) have emerged as a viable solution in various traffic forecasting problems,owing to their exceptional compatibility with traffic systems characterized by graph structures.For the temporal prediction task of traffic flow,Yu et al.[11]proposed a Spatio-Temporal Graph Convolutional Network(STGCN)model.This model considers the dependencies of space and time and effectively captures the spatio-temporal correlation of traffic flow.Ling et al.[31] proposed a network traffic prediction model based on Graph Convolution Network(GNN)to generate network traffic distribution for efficient edge server placement,resulting in significant reductions in cost and overloading of edge servers in a 5G network.

    2.5 Transformer Based Long-Term Temporal Prediction Methods

    The models above have achieved good results in short-term forecasting tasks.However,their performance in capturing long-term temporal patterns is suboptimal.To address this issue,researchers have developed Transformer-based time series prediction models specifically designed for long-term temporal prediction tasks.For long-term temporal prediction tasks,the Transformer model proposed by Vaswani et al.[20] can capture the relationships between different positions in a sequence.This model adopts an encoder-decoder structure and self-attention mechanism,effectively capturing the temporal characteristics of time series.Based on the Transformer model,Li et al.[21]completed time series prediction tasks.To solve the problem of the Transformer in handling complex time patterns and information utilization bottlenecks,Wu et al.[22] proposed an Autoformer model based on a deep decomposition architecture and self-correlation mechanism.The FEDFormer model proposed by Zhou et al.[23]adopts a Frequency Enhanced Decomposition(FED)approach,performing longterm temporal prediction tasks through Fourier transform analysis of time series in the frequency domain.The Transformer-based time series prediction models have achieved good results in generic long-term prediction tasks.However,these models still cannot capture the characteristics of highfrequency variations in time series prediction tasks based on population changes.Consequently,we propose the mechanism of multi-view in this paper and enhance the detection rate of high-frequency variation rules by combining time-domain signals and frequency-domain signals from the data.

    3 Dataset Introduction and Preprocessing

    In this section,we will elaborate on the base station dataset used in this research and its preprocessing steps.

    3.1 Dataset Description

    Our dataset is provided by a mobile communication service provider located in Shenzhen,a major city in China,and includes anonymized base station signaling data collected from August 01 to August 31,2020.Each piece of signaling data is a quintuple,containing user ID,start time,end time,base station longitude,and base station latitude.The dataset we received comprises about 7.8 million users(including roaming users from other regions).The dataset is divided into 31 files,corresponding to each day of August 2020.Each file is approximately 3 GB after compression,totaling about 100 GB.An example of the signaling dataset is shown in Table 1.

    Table 1: An example of the signaling dataset from cellular data stations

    3.2 Dataset Preprocessing

    Since the original base station data is vast and contains a large amount of invalid data,these could negatively impact the subsequent population density prediction task.For instance,invalid duplicate data could lead to the overcounting of crowd numbers at specific times and places.Furthermore,“pingpong”data and“drift”data in the base station signaling data could cause erroneous crowd counting.To enhance the usability and accuracy of the data,we performed preprocessing on the original base station data.

    3.2.1 Removal of Abnormal Invalid Data

    Invalid data refers to data that does not conform to the format listed in Table 1.During the data collection process,it is possible for certain anomalous and invalid data to be produced.Such anomalies can arise from issues such as cellular base station malfunctions or transmission errors.The presence of these anomalous and invalid data points can have a detrimental impact on subsequent data analysis and modeling,and therefore,it becomes necessary to remove them.The purpose of this step is to improve the overall quality and accuracy of the data,enabling subsequent analyses to be conducted in a more reliable and effective manner.We identify and remove invalid data by checking whether each field of every data entry adheres to the specified format and whether the number of fields is correct.

    3.2.2 Sorting the Signaling Data by User ID

    After removing the invalid data caused by technical issues,we need to sort the original data based on user IDs to facilitate subsequent preprocessing steps.However,due to the large scale of the data,the time and space costs of directly sorting the original data are unacceptable.Therefore,we chose to use a block processing method.First,we distributed the data into 4096 different files based on the first three digits of the user IDs.Then,we used a multi-process parallel processing approach to sort the records in these files by user ID and time.In this way,we obtained base station signaling data sorted by user ID.

    3.2.3 Removal of Duplicate Data

    In the base station signaling data,duplicate data refers to data with the same user ID and base station coordinates within the same time period.The purpose of this step is to remove duplicate data that have identical timestamps and location information.These duplicate data may be caused by instrument malfunctions,errors in data acquisition devices,or other factors.These duplicate data could lead to statistical errors in crowd numbers,as the same user might be counted multiple times.Deleting duplicate data helps to ensure that our analysis results are not influenced by repeated observations and reduces redundant calculations in subsequent data processing steps.Therefore,we need to read the sorted signaling data from the previous step,identify,and remove duplicate data within the same time slice.

    3.2.4 Removal of Ping-Pong Data

    In the context of base station signaling data,ping-pong data refers to the situation where a user frequently switches between two base stations within a short period of time.This typically occurs when a user is located at the boundary between two base stations,as shown in Fig.1.The presence of ping-pong data can result in a user being counted multiple times,leading to inaccuracies in analysis results and introducing bias.As ping-pong data does not represent the actual behavior of the base stations,it can also affect the subsequent processing of crowd density data.Therefore,we eliminate ping-pong data by examining the frequency of time-based switching and the hopping of connections in the data.We extract the sorted and deduplicated signaling data from the previous step,identify and remove ping-pong data that occurs within a 10-min interval.This helps ensure the continuity and consistency of the data,reduces statistical bias caused by ping-pong data,and improves the reliability of subsequent analysis.The pseudocode description of the ping-pong data processing procedure can be found in Algorithm 1.

    Figure 1: Ping-pong data occurrence schematic diagram.Ping-pong data may be generated when a user is located at the border of the coverage area of two cellular data stations

    3.2.5 Removal of Abnormal Drift Data

    Abnormal drift data refers to situations in which a user’s consecutive signaling data shows that the user has moved a significant distance in a short period of time,as illustrated in Fig.2.This can occur when the user is in an environment that interferes with the reception of the mobile signal,resulting in sudden changes in the recorded user location at the base station,among other abnormal behaviors.These deviations can stem from environmental changes,equipment malfunction,or other causes.The presence of such drift data can lead to inaccurate analysis results and distort our understanding of network performance and behavior.Therefore,it is essential to handle such data,eliminating the impact of these temporary anomalies to ensure the authenticity and rationality of the data,thereby improving the accuracy of data analysis.To address this issue,we have adopted specific thresholds and statistical rules to detect and remove these abnormal drift data.The specific steps of the abnormal drift data handling process are described in pseudocode in Algorithm 2.

    Figure 2: A schematic diagram suggesting reasons for the occurrence of abnormal drift data.Abnormal drift data may be generated when a user moves a large distance within a short time

    After the aforementioned data processing steps,we organized and counted the base station signaling data at intervals of five minutes,resulting in a time series dataset of base station signals.Slicing data by time allows for a more structured organization of data and enables data analysis based on different time periods.This approach is beneficial for identifying the temporal correlation and periodicity of data,particularly in predicting population quantities.Additionally,it provides a more comprehensive and targeted foundation for subsequent analysis.First,we performed a time-dimension visualization analysis on these data and discovered clear regularities at three different time scales:daily,weekly,and monthly.

    Fig.3 shows the distribution of base station signaling quantities on August 03,2020(Monday).It can be clearly seen that the number of signals strongly correlates with human daily activity patterns.During the day,the number of signals is high,while at midnight,the number of signals significantly decreases.In addition,there are two peak periods for the number of signals each day:one around 8 am and another around 7 pm.In Fig.4,we can also see these two peak periods on weekdays.

    Figure 3: The changes in the number of cellular data station signals detected on August 03,2020

    Fig.4 describes the distribution of base station signaling quantities from August 03,2020,to August 09,2020.Fig.5 shows the distribution of base station signaling quantities from August 01,2020,to August 31,2020.From these two figures,we can clearly see that the base station signaling data presents a cyclical pattern on a weekly timescale,and the number of signals on weekends is less than on weekdays.These variations in the number of signals primarily stem from people’s weekly work and life arrangements.

    Figure 5: The changes in the number of cellular data station signals detected within a month.It can be observed that the changes display a periodicity on a weekly basis

    To facilitate regional population prediction based on base station signaling data,it is necessary to partition the area according to certain rules.Past related studies have mostly adopted fixed-size grid partitioning.However,this partitioning method may lead to the inclusion of areas where no signals are generated within the prediction scope,and grids spanning multiple areas may lack practical significance in prediction[32].Meanwhile,manual partitioning may render the prediction work lacking in universality and portability.Therefore,in this study,we choose to use a clustering method to cluster all base stations that generate signaling data using the K-Means algorithm.Based on the Elbow rule,we calculate the intra-cluster distance of base stations under different K values.Ultimately,we set K to 30,which means we divide all base stations into 30 categories based on their longitude and latitude,serving as 30 prediction regions.

    Finally,we organize the original base station signaling data according to five-minute time slots and tag each base station signaling data with its corresponding research region ID.The start time in the original base station data is precise to the second,but in actual time series prediction tasks,we do not need such precise time slots.Therefore,during data preprocessing,we can organize crowd number data by time slots and mark the region ID to which the base station data belongs according to the partitioned areas.Ultimately,we count the crowd number data in different regions based on the region ID and corresponding time slot number to obtain the crowd number dataset for time series prediction.

    4 Crowd Prediction Model

    To more effectively predict the number of people in each area using base station signaling data,we designed a crowd number prediction model called Multi-View Transformer(MVformer).Firstly,our MVformer model is based on Transformer,a neural network model with self-attention mechanism that has a powerful ability to capture long-range dependencies in sequences.Secondly,the MVformer model combines multi-view modules and cross-view attention modules to better utilize the information among different views.The schematic diagram of the model is shown in Fig.6.The multi-view modules are responsible for extracting and integrating features,while the cross-view attention modules are responsible for utilizing the correlation information between views.

    4.1 Multi-View Division Method

    For a given time series,our model first divides the series S into a set of equally long sub-sequences{s1,s2,...,sn} based on the preset number of views.We define such a set of sub-sequences as a view.Therefore,multiple views refer to multiple sets of such sub-sequences,which we mathematically represent as:

    Here,S represents the complete time series,s represents the sub-time series,nis the index of the sub-time series,andvis the view index.represents then-th sub-time series in thev-th view,and S(v)represents the set of sub-time series in thev-th view.

    Our multi-view encoder consists of independent Transformer encoders for each view,which integrate cross-view information through a cross-view attention mechanism.The main objective of the multi-view module is to extract and fuse features from multiple views.We adopt a parallel approach by inputting each view separately into different Transformer encoders.Each encoder independently learns the representation of each view.Then,we use an attention mechanism to fuse the representations of different views together.Specifically,we calculate attention weights between different views,so that when fusing the representations,we can focus more on the views that have significant influence on the population prediction task.This allows us to better capture the high-frequency variation features in time series forecasting tasks based on population changes.

    4.2 Cross-View Attention Mechanism

    The Cross-View Attention Module is designed to better utilize the correlation information between each view.We employ an attention mechanism to calculate the correlations between different views and apply these correlations to the representations of each view.In this way,we can better capture the interactions between different views during the feature fusion process,thereby improving the accuracy of population prediction.

    As previously mentioned,we utilize a cross-view attention mechanism between multi-view encoders to share and integrate feature information across views.We fuse all pairwise information between adjacent views in an order that increases with the number of tags in the view.Specifically,we calculate attention to update the tags in the larger view.Since the tags between two views may have different hidden dimensions,we begin by projecting the keys and values to the same dimension.The cross-view attention mechanism is a direct method to combine various views.Sequentially,we blend information between every pair across adjacent viewsiand viewi+1.In order to update tags originating from a larger view s(i),we compute attention,where the queries are s(i),and the keys and values are all s(i+1).

    In addition,we encompass a residual connection around the cross-view attention operation and initialise the parameters of this operation at zero.The architecture of the Cross-View Attention Module is illustrated in Fig.7.

    Figure 7: The structure of cross-view attention

    4.3 View Encoder

    After the sequence data is processed through the multi-view module as described above,we obtain the sequence data in different views.These segment sequences in different views vary in length.This sequence data,post embedding encoding processes,enters the attention-based encoder.Each view has a unique encoder.In this paper,we propose two different forms of attention layers:the temporal vanilla attention module,and the frequency enhanced attention module.

    4.3.1 Temporal Vanilla Attention Module

    The architecture diagram of this module is shown in Fig.8.The attention block in the time domain utilizes the scaled dot-product attention method[20].The input consists of queries and keys of dimensiondk,and values of dimensiondv.We choose the dot-product(multiplication)attention as our attention function.We compute the dot product of the query with all keys,divide each product byand then apply the softmax function to obtain the weights of the values.We compute the attention function of a set of queries concurrently,bundling them into a matrixQ.The keys and values are also packed into matricesKandV.This allows us to compute the attention output matrix as:

    Figure 8: The structure of temporal vanilla attention module

    4.3.2 Frequency Enhancement Attention Module

    The structure of this module is shown in Fig.9.The attention block in the frequency domain is generally inspired by the Frequency Enhanced Block with Fourier Transform proposed by Zhou et al.[23].The input data x ∈RN×Dwill first undergo a Fourier transformation,converting time-domain data to frequency-domain data.The transformed frequency-domain data is represented by Q ∈CN×D.Then,for each frequency domain sequence,we retain only the first half.The aim of this operation is to reduce the occurrence of overfitting during training.The operation is represented by the following equation:

    wheredi=1,2,...,Ddenotes the input channel anddo=1,2,...,Ddenotes the output channel.The result of⊙R is then zero-padding to CN×Dbefore performing inverse Fourier transform back to the time domain.

    Figure 9: The structure of frequency enhancement attention module

    4.4 Global Encoder and Decoder

    After the multi-view encoding is completed,data from each view,,will be integrated and entered into a linear transformation layer for dimension unification.Then,these data are sent into a global encoder,following Vaswani et al.[20].The global encoder uses multiple Transformer layers to integrate the temporal trend information under different views.After data integration,it enters a global frequency-enhanced decoder for processing.Finally,the model outputs the final time series prediction results.

    5 Experimental Section

    5.1 Experimental Setup

    5.1.1 Dataset

    To evaluate the performance of our proposed model,we conducted a series of experiments.The base station signaling dataset we used has been described in detail in Section 3.After obtaining the preprocessed data,we standardized the dataset and divided it into training,validation,and test sets in chronological order,with a ratio of 7:1:2.Both the validation and test sets cover at least one full week,including weekdays and holidays.

    5.1.2 Hyperparameter Settings

    In this study,we set the time span of the input data to be 24 h.Due to the 5-min interval of the input data,the length of the input time series is 288.Simultaneously,we set each experimental model to predict the changes in the number of people over the next 24 h,so the length of the output prediction series is also 288.We evaluate the performance of different models by comparing the output prediction series with the actual series.In the experiments with the Transformer baseline model,to ensure fairness in model comparisons,we uniformly set the number of attention layers in the encoder to 2 and in the decoder to 1.We selected the mean squared error as the loss function and used the Adam optimizer for deep model training.The batch size and learning rate during training were set to 64 and 0.0001,respectively.

    5.1.3 Performance Evaluation Metrics

    We chose Mean Absolute Error(MAE),Mean Squared Error(MSE),Mean Absolute Percentage Error(MAPE),and Root Mean Squared Error(RMSE)as the metrics for evaluating model accuracy.Their definitions are as follows:

    Firstly,we selected Mean Absolute Error(MAE)and Mean Squared Error(MSE)as evaluation metrics.We chose these metrics because they are commonly used for regression tasks and widely applied in various prediction problems,including population forecasting.MAE measures the average absolute difference between predicted values and true values,while MSE considers the square of the errors,thereby assigning higher penalties to larger errors.Since our task involves population prediction based on regression models,these metrics provide intuitive measures of error to help assess the accuracy and performance of our models.

    Secondly,we included Mean Absolute Percentage Error(MAPE)as one of the evaluation metrics.This metric offers a relative measure of error,helping us understand the average percentage difference between the model’s predictions and the true values.In population forecasting tasks,measuring relative error allows us to evaluate the model’s performance across different scales,which is easier to comprehend and explain for decision-makers.

    Finally,we chose Root Mean Squared Error (RMSE) as another evaluation metric widely used in regression tasks.RMSE provides the standard deviation of prediction errors by taking the square root of the mean of squared errors.This metric is useful in assessing the predictive performance and error magnitude of the models,particularly when comparing mean squared errors.

    In summary,we selected these evaluation metrics to comprehensively consider different aspects of model performance and assess the accuracy and effectiveness of the proposed approach.

    5.1.4 Baseline Models

    We selected four Transformer-based models for comparison,namely Informer [33],Reformer[34],Autoformer [22],and FEDformer [23],as baseline models to evaluate the performance of our MVformer model.

    Informer[33].The Informer is an improved model built upon the Transformer,with the primary goal of enhancing model performance while reducing time complexity.

    Reformer[34].The Reformer is a memory-efficient and faster alternative to standard Transformer models,achieved through the use of reversible layers,locality-sensitive hashing attention,and other techniques.

    Autoformer [22].The Autoformer adds a decomposition block to extract the inherent complex temporal trends of the hidden states in the model,built on the design foundation of the Transformer framework.

    FEDformer[23].The FEDformer is a model proposed based on the Autoformer architecture.It applies attention mechanisms in the frequency domain and projects the time series signal into the frequency domain using the Fourier transform.

    5.2 Results Analysis

    To compare the performance of various models in long-term time series prediction tasks,and their behavior under different input and output lengths,we designed four sets of experiments,setting the input and output lengths to 144 and 288,respectively.Specifically,we input historical data spanning one and two days,and output long-term time series predictions spanning one and two days.In the context of our proposed MVformer,we conducted experiments separately using both temporal attention module(MVformer-t)and frequency enhancement attention module(MVformer-f).

    5.2.1 Univariate Results

    The univariate time series prediction results are presented in Table 2.Compared to FEDformer,MVformer achieved an average reduction of 34.7% in MSE.In the prediction task with an input history length of 144 and an output prediction length of 288,the improvement can even exceed 50%.It is worth noting that due to the differences in the temporal and frequency domain transformation modules,MVformer-f and MVformer-t exhibit different performances in different types of prediction tasks,complementing each other.This indicates the advantage of MVformer in long-term univariate time series prediction on the dataset used in this study.

    Table 2: Univariate long-term series forecasting results

    To provide a comprehensive view of these various models on the population count time-series prediction task using our proposed dataset,we visualized both the predicted and actual data from the models.We plotted the whole predicted outcomes and corresponding ground truth from the univariate test data set in Fig.10.Additionally,we plotted the single prediction data and corresponding ground truth data for various models where both the input and output sequence lengths are 288,as illustrated in Fig.11.As evident from the graph,the MVformer model we proposed,whether based on the timedomain or frequency-domain method,successfully captures the cyclical changes in population density.Baseline models such as Autoformer diverge from the peaks and troughs of the ground truth curve,yet our MVformer model still accurately predicts the ground truth curve’s peaks and troughs.Compared to other models,our proposed MVformer is superior.

    Figure 10: The prediction results of our proposed MVformer and baseline models during a period time

    Figure 11: The results of our proposed MVformer and baseline models in a single prediction

    5.2.2 Multivariate Results

    We conducted experiments on our multivariate dataset that includes population variations in all regions to evaluate the performance of our proposed MVformer in multivariate time series prediction.The experimental results are shown in Table 3.Compared to FEDformer,MVformer-t achieved an average reduction of 31.3%in MSE.In terms of the MSE metric,the MVformer-t model demonstrated the best performance across various types of prediction tasks.This highlights the advantage of MVformer in long-term multivariate time series prediction on the dataset used in this study.

    Table 3: Multivariate long-term series forecasting results

    The results presented above illustrate a comparison of various models on a multivariate population density dataset.Changes in population numbers exhibit a certain periodicity,which impacts the performance of the models in population count forecasting tasks.To investigate the spatial distribution of forecast errors in different regions,we calculated the average error based on the prediction data and actual data from different areas in the multivariate population dataset.We visualized the MAE and MAPE metrics as line graphs,as shown in Fig.12.As can be seen from the figure,the model we proposed achieved the lowest MAE and MAPE values in most regions of this population count dataset.This suggests that our proposed model can adapt to various regions,achieving better longterm prediction performance in areas with strong population dynamics,and thereby yielding improved overall performance.

    Figure 12: Prediction MAE(left)and MAPE(right)of every region

    5.3 Ablation Study

    In this study,we conduct an ablation study on the temporal dataset of the population of cellular base stations mentioned earlier,to explore the impact of different modules on the overall performance and prediction results of the MVformer model.

    5.3.1 Model-View Assignments

    In this study,the size of the view refers to the size of the sequence fragment.A larger view corresponds to larger subsequence fragments,which means fewer Transformer tokens.Conversely,a smaller view corresponds to smaller subsequence fragments,which means more Transformer tokens.Therefore,in the ablation experiments,we fixed the population size and the length of the input and output time series to be 288,and only considered a single view.By changing the size of the view,we examined the impact of the size of the individual view on the prediction results.We conducted three experiments with different view sizes,with the number of view fragments being 16,72,and 144,respectively.The experimental results are shown in Table 4.

    Table 4: Ablation study results of model-view assignments

    From the experimental results,it can be observed that as the size of the view decreases,the corresponding Transformer tokens increase,resulting in better predictive performance of the model.We interpret this as follows:in the case of a single view,a greater number of view tokens allows the model to learn more variations in the sequence.Consequently,the model can better fit and predict subsequent time series.

    5.3.2 Effect of the Number of Views

    To validate the impact of the number of views in the multi-view module on prediction results,we conducted three sets of experiments with varying numbers of views while keeping the population size and time series length of input and output fixed at 288.In the first set,we employed two segmentation strategies for the views (with 1 and 16 segments).In the second set,we used three segmentation strategies(with 1,16,and 72 segments).In the third set,we utilized four segmentation strategies(with 1,16,72,and 144 segments).By conducting these three sets of experiments,we aim to investigate the influence of different numbers of views on the prediction results of the multi-view module regarding population size.The experimental results for this section are shown in Table 5.

    Table 5: Ablation study results of the number of views

    From the experimental results,it can be observed that the predictive performance of the model increases with the number of views.One possible explanation for this observation is that by increasing the number of views,the model could learn the pattern variations of the temporal sequence from different perspectives.As a result,the model can enhance the accuracy of temporal prediction through the multi-view module.

    6 Conclusion

    This paper presents a comprehensive method for preprocessing base station data,which handles temporal and spatial information,providing a reliable data foundation for predicting population quantities.Furthermore,a Transformer-based multi-view enhancement model,MVformer,is proposed to improve the accuracy of long-term population forecasting by incorporating Transformer structures and frequency domain enhancement techniques.Experimental results demonstrate the exceptional performance of our proposed model in population prediction tasks.To further enhance the accuracy and reliability of population density forecasting,future research could focus on refining existing models.This may involve incorporating additional data sources,exploring different network architectures,or developing hybrid models that combine multiple techniques.Moreover,future studies can further explore other application areas of base station data,improve the performance of prediction models,and strengthen research on model interpretability to enhance the accuracy and credibility of prediction results.Through continuous and in-depth research,we believe that the value of base station data can be better exploited and applied.In summary,continuous exploration and research on base station data,along with improvements in prediction models and interpretability,will maximize the value and applicability of this data source.As researchers delve deeper into these areas,the potential applications of base station data in smart cities will be fully realized,influencing critical areas such as urban planning,transportation management,industry applications,and public safety.

    Acknowledgement:The authors would like to express their gratitude for the valuable feedback and suggestions provided by all the anonymous reviewers and the editorial team.

    Funding Statement:This work was supported in part by Guangdong Basic and Applied Basic Research Foundation under Grant No.2024A1515012485 and in part by the Shenzhen Fundamental Research Program under Grant JCYJ20220810112354002.

    Author Contributions:Conceptualization,Yu Zhou and Bosong Lin;Data curation,Dandan Yu;Formal analysis,Siqi Hu and Dandan Yu;Investigation,Siqi Hu;Methodology,Yu Zhou,Bosong Lin and Siqi Hu;Validation,Bosong Lin and Siqi Hu;Visualization,Dandan Yu;Writing original draft,Yu Zhou and Bosong Lin.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Data is available upon request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    可以在线观看毛片的网站| 在线观看免费视频网站a站| 女性生殖器流出的白浆| a在线观看视频网站| 国产97色在线日韩免费| 久久人人爽av亚洲精品天堂| 国产亚洲精品久久久久久毛片| av天堂在线播放| 亚洲激情在线av| 欧美日本中文国产一区发布| 欧美激情极品国产一区二区三区| 亚洲av熟女| 色播亚洲综合网| 最近最新免费中文字幕在线| 男女之事视频高清在线观看| cao死你这个sao货| 18禁裸乳无遮挡免费网站照片 | 亚洲少妇的诱惑av| 黄色 视频免费看| 精品一区二区三区视频在线观看免费| 黄片小视频在线播放| 高清黄色对白视频在线免费看| 一进一出抽搐动态| 国内精品久久久久久久电影| 精品国产一区二区久久| 久久青草综合色| 久久国产精品影院| 一区二区日韩欧美中文字幕| a在线观看视频网站| 免费久久久久久久精品成人欧美视频| 两性午夜刺激爽爽歪歪视频在线观看 | 国产精华一区二区三区| 国产精品一区二区免费欧美| 免费在线观看影片大全网站| 亚洲中文av在线| 真人一进一出gif抽搐免费| 亚洲天堂国产精品一区在线| 亚洲五月色婷婷综合| 男女下面进入的视频免费午夜 | 欧美成人一区二区免费高清观看 | 国产三级黄色录像| 精品一区二区三区四区五区乱码| 久久精品亚洲精品国产色婷小说| 涩涩av久久男人的天堂| 亚洲七黄色美女视频| 中出人妻视频一区二区| 欧美日韩精品网址| 1024香蕉在线观看| 国产av又大| 久久热在线av| 激情视频va一区二区三区| 1024香蕉在线观看| 欧美日韩福利视频一区二区| 日日夜夜操网爽| 丝袜在线中文字幕| 淫秽高清视频在线观看| 午夜老司机福利片| 中文字幕精品免费在线观看视频| 成人三级黄色视频| 久久久久国内视频| 淫妇啪啪啪对白视频| 亚洲欧美激情在线| 一二三四社区在线视频社区8| 黄片大片在线免费观看| 两个人看的免费小视频| 国产精品综合久久久久久久免费 | 久久午夜综合久久蜜桃| 人人妻人人澡人人看| 高清在线国产一区| 女人被躁到高潮嗷嗷叫费观| 91精品三级在线观看| tocl精华| 亚洲中文日韩欧美视频| 好男人电影高清在线观看| av有码第一页| 国产区一区二久久| 成年版毛片免费区| 免费在线观看完整版高清| 视频在线观看一区二区三区| 久久青草综合色| 日韩免费av在线播放| av有码第一页| 精品日产1卡2卡| 成人国产综合亚洲| 美女高潮喷水抽搐中文字幕| 午夜老司机福利片| 搞女人的毛片| 国产主播在线观看一区二区| 麻豆av在线久日| 亚洲天堂国产精品一区在线| 国产午夜福利久久久久久| 国产成年人精品一区二区| 欧美乱妇无乱码| 亚洲在线自拍视频| 亚洲伊人色综图| av在线天堂中文字幕| 国产单亲对白刺激| 久久香蕉精品热| 欧美黄色片欧美黄色片| 久久这里只有精品19| 国产精品亚洲美女久久久| 村上凉子中文字幕在线| a级毛片在线看网站| 在线播放国产精品三级| 久久久久久人人人人人| 亚洲狠狠婷婷综合久久图片| 亚洲avbb在线观看| 国产一区二区三区视频了| 亚洲中文av在线| 夜夜夜夜夜久久久久| 亚洲自偷自拍图片 自拍| 精品国产一区二区三区四区第35| 高清毛片免费观看视频网站| 久久久久久久久中文| 国产男靠女视频免费网站| 免费女性裸体啪啪无遮挡网站| 精品久久蜜臀av无| 亚洲欧美激情在线| 国产精品精品国产色婷婷| 欧美中文日本在线观看视频| 九色亚洲精品在线播放| 亚洲精品美女久久av网站| 国产极品粉嫩免费观看在线| 老司机深夜福利视频在线观看| 国产一区在线观看成人免费| 久久久久九九精品影院| 久久 成人 亚洲| 最好的美女福利视频网| 一级片免费观看大全| 国产人伦9x9x在线观看| 亚洲专区中文字幕在线| 久久中文字幕一级| 精品第一国产精品| 午夜福利18| 电影成人av| 亚洲在线自拍视频| 日韩成人在线观看一区二区三区| 欧美在线一区亚洲| 黑人欧美特级aaaaaa片| 日本五十路高清| 日日干狠狠操夜夜爽| 丰满人妻熟妇乱又伦精品不卡| 亚洲一区二区三区色噜噜| 成人免费观看视频高清| 日本三级黄在线观看| 亚洲 欧美 日韩 在线 免费| 国产成人精品在线电影| 国产av一区在线观看免费| 色哟哟哟哟哟哟| 亚洲狠狠婷婷综合久久图片| 亚洲天堂国产精品一区在线| 最好的美女福利视频网| 在线观看免费日韩欧美大片| 精品卡一卡二卡四卡免费| 国产xxxxx性猛交| 亚洲精品久久国产高清桃花| 国产1区2区3区精品| 91成人精品电影| 亚洲七黄色美女视频| 久久午夜综合久久蜜桃| 精品久久久久久,| 在线观看午夜福利视频| 真人做人爱边吃奶动态| 999久久久精品免费观看国产| 不卡一级毛片| 中文字幕av电影在线播放| 亚洲欧美精品综合一区二区三区| 国产成人系列免费观看| 亚洲中文av在线| 99在线视频只有这里精品首页| 国产片内射在线| 国产精品亚洲一级av第二区| 国产91精品成人一区二区三区| 亚洲七黄色美女视频| 国产单亲对白刺激| 国产精品野战在线观看| 亚洲精品国产区一区二| 日本 av在线| 叶爱在线成人免费视频播放| 久久香蕉激情| av电影中文网址| 一a级毛片在线观看| 后天国语完整版免费观看| 在线观看舔阴道视频| 两个人免费观看高清视频| 国产欧美日韩一区二区精品| 欧美国产精品va在线观看不卡| 国产精品久久久人人做人人爽| 久久香蕉激情| 女同久久另类99精品国产91| 精品国产乱码久久久久久男人| 亚洲第一电影网av| 97碰自拍视频| 给我免费播放毛片高清在线观看| 色老头精品视频在线观看| 精品乱码久久久久久99久播| 久久草成人影院| 国产午夜福利久久久久久| 极品教师在线免费播放| 看免费av毛片| 两个人免费观看高清视频| 日本 欧美在线| 色av中文字幕| 韩国av一区二区三区四区| 亚洲 国产 在线| 人人妻人人爽人人添夜夜欢视频| 一级,二级,三级黄色视频| 好男人在线观看高清免费视频 | 可以在线观看的亚洲视频| 欧美日韩亚洲国产一区二区在线观看| 超碰成人久久| 97人妻精品一区二区三区麻豆 | 自线自在国产av| 老汉色av国产亚洲站长工具| 欧美乱妇无乱码| 欧美一级a爱片免费观看看 | 精品欧美一区二区三区在线| 久久国产亚洲av麻豆专区| 亚洲中文字幕一区二区三区有码在线看 | 国产高清视频在线播放一区| 十八禁人妻一区二区| 中亚洲国语对白在线视频| www.精华液| 亚洲午夜精品一区,二区,三区| 丁香欧美五月| 日本免费a在线| 国产亚洲欧美98| 亚洲专区字幕在线| 国产野战对白在线观看| 香蕉丝袜av| 精品国产亚洲在线| 国产精品久久久久久精品电影 | 午夜精品在线福利| 黑人欧美特级aaaaaa片| 国产一级毛片七仙女欲春2 | 免费人成视频x8x8入口观看| 精品久久久久久成人av| 欧美日韩瑟瑟在线播放| 欧美在线黄色| 欧美成人性av电影在线观看| 亚洲avbb在线观看| 在线观看免费视频日本深夜| 国产精品免费一区二区三区在线| 亚洲人成电影观看| 亚洲欧美日韩高清在线视频| 亚洲 欧美 日韩 在线 免费| 宅男免费午夜| 麻豆av在线久日| av天堂久久9| 亚洲中文字幕一区二区三区有码在线看 | 少妇被粗大的猛进出69影院| 日韩高清综合在线| 久久影院123| 亚洲 欧美 日韩 在线 免费| 老汉色av国产亚洲站长工具| 日本一区二区免费在线视频| 亚洲午夜理论影院| 国产野战对白在线观看| 两个人视频免费观看高清| 熟女少妇亚洲综合色aaa.| 日韩精品中文字幕看吧| av在线播放免费不卡| 夜夜爽天天搞| 最新美女视频免费是黄的| 99久久久亚洲精品蜜臀av| 极品教师在线免费播放| 欧美一级毛片孕妇| 国语自产精品视频在线第100页| 少妇被粗大的猛进出69影院| 桃红色精品国产亚洲av| 亚洲 欧美一区二区三区| 999久久久国产精品视频| 欧美成人性av电影在线观看| 国产亚洲欧美精品永久| 亚洲精品一卡2卡三卡4卡5卡| 亚洲视频免费观看视频| 在线观看免费日韩欧美大片| 免费在线观看日本一区| 好男人电影高清在线观看| 极品教师在线免费播放| 欧美一级毛片孕妇| 少妇 在线观看| 黄片小视频在线播放| 精品国产乱码久久久久久男人| 亚洲人成伊人成综合网2020| 久久精品亚洲精品国产色婷小说| 国产一区在线观看成人免费| 又大又爽又粗| 久久 成人 亚洲| 久久国产乱子伦精品免费另类| 最近最新中文字幕大全免费视频| 亚洲国产中文字幕在线视频| 久久国产精品影院| 亚洲av成人av| av片东京热男人的天堂| 美女国产高潮福利片在线看| 国产一卡二卡三卡精品| 成人手机av| 午夜免费鲁丝| 亚洲精品在线观看二区| 制服丝袜大香蕉在线| 中出人妻视频一区二区| 国产视频一区二区在线看| 亚洲av成人av| 女人精品久久久久毛片| 成年版毛片免费区| 亚洲成av片中文字幕在线观看| 18禁黄网站禁片午夜丰满| 精品久久久精品久久久| 一区二区日韩欧美中文字幕| 男人操女人黄网站| 热re99久久国产66热| 美女大奶头视频| av网站免费在线观看视频| 我的亚洲天堂| 欧美+亚洲+日韩+国产| 黄频高清免费视频| 天堂影院成人在线观看| 亚洲精品中文字幕在线视频| 国产av精品麻豆| 在线观看www视频免费| 精品人妻在线不人妻| 禁无遮挡网站| 日日爽夜夜爽网站| 村上凉子中文字幕在线| 这个男人来自地球电影免费观看| 中文字幕av电影在线播放| 国产亚洲精品第一综合不卡| 一级,二级,三级黄色视频| 亚洲色图综合在线观看| 久久久久久人人人人人| 欧美激情高清一区二区三区| 日韩欧美在线二视频| 又黄又爽又免费观看的视频| 宅男免费午夜| 亚洲欧美激情综合另类| 激情在线观看视频在线高清| 深夜精品福利| 久99久视频精品免费| 欧美日韩瑟瑟在线播放| 99re在线观看精品视频| www.www免费av| 丝袜美足系列| 午夜a级毛片| 精品熟女少妇八av免费久了| 熟妇人妻久久中文字幕3abv| 亚洲视频免费观看视频| 非洲黑人性xxxx精品又粗又长| 老司机在亚洲福利影院| 一区二区日韩欧美中文字幕| 亚洲成av人片免费观看| 亚洲中文日韩欧美视频| 一a级毛片在线观看| 别揉我奶头~嗯~啊~动态视频| 美国免费a级毛片| 国产成年人精品一区二区| 十分钟在线观看高清视频www| 亚洲专区国产一区二区| 亚洲精品国产精品久久久不卡| 国产精品亚洲av一区麻豆| 国产一区二区三区视频了| 女生性感内裤真人,穿戴方法视频| 亚洲免费av在线视频| 亚洲 欧美 日韩 在线 免费| 叶爱在线成人免费视频播放| 天堂影院成人在线观看| 人人妻人人澡欧美一区二区 | 欧美国产精品va在线观看不卡| 国产黄a三级三级三级人| 夜夜夜夜夜久久久久| 精品人妻1区二区| 性少妇av在线| 一级a爱片免费观看的视频| 亚洲第一欧美日韩一区二区三区| 久久国产乱子伦精品免费另类| 在线av久久热| 国产在线观看jvid| 色av中文字幕| 国产精品久久久久久精品电影 | 亚洲一区二区三区不卡视频| 女性生殖器流出的白浆| 国产片内射在线| 每晚都被弄得嗷嗷叫到高潮| 不卡一级毛片| 性色av乱码一区二区三区2| 啦啦啦韩国在线观看视频| 国产成人系列免费观看| 91国产中文字幕| 女性被躁到高潮视频| 露出奶头的视频| 久久国产精品人妻蜜桃| 国产三级在线视频| 久久性视频一级片| 天天一区二区日本电影三级 | 少妇的丰满在线观看| 777久久人妻少妇嫩草av网站| АⅤ资源中文在线天堂| 亚洲色图 男人天堂 中文字幕| 午夜福利影视在线免费观看| 日韩欧美三级三区| 亚洲第一欧美日韩一区二区三区| 每晚都被弄得嗷嗷叫到高潮| 国产高清有码在线观看视频 | 校园春色视频在线观看| 精品人妻在线不人妻| 黄色丝袜av网址大全| 免费在线观看完整版高清| 淫妇啪啪啪对白视频| 19禁男女啪啪无遮挡网站| 一级毛片精品| 少妇裸体淫交视频免费看高清 | 熟女少妇亚洲综合色aaa.| 久久精品国产亚洲av香蕉五月| 丰满人妻熟妇乱又伦精品不卡| 宅男免费午夜| 久久 成人 亚洲| 久久久精品欧美日韩精品| 韩国av一区二区三区四区| 性色av乱码一区二区三区2| 曰老女人黄片| 青草久久国产| 18禁观看日本| 老司机午夜福利在线观看视频| 午夜a级毛片| 久久影院123| 亚洲熟妇熟女久久| 9191精品国产免费久久| 99国产精品一区二区蜜桃av| 国产视频一区二区在线看| 真人一进一出gif抽搐免费| 自线自在国产av| 欧美丝袜亚洲另类 | 日韩大尺度精品在线看网址 | 亚洲色图av天堂| 久久精品成人免费网站| 色哟哟哟哟哟哟| 非洲黑人性xxxx精品又粗又长| 欧美av亚洲av综合av国产av| 亚洲第一电影网av| 变态另类丝袜制服| 久久久国产欧美日韩av| av福利片在线| 国产av在哪里看| 免费在线观看亚洲国产| 精品久久久久久,| 欧洲精品卡2卡3卡4卡5卡区| 人人妻人人爽人人添夜夜欢视频| 看黄色毛片网站| 午夜亚洲福利在线播放| 国产亚洲av高清不卡| 成年人黄色毛片网站| 伊人久久大香线蕉亚洲五| 国产单亲对白刺激| 1024香蕉在线观看| 亚洲中文日韩欧美视频| 69av精品久久久久久| 夜夜看夜夜爽夜夜摸| 亚洲人成电影免费在线| 国产在线观看jvid| 久久热在线av| 嫩草影院精品99| 亚洲午夜精品一区,二区,三区| 99国产精品一区二区蜜桃av| 一区二区三区激情视频| 无人区码免费观看不卡| 美女大奶头视频| 午夜久久久久精精品| 女生性感内裤真人,穿戴方法视频| av电影中文网址| 亚洲精华国产精华精| 精品国产超薄肉色丝袜足j| 精品国产一区二区久久| 日韩精品免费视频一区二区三区| 国内精品久久久久精免费| 黑丝袜美女国产一区| 757午夜福利合集在线观看| 一边摸一边抽搐一进一小说| 91字幕亚洲| 国产单亲对白刺激| 日韩欧美三级三区| 韩国av一区二区三区四区| 亚洲免费av在线视频| 亚洲精品久久成人aⅴ小说| 精品一区二区三区av网在线观看| 丝袜美腿诱惑在线| 少妇 在线观看| 日本一区二区免费在线视频| 一a级毛片在线观看| 国产成人av激情在线播放| 99国产综合亚洲精品| 女人爽到高潮嗷嗷叫在线视频| 国产一区二区三区在线臀色熟女| 国产亚洲精品综合一区在线观看 | 男男h啪啪无遮挡| 久久亚洲精品不卡| 精品日产1卡2卡| 无遮挡黄片免费观看| 日本vs欧美在线观看视频| 亚洲 国产 在线| 日本一区二区免费在线视频| 午夜精品国产一区二区电影| 亚洲色图 男人天堂 中文字幕| 一区二区三区激情视频| a在线观看视频网站| 亚洲熟女毛片儿| 午夜福利在线观看吧| 精品不卡国产一区二区三区| 久久精品91无色码中文字幕| 国产精品1区2区在线观看.| 亚洲国产欧美一区二区综合| 国产精品98久久久久久宅男小说| 电影成人av| 少妇 在线观看| 丝袜人妻中文字幕| 亚洲精品一卡2卡三卡4卡5卡| 色精品久久人妻99蜜桃| 国产亚洲精品久久久久久毛片| 美女 人体艺术 gogo| 久久国产精品男人的天堂亚洲| 日韩欧美国产在线观看| 国产黄a三级三级三级人| 中文字幕精品免费在线观看视频| 久久性视频一级片| 天天躁狠狠躁夜夜躁狠狠躁| 久久影院123| 91成年电影在线观看| 9热在线视频观看99| 欧美日韩福利视频一区二区| 亚洲国产精品999在线| 日韩欧美国产在线观看| av在线天堂中文字幕| 国产亚洲av高清不卡| 日本vs欧美在线观看视频| 久久天堂一区二区三区四区| 搡老熟女国产l中国老女人| 十八禁人妻一区二区| 最好的美女福利视频网| 9色porny在线观看| 久久人妻福利社区极品人妻图片| 在线观看www视频免费| 欧美中文日本在线观看视频| 日本精品一区二区三区蜜桃| 一本大道久久a久久精品| 亚洲成av人片免费观看| 免费少妇av软件| 久久影院123| 91av网站免费观看| 久久婷婷成人综合色麻豆| 老司机午夜福利在线观看视频| 91精品三级在线观看| av电影中文网址| 成年人黄色毛片网站| 久久国产精品男人的天堂亚洲| 欧美最黄视频在线播放免费| 亚洲黑人精品在线| 精品电影一区二区在线| 国产av一区在线观看免费| av在线天堂中文字幕| 日本黄色视频三级网站网址| 久久久久国产一级毛片高清牌| 国产精品二区激情视频| 99国产精品一区二区三区| 久久午夜亚洲精品久久| 国产不卡一卡二| 波多野结衣巨乳人妻| 国产精品 国内视频| av天堂在线播放| 亚洲情色 制服丝袜| 欧美成人性av电影在线观看| 亚洲专区中文字幕在线| 国产高清videossex| 日韩欧美免费精品| 国产99白浆流出| www国产在线视频色| 欧美日本中文国产一区发布| 国产精品99久久99久久久不卡| 亚洲人成伊人成综合网2020| 日韩欧美三级三区| 欧美日韩瑟瑟在线播放| bbb黄色大片| 欧美午夜高清在线| 99国产精品一区二区三区| 操出白浆在线播放| 国产一卡二卡三卡精品| 欧美国产精品va在线观看不卡| 久99久视频精品免费| 久久天堂一区二区三区四区| 亚洲自偷自拍图片 自拍| 日韩有码中文字幕| 日韩精品免费视频一区二区三区| 亚洲中文日韩欧美视频| a在线观看视频网站| 成人三级黄色视频| 日韩欧美国产在线观看| 亚洲精品中文字幕一二三四区| 日韩免费av在线播放| 欧美大码av| 日日干狠狠操夜夜爽| 日韩精品青青久久久久久| 久久久久久久久免费视频了| 中文字幕另类日韩欧美亚洲嫩草| 国产高清视频在线播放一区| 国产高清videossex| 一夜夜www| 成人国产一区最新在线观看| 午夜久久久久精精品| 最近最新中文字幕大全免费视频| 韩国精品一区二区三区| 日韩欧美国产在线观看| 狠狠狠狠99中文字幕| 不卡一级毛片| 亚洲一区高清亚洲精品| 久久天堂一区二区三区四区| 中国美女看黄片| 色播在线永久视频|