• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Improving Stock Price Forecasting Using a Large Volume of News Headline Text

    2021-12-15 07:10:26DaxingZhangandErguanCai
    Computers Materials&Continua 2021年12期

    Daxing Zhangand Erguan Cai

    1Department of Mathematics,Clemson University,Hangzhou,310018,China

    2Institute of Graphics and Image,Hangzhou Dianzi University,Hangzhou,310018,China

    Abstract: Previous research in the area of using deep learning algorithms to forecast stock prices was focused on news headlines,company reports,and a mix of daily stock fundamentals, but few studies achieved excellent results.This study uses a convolutional neural network(CNN)to predict stock prices by considering a great amount of data,consisting of financial news headlines.We call our model N-CNN to distinguish it from a CNN.The main concept is to narrow the diversity of specific stock prices as they are impacted by news headlines, then horizontally expand the news headline data to a higher level for increased reliability.This model solves the problem that the number of news stories produced by a single stock does not meet the standard of previous research.In addition, we then use the number of news headlines for every stock on the China stock exchange as input to predict the probability of the highest next day stock price fluctuations.In the second half of this paper,we compare a traditional Long Short-Term Memory (LSTM) model for daily technical indicators with an LSTM model compensated by the N-CNN model.Experiments show that the final result obtained by the compensation formula can further reduce the root-mean-square error of LSTM.

    Keywords: Deep learning; recurrent neural network; convolutional neural network; long short-term memory; stocks forecasting

    1 Introduction

    The price changes in financial markets are nonlinear and the influencing factors are variable.Stock markets attract many wealthy shareholders and researchers.Henrique et al.[1] study mentions a theory called efficient-market hypothesis (EMH), arguing that because financial markets will immediately adapt to any information available, people cannot predict the market’s movements.In many studies, securities analysts research the price changes in the stock market, then analyze this vast amount of data and come to useful conclusions.Professionals in this field invest a good deal of time in learning how to analyze data.Artificial intelligence technology is increasingly being incorporated in data analysis, particularly in the area of predicting stock movements.

    Stock price forecasting methods can be divided into two categories:a time series method that was modeled by Aue et al.[2] using a process called autoregressive conditional heteroscedasticity(ARCH); and machine learning algorithms.As of 2018, models based on artificial neural networks(ANNs) and support vector machines (SVMs) were still at the core of a stock market’s analysis.ANNs contain many neurons, can handle the relationship between large amounts of data, and have excellent performance with financial time series data [3,4].Multilayer feedforward networks(also called the BP network model) were used for time series data prediction [5].Neural networks continue to improve.Convolutional neural networks (CNNs) and recurrent neural networks(RNN) offer good performance in a specific mode.Long short-term memory (LSTM) represents an improvement over RNNs and is considered to be the most suitable network for predicting stock data [6].However, Kim et al.[7–9] expressed data and indicators of stock as a chart, leveraged CNN, and proposed a CNN-LSTM model, proving that CNN still plays an important role in the research of predicting stock prices.

    LSTM inputs time series data for the purpose of prediction.Most stocks have more than 6,000 fundamentals, which means that technical analysis (TA) indicators have the same data sample.On the basis of a time series, news related to daily fluctuations in stock prices can also be part of another CNN algorithm input.Ding et al.[10] shows that shareholders can view stockrelated news positively or negatively, which will determine whether they buy or sell stocks, thereby affecting the fluctuation of stock prices.

    Improved algorithms related to LSTM all use fundamental indicators and TA indicators as input, but differ in the length of the predicted period.Rather et al.[11] concentrated on every five minutes when a stock price changes, Chung et al.[12] predicted daily stock price, and Weng et al.[13,14] predicted the price of stocks one month out for major U.S.stock and sector indices.However, daily price research is the most common time period.For the input dimension of daily price data, more data represents more information, which can reduce the likelihood of errors in a prediction.

    Ding et al.[15] showed that no matter whether the news is weekly or monthly, it affects the stock market trend of the next day.Therefore, this study uses two weeks’worth of news headlines as data.However, when they did related research, they only tracked the news data of current stocks, which resulted in an insufficient sample size for deep learning algorithms, and the final experimental results were not very promising.When using a supervised deep learning algorithm,at least 5000 of each type of labeled sample is required to reach an acceptable range.When the number of samples is 10 million, it is generally sufficient to surpass human performance.Goodfellow et al.[16], mentioned in Fig.1 below, shows the amount of sample data that emerged from deep learning in 2015.The results of deep learning algorithms are positively correlated with the amount of data.While tracking a single stock, even when it starts to record news from the time a company goes public, most companies in China can only possess a sample of around 5,000.Emerging companies do not generate enough news samples, and so an emerging company cannot guarantee enough weight for an algorithm to predict the future price of stock.To expand the amount of stock news sample data, this paper unifies all stock news that can be collected,removes redundant information, improves the accuracy of experimental results, and expands the use of the algorithm to stock prediction.

    Figure 1:Increasing dataset size over time

    2 Materials and Methods

    This section describes the sources of the datasets used in this paper and distributions to each network model.There were two input types:TA indicators, which display daily financial technical directional-movements series called Stock datasets, and half-month financial news titles of each stock, called News datasets.

    2.1 Stock Datasets

    The basic stock data were obtained from Tonghuashun (China’s professional economic website), marked asxS.

    Web crawler technology was used to crawl daily data and TA indicators from the website.Data on 1014 individual stocks were obtained from listing up to 3/8/2019, including Open,Close, High, Low, Volume, Vibration, as LSTM first layer input, with 80% of stocks representing the training set, and 20% the test set.The calculation of TA indicators is similar to that in Guresen’s [4].

    The majority of papers predict or calculate the composite index as their main research target,such as the Shanghai Stock Exchange (SSE) Composite Index, which makes result more stable and more representative.The LSTM input indicators dataset is the largest one.The SSE Composite Index contains all the stocks listed on the SSE.In terms of time, the prices of stocks on the SSE tend to be more affected by political factors than ordinary events.The bigger the news, the greater the impact on the index [10].However, the prices of individual stocks are easily influenced by significant company events.Therefore, the daily basic data of the SSE Composite Indexexperimental results are only used as a sample comparison with individual stocks on LSTM.This is why this paper also focuses on the effect of deep learning on individual stocks.

    DataxSwas used to prove the link between the composite index and individual stock data in LSTM.The training model of all 1014 stocks can be used to predict the final stock price fixed by news trained model in last formula.The results can further reduce the number of errors of LSTM.

    2.2 News Datasets

    News headlines text were also obtained from Tonghuashun.We call this dataxN.

    Ding’s algorithm uses about 100 news headlines and content in order to predict each single stock price.Deep learning algorithms are often dependent on big data.If the data volume is low, the final result will be unsatisfactory for the LSTM algorithm.In Ding’s paper, the highest accuracy achieved by the algorithm is only 56%.The easiest way to increase accuracy of stock news on stock price forecasting algorithms is to increase the amount of text data.This paper proposes ideas for achieving this target volume.

    Existing research tracked only single stock news series as input for calculating a specific stock price.That’s why the CNN net input dataset number is small.We think that, under normal circumstances, different stocks have different sensitivities to similar news, especially financial news.The price trends of A-shares will be stable even if the news said “Huge Funds Coming Soon,”but B-shares are more likely to be affected.A single stock may only produce 200 news stories in an entire year.To improve the final accuracy of using stock headlines for stock forecasts, this paper makes the assumption that each stock has exactly the same impact on similar news, and gathers all stocks’news headlines within 15 days to calculate the next day’s price fluctuations.This method can improve the accuracy of the algorithm because of the large amount of data, while ignoring the negative effects caused by different stock sensitivity.In the Chinese stock market,there are now 3000 stocks.This means the number of news headlines for a single stock can be expanded by 3000 times.If a stock’s news is tracked within half month, the amount of sample data that can theoretically be news headlines could reach 1 million.The experiment described in Section 3 of this paper also studies the impact of news headline text data on the accuracy of the CNN.

    In summary, this paper shifts from research on vertical data to research on horizontal data.All stocks’price impact factors are considered the same, and stock headline data is horizontally enlarged at the same time.Unfortunately, this paper only uses the news headlines of 1014 stocks vs.all 3000 stocks in the Chinese stock market, and the amount of data is therefore expanded by 1014 times.Detailed data when tracking each stock’s news spans within 15 days, until 5/23/2019,and the news’s time limit on May 23rd, for example, is for the closing time of 3 pm, which is used to predict the highest price of May 24.This assumption is feasible if it can increase the amount of data expansion in the CNN in order to improve final accuracy.To further lessen the impact of different companies on the news sensitivity, after the text data is unified, the redundant text should be reduced.It has been observed that many news headlines contain names of companies and/or stocks.These names are not valid information for the actual algorithm, because this type of information is deleted in text preprocessing.For the 1014 stocks, there were 29,700 news headlines.Therefore, every stock had 29.3 news headlines on average.The training and test set settings are shown in Tab.1.The training forecast set allocation ratio is the same as the stock’s fundamental indicator dataset, and the news headlines are obtained.Of the text in the dataset, 80% is used for training, and the remaining 20% is used as a test set.

    Table 1:Statistics of news dataset

    To verify the final network effect, we should re-establish verification set outside the time points of dataxSandxN, so we also get stock base data from the 23rd (Thursday) and the following day.The highest price is calculated as the accuracy of the label layer compared to previous day’s highest price.The higher price on the second day is set to 1 in the label layer, and vice versa, the lower price is set to 0, which predicts the stock price of the following day, all the way through 15 days of news headlines.We call this data news ofxF, which contains the TA indicators of Stock Data LSTM (D-LSTM), and also contains News Data CNN (N-CNN)’s news headlines text.The final experiment will take the next day’s highest price H from D-LSTM andH′contrast corrected by a compensation formula in which parameters are obtained from N-CNN.

    2.3 Nets Design

    2.3.1 N-CNN Design

    Fig.2 shows the design of the CNN for news.First, data language type obtained from news headlines is Chinese.An embedding layer is needed to convert text into a word vector.In this step,we use the Word2vec tool and the continuous Bag of Words (CBOW) model for the conversion of Chinese word embedding.

    Figure 2:News CNN model architecture

    The output of a word vector is a two-dimensional matrix, which is used as input of the convolutional layer.The convolutional layer is often followed by the pooling layer to reduce dimension, the number of parameters, and training time.Immediately after the pooling layer is a three fully connected (FC) layers.The third FC layer activation function is softmax, whose formula is shown as Eq.(1).Since we only need to convert stock-derived information into a rising or falling output, Softmax outputs only two types, a label that is 0 or 1.Score set S calculated by the pre-neuron is mapped by formula to interval (0, 1).The result is [O1,O2], whereO1represents the probability of falling, andO2represents the probability of rising.

    2.3.2 LSTM Introduction

    A recurrent neural network (RNN) differs from a traditional neural network in that it can consider the influence of information that has already appeared on the next prediction.LSTM performs better than the other RNNs [17] in long-term sequence data.Unlike an RNN that simply uses the hyperbolic tangent function (tanh) layer as the structure of repeating module, LSTM has more interactive layers, as shown in Fig.3.

    Eqs.(2) to (5) show calculations for each gate and cell state, whereit,ft,otare all sigmoid layers expressed byσ(·),

    W represents weight matrices andbis a bias term.The cell’s workflow is as follows:xtis current input andht-1is previous input.The first step is the forget gate deciding to discard the information, theσ(·)function only outputs value between 0 and 1, with 1 completely reserved and 0 completely discarded.Through this door, we have deleted information, and then the input gate has to decide to add some information.In addition to the sameσlayer, there is a step regarding.Use tanh function to create candidate vectors, then use the result of point-wise multiplication ofitandwithctupdated as in Eq.(4).Current information is added by deleted information.The final step is to determine the content of output.In Eq.(5),xtandht-1asσ(·)function parameters are used to obtainot, which performs point-wise multiplication with tanh(ct)to obtainht.With continuous iteration of the algorithm, the simple gradient descent enables weight W,bto gradually approach the optimal solution, after training is completed.Calculatedhtis another result we want to predict.

    Figure 3:LSTM model architecture

    2.3.3 D-LSTM Design

    So that data TA indicators would have a better effect on the actual network, we designed the D-LSTM model shown in Fig.4.As described in Section 2.1, indicators include data Open, Close,High, Low, Volume, and Turnover.The data of the first four groups and transaction volume are of different orders of magnitude.Previous data should be normalized as input to the network before training.As shown in Eq.(6), where x represents the original value, after normalization,all data is compressed into a [-1,1] interval.This normalization method can not only reduce the difference in magnitude of the data, which may disable gradient decline, but also increase convergence speed and prediction accuracy [18].As shown in Fig.4, the increasing number of FC layers helps improve the network’s ability to predict nonlinearity.The output of this network will be used as a member of the final network, which is combined with news N-CNN network for the final calculation.Normalized dataxSis pooled through one FC layer and the LSTM layer,and then through the tanh function layer and finally the FC layer to generate the final predicted value.The loss is calculated as the difference between the normalized predicted value and the actual value.It is worth noting that we spend enough time to train all 1014 stocks in the TA indicator dataset, and make sure that the D-LSTM model is stable enough to be used in the final CNN-LSTM model.

    Figure 4:D-LSTM network

    2.3.4 CNN-LSTM Design

    The CNN is designed to verify the effectiveness of the N-CNN network.When using the network, D-LSTM is trained using dataxS, and N-CNN is trained with dataxN.They are both input into CNN-LSTM.Final text dataxFis different from the time period obtained byxSandxN.DatasetxFcontains indicators of D-LSTM and includes news headlines text of the N-CNN network, as shown in Fig.5.

    Figure 5:CNN-LSTM model architecture

    (1) The CNN network results predict the value of the stock’s rise and fall on next day from news headline marked as [O1,O2].O1shows probability of the stock’s price falling, andO2shows its probability of rising.Due to the property of the softmax function, the two probabilitiesO1+O2= 1.UsingO2as the final compensation formula, ifO2>0.5, it means the next day data HIGH will be rising; ifO2<0.5, it means it will go down.In CNN-LSTM netO2rewrite byYeas a parameter of final compensation formula.

    (2) The result of LSTMYpis the prediction calculated from the trained network D-LSTM based on TA indicator part ofxF,Ypgive a specific price next day.Stock price trend is mainly controlled by shareholders, and shareholders will obviously influenced by news.When the stock news forecast’s probability of rise and fall is accurate enough to predict the trend of the stock price, then the correct value of the final algorithm is obtained by adding the compensation value calculated byO2on basis ofYp, so that ordinary LSTM can further reduce the loss.

    (3) Compensation parameter?

    Based on the D-LSTM model, calculating the highest price of stock on the second day is done by the test set.When set toYt, there is

    N is a number of test sets per stock, andΩis called predictive absolute error.The final predicted valueYfiis calculated by compensation Eq.(8), where?is the weight of the compensation formula.Because only onexFset of datasets used to verify the final prediction result is not enough for network training to obtain the optimal solution, we groupedxFmanually in experimental part and set?to different values in order to select the best result.

    3 Experiment

    This section discusses the actual situation of N-CNN network accuracy using horizontally extended news headline data, D-LSTM, and the test results of CNN-LSTM from final dataxF.

    3.1 Results of N-CNN

    The training set uses cross-entropy loss function.The formula for calculating this is shown in(9).Parameter p indicates the distribution of true tokens, that is, the real price of the label layer on the next day.Parameter q is the predicted distribution of the training model, and it reflects the similarity between the predicted value of text and the actual value.The lower the loss value,the closer the training set to real data.This training dataset model used in this paper results in highly consistent accuracy under different cycle times.However, in this experiment, training set accuracy is a less important criterion than test set accuracy, which is used as the final evaluation standard for proving the generalization ability of CNN in natural language processing (NLP).

    If all dataxNi (0<i ≤k)in dataxN1,xN2,...,xNk(k= 1014) havelx (1 ≤x ≤30)news headlines, thenlxnews is divided into words and converted into input word vectors that are accepted by the CNN.For the initial parameter settings of the N-CNN model, the number of convolutional layer filters is 128, and the fully connected layer is 96 neurons.As can be seen in Tab.2, the experimental results show that cross-entropy of the training set decreases with the increase in training times, which indicates that the performance of the training set is gradually improved, but the constantly increasing accuracy of the test set means a decline in generalization ability.The best cycle time for the N-CNN algorithm is 200, and the test set has the highest accuracy.Therefore, in all subsequent experiments, the number of training cycles of the N-CNN algorithm was 200, and its accuracy reached as high as 73.40%.For an algorithm that tracked only one stock and had an actual sample size of about 100, accuracy was less than 56%.This meant it was feasible to increase the sample size without taking into consideration the sensitivity of different stock news.Algorithms used to predict stock market trends need to have an accuracy of more than 50%.This way, shareholders can rely on news to buy or sell stocks, rather than trying to find useful information from complex data indicators.

    Table 2:Results of N-CNN

    The following experiment was used to verify the impact of data sample size on N-CNN.We divided the 1014 stocks into 20%, 40%, 60%, 80%, and 100% samples, and kept the training set and test set ratio at 8:2.The input news text was also reduced accordingly.The most effective of all other parameters from the previous experiment, including training cycle of the training set,were the same.These experimental results are shown in Tab.3 and Fig.6.

    Table 3:Relationship between sample size and accuracy

    Figure 6:Relationship between sample size and accuracy

    It can be seen that when data volume was below the 30,000 level, sample size is proportional to accuracy.This means that as long as the number of samples obtained is increased, accuracy will inevitably continue to increase until it reaches a level of saturation.This means that theoretically,if the amount of text for all news headlines for stocks listed on the China stock exchange, that is, a level of 10 million, then accuracy will increase significantly.

    The other algorithm models used for forecasting stock prices are shown in Tab.4.The Ding and Wang Y algorithms use the time series of several stocks, which results in different accuracy between different stock data samples.The algorithm proposed in this paper uses the average of all 1014 stocks.

    Table 4:Comparison of accuracy of algorithms

    3.2 Results of D-LSTM

    The results of D-LSTM are similar to forecasting the trend of the next day’s stock market with that day’s basic indicator.The most commonly used equations for evaluating an algorithm are Eqs.(10) to (12)

    This paper uses root-mean-square-error (RMSE) as the loss function for training of the training set in the D-LSTM, because RMSE is a good indicator for verifying the sample prediction error [9].After training with RMSE loss function, we verified the prediction results of the other two indicators for the test set, and used root-mean-absolute-error (RMAE) to verify the deviation of the model system.In Eqs.(10) to (12), N is the number of samples to be calculated,x1,iis the predicted value, andx2,iis the actual value.

    In the D-LSTM model, we set the number of trainings to 200, and the number of neurons in the FC and LSTM layers as 256.In the final CNN-LSTM model, the D-LSTM part of the experimental data used basic indicators of 600718–603993.The results are randomly selected and shown in Tab.5.

    The Shanghai Composite Index has the largest amount of data and seems to be the most stable.The remaining stocks are all individual stocks.Some effects in the D-LSTM algorithm are worse than SSE, and some are better than SSE.The last line shows the average of the results of all the experiments.This average shows that individual stocks and the SSE have similar effects using D-LSTM model, and can reduce the RMSE to a level below 0.02.

    Table 5:Results of D-LSTM

    3.3 Results of CNN-LSTM

    Once we determined that news headlines did play a role in forecasting stock prices on the second day of the stock market, and the accuracy rate was greater than 70%, then we discovered that there was a way to further increase the accuracy of forecast value.An experiment was carried out by the steps described in CNN-LSTM design part.The time of final dataxFis independent of news data and TA data, which can to a great extent verify the final results of the experiment.The results of this experiment are shown is Tab.6.

    Table 6:Results of CNN-LSTM

    The experimental results show that the values of RMSE, RMAE, and mean absolute percentage error (MAPE) are the smallest when?is set to 1.5.The result of?=1.5 is compared with the original D-LSTM experiment, where RMSE, RMAE, and MAPE increased by about 8.3%,about 4%, and about 36.6%, respectively.

    4 Conclusion

    This paper proposed a model for looking at an increased volume of data, consisting of numerous stock news headlines, to improve the accuracy of CNN.Our proposed N-CNN model calculates the difference between predicted stock price values and actual values through the use of a compensation formula, which further improves the results from LSTM in stock price forecasting.

    There are still some things we need to do in the future.First, news headlines lack a standardized time format.Since stock-related news is not subject to standardized regulation, several articles are produced each day, with some stocks having three articles on the same day and some having only one article.As a result, N-CNN input dimensions are not unified, which result in a flaw in training of the CNN model, which in turn reduces the accuracy of the test set.Second,when news headline text is converted into a word vector, numbers will be filtered out.This aspect can successfully filter out information of stock search index that often appears in a news headline,so that the model can focus on information and filter out the value of financial or stock price data, which can improve the accuracy of forecast results.Third, different stocks have different sensitivities to news.If the sensitivity of each stock to news is considered in the N-CNN model,deep learning gives each stock different sensitivity, which will also improve the results.

    Acknowledgement:We thank LetPub (www.letpub.com) for its linguistic assistance during the preparation of this manuscript.

    Funding Statement:This work was supported by the Natural Science Foundation of China(61572160).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    又粗又硬又长又爽又黄的视频| 久久国内精品自在自线图片| 在线免费观看不下载黄p国产| 国产高清三级在线| 波野结衣二区三区在线| 在线观看一区二区三区| av一本久久久久| 国产亚洲精品久久久com| 亚洲精品乱码久久久v下载方式| 人妻一区二区av| 日韩人妻高清精品专区| 一级a做视频免费观看| 国模一区二区三区四区视频| 欧美日韩亚洲高清精品| 久久精品国产亚洲av涩爱| 91午夜精品亚洲一区二区三区| 亚州av有码| 一区二区三区免费毛片| 一级av片app| 久久这里有精品视频免费| 成人亚洲欧美一区二区av| 免费播放大片免费观看视频在线观看| 亚洲,一卡二卡三卡| 超碰97精品在线观看| 男人添女人高潮全过程视频| 纵有疾风起免费观看全集完整版| 五月玫瑰六月丁香| a 毛片基地| 国产精品麻豆人妻色哟哟久久| 欧美日韩综合久久久久久| 美女主播在线视频| 国产免费视频播放在线视频| 亚洲不卡免费看| 国产在线免费精品| 亚洲欧美日韩无卡精品| 国产高清有码在线观看视频| 国产v大片淫在线免费观看| 在线精品无人区一区二区三 | 国产中年淑女户外野战色| 大又大粗又爽又黄少妇毛片口| 肉色欧美久久久久久久蜜桃| 欧美精品国产亚洲| 久久精品夜色国产| 人妻一区二区av| av在线观看视频网站免费| 伦理电影免费视频| 欧美精品亚洲一区二区| h视频一区二区三区| 欧美xxⅹ黑人| 国产精品欧美亚洲77777| 丝袜脚勾引网站| 成人综合一区亚洲| 看十八女毛片水多多多| 日韩欧美 国产精品| 黄色视频在线播放观看不卡| 女人十人毛片免费观看3o分钟| 日本猛色少妇xxxxx猛交久久| 国产精品免费大片| 最近中文字幕高清免费大全6| 春色校园在线视频观看| 久久99热6这里只有精品| 97在线人人人人妻| 亚洲人成网站在线观看播放| 亚洲av二区三区四区| 日韩人妻高清精品专区| 色婷婷av一区二区三区视频| 深爱激情五月婷婷| 五月天丁香电影| 亚洲精品成人av观看孕妇| 99热这里只有精品一区| 国产熟女欧美一区二区| 国产精品欧美亚洲77777| 狂野欧美白嫩少妇大欣赏| 久久精品熟女亚洲av麻豆精品| 国产免费一级a男人的天堂| 国产国拍精品亚洲av在线观看| 麻豆成人午夜福利视频| 国产精品不卡视频一区二区| 久久精品国产自在天天线| 日日摸夜夜添夜夜爱| 狠狠精品人妻久久久久久综合| 五月开心婷婷网| 国产精品福利在线免费观看| 亚洲自偷自拍三级| 王馨瑶露胸无遮挡在线观看| 狂野欧美激情性bbbbbb| 男人狂女人下面高潮的视频| 国产成人a区在线观看| 亚洲欧美日韩东京热| 国内少妇人妻偷人精品xxx网站| 搡女人真爽免费视频火全软件| 狂野欧美激情性xxxx在线观看| 国产精品久久久久久av不卡| 韩国av在线不卡| 久久精品国产鲁丝片午夜精品| 亚洲人成网站在线观看播放| 精品国产一区二区三区久久久樱花 | 人人妻人人爽人人添夜夜欢视频 | 五月伊人婷婷丁香| 国产成人a∨麻豆精品| 免费在线观看成人毛片| 精品一区二区三卡| 国产在线视频一区二区| 久久热精品热| 在线观看三级黄色| av在线播放精品| 日韩一区二区三区影片| 亚洲国产精品国产精品| 国产精品不卡视频一区二区| 制服丝袜香蕉在线| 精品国产三级普通话版| 少妇裸体淫交视频免费看高清| 国产精品嫩草影院av在线观看| 国产v大片淫在线免费观看| av专区在线播放| 99精国产麻豆久久婷婷| 深夜a级毛片| 色哟哟·www| 亚洲va在线va天堂va国产| 国产一级毛片在线| 高清午夜精品一区二区三区| a级毛片免费高清观看在线播放| 美女高潮的动态| 久久久久国产网址| 高清日韩中文字幕在线| 国产美女午夜福利| 国产精品.久久久| 国产精品一及| 人妻系列 视频| 亚洲最大成人中文| 少妇的逼好多水| 久久久久性生活片| 国产高清不卡午夜福利| 麻豆国产97在线/欧美| 狂野欧美白嫩少妇大欣赏| 亚洲久久久国产精品| 亚洲欧美成人综合另类久久久| 夜夜骑夜夜射夜夜干| 精品一品国产午夜福利视频| 精品国产乱码久久久久久小说| 97热精品久久久久久| 成人美女网站在线观看视频| 亚洲美女黄色视频免费看| 91久久精品国产一区二区成人| 乱码一卡2卡4卡精品| 成人毛片a级毛片在线播放| 国产亚洲最大av| 国产一级毛片在线| 亚洲精品aⅴ在线观看| 在线观看国产h片| 久久久久久久大尺度免费视频| 美女脱内裤让男人舔精品视频| 2018国产大陆天天弄谢| 久久99热6这里只有精品| 久久久久久久久久久免费av| 国产亚洲欧美精品永久| 18禁动态无遮挡网站| 国产有黄有色有爽视频| 久久ye,这里只有精品| 国产精品蜜桃在线观看| 午夜福利高清视频| 午夜免费鲁丝| 国产淫语在线视频| 日韩伦理黄色片| 国产精品久久久久成人av| 国产午夜精品一二区理论片| 精品国产乱码久久久久久小说| 我要看黄色一级片免费的| 少妇的逼水好多| 26uuu在线亚洲综合色| 亚洲av免费高清在线观看| 欧美日本视频| 五月玫瑰六月丁香| 中文字幕亚洲精品专区| 天天躁夜夜躁狠狠久久av| 日本av手机在线免费观看| 国产日韩欧美亚洲二区| 国产 一区精品| 性色avwww在线观看| 欧美变态另类bdsm刘玥| 亚洲国产av新网站| 黄色视频在线播放观看不卡| 久久久久国产网址| 国产午夜精品久久久久久一区二区三区| 最近最新中文字幕免费大全7| 国产精品免费大片| 欧美xxxx性猛交bbbb| 国产探花极品一区二区| 欧美+日韩+精品| av国产免费在线观看| 2021少妇久久久久久久久久久| 秋霞伦理黄片| 亚洲欧美精品专区久久| 久久久久精品性色| av福利片在线观看| 综合色丁香网| av视频免费观看在线观看| 3wmmmm亚洲av在线观看| 秋霞在线观看毛片| 精品熟女少妇av免费看| 久久 成人 亚洲| 美女中出高潮动态图| 亚洲成人中文字幕在线播放| 97精品久久久久久久久久精品| 大香蕉久久网| 国产高清国产精品国产三级 | 亚洲欧美成人综合另类久久久| 下体分泌物呈黄色| 国产熟女欧美一区二区| 日日摸夜夜添夜夜爱| 亚洲丝袜综合中文字幕| 一级毛片 在线播放| 日韩一区二区三区影片| 永久免费av网站大全| 日本av手机在线免费观看| 国产成人精品一,二区| 国产精品偷伦视频观看了| 亚洲综合精品二区| 免费大片黄手机在线观看| 九色成人免费人妻av| 国产精品蜜桃在线观看| 国产黄色免费在线视频| 能在线免费看毛片的网站| 欧美高清成人免费视频www| 日本wwww免费看| 久久精品夜色国产| 国产av精品麻豆| 国产高清国产精品国产三级 | 噜噜噜噜噜久久久久久91| 久久精品国产a三级三级三级| 日韩av在线免费看完整版不卡| 久久韩国三级中文字幕| 久久av网站| 亚洲成人中文字幕在线播放| 精品久久国产蜜桃| 91久久精品电影网| 国产一区二区在线观看日韩| 少妇猛男粗大的猛烈进出视频| tube8黄色片| 国产女主播在线喷水免费视频网站| 一级毛片电影观看| 日本黄色片子视频| 看免费成人av毛片| 中国三级夫妇交换| 在现免费观看毛片| 国产伦精品一区二区三区四那| freevideosex欧美| 综合色丁香网| 久久精品国产亚洲av天美| 欧美日韩综合久久久久久| 午夜激情久久久久久久| 亚洲欧美一区二区三区国产| 中国美白少妇内射xxxbb| 精品国产乱码久久久久久小说| 性色avwww在线观看| 久久婷婷青草| 亚洲熟女精品中文字幕| 91久久精品国产一区二区成人| 亚洲精品乱码久久久v下载方式| 成人亚洲欧美一区二区av| www.色视频.com| 高清不卡的av网站| 亚洲成人一二三区av| 亚洲无线观看免费| 精华霜和精华液先用哪个| 综合色丁香网| 免费黄网站久久成人精品| 国产白丝娇喘喷水9色精品| 国产成人91sexporn| 精品久久久精品久久久| 乱码一卡2卡4卡精品| 纵有疾风起免费观看全集完整版| 精品一区在线观看国产| 麻豆乱淫一区二区| 久久精品国产亚洲av涩爱| 日韩制服骚丝袜av| 午夜激情福利司机影院| 成人毛片a级毛片在线播放| 最后的刺客免费高清国语| 国产探花极品一区二区| 天堂8中文在线网| 在线观看免费视频网站a站| 99热网站在线观看| 亚洲图色成人| 免费播放大片免费观看视频在线观看| 女的被弄到高潮叫床怎么办| 精品人妻视频免费看| 亚洲人与动物交配视频| 国产永久视频网站| 熟女人妻精品中文字幕| 91久久精品电影网| av女优亚洲男人天堂| 成人亚洲欧美一区二区av| 国产av国产精品国产| 蜜桃在线观看..| 一边亲一边摸免费视频| 联通29元200g的流量卡| 亚洲av福利一区| 亚洲国产精品国产精品| 99久久精品热视频| 99久久人妻综合| 一区在线观看完整版| 久久久久国产精品人妻一区二区| 久久久久久九九精品二区国产| 高清欧美精品videossex| 成人亚洲精品一区在线观看 | 熟妇人妻不卡中文字幕| 精品国产乱码久久久久久小说| 99热这里只有是精品在线观看| 国产永久视频网站| 亚洲欧美一区二区三区国产| 日韩人妻高清精品专区| 狂野欧美白嫩少妇大欣赏| 国产精品一区二区性色av| 亚洲va在线va天堂va国产| 女性生殖器流出的白浆| 久久国产亚洲av麻豆专区| 欧美人与善性xxx| 成人影院久久| 欧美老熟妇乱子伦牲交| 国产精品爽爽va在线观看网站| 美女中出高潮动态图| 日韩三级伦理在线观看| 国产高清三级在线| 精品午夜福利在线看| videos熟女内射| 亚洲欧洲国产日韩| 亚洲美女搞黄在线观看| 视频区图区小说| 女的被弄到高潮叫床怎么办| 少妇丰满av| 欧美成人精品欧美一级黄| 亚洲美女黄色视频免费看| 国产人妻一区二区三区在| 一个人看视频在线观看www免费| 岛国毛片在线播放| 精品少妇黑人巨大在线播放| 99热国产这里只有精品6| 日本欧美国产在线视频| 日韩三级伦理在线观看| 涩涩av久久男人的天堂| 亚州av有码| 免费黄频网站在线观看国产| 国产一区亚洲一区在线观看| 在线观看人妻少妇| av免费在线看不卡| 国产一区有黄有色的免费视频| 99热网站在线观看| 少妇精品久久久久久久| 又粗又硬又长又爽又黄的视频| 亚洲人成网站高清观看| 一级片'在线观看视频| tube8黄色片| 国产精品国产三级国产av玫瑰| 又爽又黄a免费视频| 国产伦精品一区二区三区视频9| 国产精品伦人一区二区| 中文欧美无线码| 日本黄色片子视频| 日本午夜av视频| 国产无遮挡羞羞视频在线观看| 国产精品福利在线免费观看| 国产精品久久久久久久久免| 少妇丰满av| 熟妇人妻不卡中文字幕| 久久久久久伊人网av| 伊人久久国产一区二区| 黑人高潮一二区| 亚洲av在线观看美女高潮| av播播在线观看一区| 在线观看免费高清a一片| 日韩成人av中文字幕在线观看| 欧美 日韩 精品 国产| 国产有黄有色有爽视频| 精品一区在线观看国产| 亚洲va在线va天堂va国产| 80岁老熟妇乱子伦牲交| 少妇人妻 视频| 国产成人a∨麻豆精品| 日韩在线高清观看一区二区三区| www.色视频.com| 香蕉精品网在线| 最黄视频免费看| 纯流量卡能插随身wifi吗| 精华霜和精华液先用哪个| 欧美 日韩 精品 国产| 看非洲黑人一级黄片| 99热这里只有精品一区| 一级毛片 在线播放| 国产高清三级在线| 亚洲欧洲日产国产| 人妻一区二区av| 综合色丁香网| 人人妻人人澡人人爽人人夜夜| videossex国产| 99精国产麻豆久久婷婷| 国产片特级美女逼逼视频| 国产高潮美女av| 日韩 亚洲 欧美在线| 午夜福利在线在线| 欧美成人精品欧美一级黄| 最新中文字幕久久久久| 91精品国产九色| 日日啪夜夜撸| 午夜福利影视在线免费观看| 99国产精品免费福利视频| 高清日韩中文字幕在线| 成年av动漫网址| 久久久午夜欧美精品| 久久精品夜色国产| 欧美精品人与动牲交sv欧美| 嫩草影院入口| 人妻制服诱惑在线中文字幕| 男女边吃奶边做爰视频| 少妇熟女欧美另类| 久久久久久久久久人人人人人人| 日本猛色少妇xxxxx猛交久久| 女的被弄到高潮叫床怎么办| 日韩欧美 国产精品| 能在线免费看毛片的网站| 看免费成人av毛片| 成人18禁高潮啪啪吃奶动态图 | 国产精品三级大全| 精品少妇久久久久久888优播| h视频一区二区三区| 97超视频在线观看视频| 亚洲成人一二三区av| 男女免费视频国产| 国精品久久久久久国模美| 91精品一卡2卡3卡4卡| 18禁裸乳无遮挡动漫免费视频| 亚洲人成网站在线播| 久久久午夜欧美精品| 高清欧美精品videossex| 大陆偷拍与自拍| 免费观看无遮挡的男女| 亚洲性久久影院| 有码 亚洲区| 国产精品麻豆人妻色哟哟久久| 欧美bdsm另类| 国产美女午夜福利| 插逼视频在线观看| 高清不卡的av网站| 伊人久久精品亚洲午夜| 超碰av人人做人人爽久久| 欧美日本视频| 有码 亚洲区| 伊人久久国产一区二区| 男的添女的下面高潮视频| 午夜福利影视在线免费观看| h日本视频在线播放| 久久人人爽av亚洲精品天堂 | 一本—道久久a久久精品蜜桃钙片| 精品国产三级普通话版| 一本久久精品| 亚洲中文av在线| 狂野欧美激情性xxxx在线观看| 联通29元200g的流量卡| 97超视频在线观看视频| 国产69精品久久久久777片| 久久国产乱子免费精品| 欧美xxxx性猛交bbbb| 午夜福利在线观看免费完整高清在| 好男人视频免费观看在线| 成人亚洲精品一区在线观看 | 久久精品国产亚洲av天美| 亚洲精品日韩av片在线观看| 在线免费观看不下载黄p国产| 免费观看在线日韩| 国产免费视频播放在线视频| 亚洲av二区三区四区| 日韩强制内射视频| 亚洲精品中文字幕在线视频 | 国产精品久久久久久精品电影小说 | 国产精品一区二区性色av| 欧美成人一区二区免费高清观看| 国产黄片美女视频| 在线观看免费高清a一片| 亚洲色图av天堂| 日韩成人av中文字幕在线观看| 亚洲欧美成人精品一区二区| 丰满乱子伦码专区| 久久久久久久久久久丰满| 亚洲美女视频黄频| 亚洲av不卡在线观看| 亚洲美女黄色视频免费看| 久久久久精品性色| 丰满人妻一区二区三区视频av| 欧美 日韩 精品 国产| 亚洲真实伦在线观看| 少妇熟女欧美另类| 国产成人精品福利久久| 成年美女黄网站色视频大全免费 | 欧美xxⅹ黑人| 欧美xxxx黑人xx丫x性爽| 伦理电影大哥的女人| 在线观看免费高清a一片| 久久国产精品大桥未久av | 麻豆乱淫一区二区| 99热这里只有是精品在线观看| 亚洲经典国产精华液单| 一级二级三级毛片免费看| 日本欧美视频一区| 少妇丰满av| 国产爽快片一区二区三区| 婷婷色综合大香蕉| 99热这里只有精品一区| 成人国产麻豆网| 在线亚洲精品国产二区图片欧美 | 天堂俺去俺来也www色官网| 欧美精品人与动牲交sv欧美| 久久国内精品自在自线图片| 我的老师免费观看完整版| av女优亚洲男人天堂| 汤姆久久久久久久影院中文字幕| 亚洲精品第二区| 国产亚洲5aaaaa淫片| 国产乱来视频区| 国产极品天堂在线| 麻豆精品久久久久久蜜桃| 国产永久视频网站| 麻豆精品久久久久久蜜桃| 国产黄色免费在线视频| 三级国产精品片| av在线观看视频网站免费| 三级国产精品片| 少妇人妻一区二区三区视频| 欧美日韩亚洲高清精品| 亚洲欧洲国产日韩| 内地一区二区视频在线| 日韩伦理黄色片| 91精品国产九色| 久久精品国产自在天天线| 香蕉精品网在线| 日本色播在线视频| 中文字幕亚洲精品专区| 菩萨蛮人人尽说江南好唐韦庄| 欧美成人午夜免费资源| 草草在线视频免费看| tube8黄色片| 日韩中文字幕视频在线看片 | 精品一区二区免费观看| av线在线观看网站| av免费在线看不卡| 欧美极品一区二区三区四区| 日韩制服骚丝袜av| 免费少妇av软件| 精品久久国产蜜桃| 成人美女网站在线观看视频| 亚洲怡红院男人天堂| 亚洲av不卡在线观看| 午夜日本视频在线| 亚洲精品日本国产第一区| 成人毛片60女人毛片免费| 99热全是精品| 日本av免费视频播放| 美女视频免费永久观看网站| 亚洲美女黄色视频免费看| 91精品国产九色| 日日撸夜夜添| 黄色欧美视频在线观看| 多毛熟女@视频| 欧美人与善性xxx| 少妇猛男粗大的猛烈进出视频| 最近中文字幕2019免费版| 2021少妇久久久久久久久久久| 国内少妇人妻偷人精品xxx网站| 激情五月婷婷亚洲| 99热这里只有是精品在线观看| av国产免费在线观看| 免费看日本二区| 成人黄色视频免费在线看| 国产乱人偷精品视频| 久久久国产一区二区| 亚洲三级黄色毛片| 2018国产大陆天天弄谢| 美女cb高潮喷水在线观看| 日本黄大片高清| 国产视频内射| 一区在线观看完整版| 高清在线视频一区二区三区| 一边亲一边摸免费视频| 在线播放无遮挡| 亚洲色图综合在线观看| 国产精品久久久久久久电影| 国产精品99久久99久久久不卡 | 国产成人免费无遮挡视频| 日本免费在线观看一区| 国产精品久久久久久精品电影小说 | 久久人人爽人人爽人人片va| 国产av一区二区精品久久 | 国产v大片淫在线免费观看| 成人国产av品久久久| 国产午夜精品一二区理论片| 欧美zozozo另类| 1000部很黄的大片| 一级av片app| 国产精品人妻久久久影院| 欧美激情国产日韩精品一区| 日本一二三区视频观看| 国产精品偷伦视频观看了| 极品少妇高潮喷水抽搐| 国产高清国产精品国产三级 | 国产精品熟女久久久久浪| 国产一区二区三区av在线| 尤物成人国产欧美一区二区三区| 国产在线男女| 国产高清国产精品国产三级 | 国产精品av视频在线免费观看| 一级爰片在线观看| 成人18禁高潮啪啪吃奶动态图 | 亚洲精品久久午夜乱码| 久久国产精品大桥未久av | 热re99久久精品国产66热6| 一区二区av电影网|