• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A deep reinforcement learning (DRL) based approach for well-testing interpretation to evaluate reservoir parameters

    2022-03-30 13:52:22PengDongZhiMingChenXinWeiLioWeiYu
    Petroleum Science 2022年1期

    Peng Dong ,Zhi-Ming Chen ,,* ,Xin-Wei Lio ,Wei Yu

    a State Key Laboratory of Petroleum Resources and Prospecting,China University of Petroleum Beijing (CUP),Beijing,102249,China

    b University of Texas at Austin,Austin,78731,Texas,USA

    Keywords:Well testing Deep reinforcement learning Automatic interpretation Parameter evaluation

    ABSTRACT Parameter inversions in oil/gas reservoirs based on well test interpretations are of great significance in oil/gas industry.Automatic well test interpretations based on artificial intelligence are the most promising to solve the problem of non-unique solution.In this work,a new deep reinforcement learning(DRL)based approach is proposed for automatic curve matching for well test interpretation,by using the double deep Q-network (DDQN).The DDQN algorithms are applied to train agents for automatic parameter tuning in three conventional well-testing models.In addition,to alleviate the dimensional disaster problem of parameter space,an asynchronous parameter adjustment strategy is used to train the agent.Finally,field applications are carried out by using the new DRL approaches.Results show that step number required for the DDQN to complete the curve matching is the least among,when comparing the naive deep Q-network (naive DQN) and deep Q-network (DQN).We also show that DDQN can improve the robustness of curve matching in comparison with supervised machine learning algorithms.Using DDQN algorithm to perform 100 curve matching tests on three traditional well test models,the results show that the mean relative error of the parameters is 7.58%for the homogeneous model,10.66%for the radial composite model,and 12.79%for the dual porosity model.In the actual field application,it is found that a good curve fitting can be obtained with only 30 steps of parameter adjustment.

    1.Introduction

    Well test analysis plays an important role in understanding the characteristics of wellbore,reservoir and boundary,especially in the discovery and development of some important oil and gas fields(Yao and Ge,2011;Chen et al.,2018;Mohammed et al.,2020).One of the most important means of well test interpretation is the type curve match.In the early days,curve matching was performed based on the chart(Earlougher,1977;Horne,1995).The parameters are obtained by moving the measured curve to find the most consistent theoretical curve on the chart.Since the number of curves on the chart is limited,this method will cause large errors.At present,manual parameter adjustment assisted by computer software become the primary method for curve matching(Bourdet,2002).However,the human bias,non-unique solutions,and low efficiency caused by artificial interpretation cannot adapt to the rapid development of oil and gas fields (AlMaraghi and El-Banbi,2015).

    In recent decades,with the improvement of calculation technology and test instrument precision,many optimization algorithms have been gradually applied to the automatic matching of well test curve.Among them,nonlinear regression is the classical method,and the least square method represented by Levenberg-Marquardt method is the most common solution method (Nanba and Horne,1992;Dastan,2010;Dastan and Horne,2011).However,the method is greatly affected by the initial value,and the parameters obtained are locally optimal.Therefore,some global optimization algorithms,such as genetic algorithm and particle swarm optimization algorithm,are also applied to the automatic matching of well test curve (Guyaguler et al.,2001;Gomez et al.,2014;Awotunde,2015).But the global optimization algorithm has the problem of low efficiency.

    With the great progress made in computer science in recent years,artificial intelligence (AI) algorithms have been widely used in the oil and gas industry (Zhu et al.,2019;Liu et al.,2020;Gao et al.,2021;Huang et al.,2021).Meanwhile,AI algorithms are also tried for automatic well test interpretation.Al-Kaabi and Lee(1990)firstly deployed artificial neural network (ANN) in well test model recognition.They used a number of independent ANN to calculate the probability that the curve fell into each type of well test model.Adibifard et al.(2014)trained an ANN to automatically identify well test curve parameters.The coefficients interpolated by Chebyshev polynomial on pressure derivative data are used as the input of ANN to improve the accuracy of parameter identification.Li et al.(2020) used the convolutional neural network to train the deep learning model for identifying curve parameters by taking the complete curve as input,without manually extracting curve features.In addition,the application of deep learning on well-test interpretation for identifying pressure behavior and characterizing reservoirs was also performed (Dong et al.,2021).However,training neural network requires the acquisition or synthesis of a large number of samples,and the interpretability of the parameter inversion results is weak.Sometimes unrealistic interpretation results are prone to appear and the reasons cannot be known(Zhang and Zhu,2018).To make the parameter inversion results more interpretable and reasonable,this work proposes for the first time a robust automatic well test curve matching method based on reinforcement learning (RL).

    RL is inspired by relevant principles of animal psychology(Sutton and Barto,2018).By imitating the trial-and-error mechanism of humans or animals,the agent can interact with the environment and learn the mapping relationship between state and behavior to obtain the maximum cumulative expected return.Compared with supervised learning algorithms,RL algorithms have the potential to achieve results beyond human performance through active learning,exploration and exploitation.Because of this,RL tends to be a harder learning task.Among RL algorithms,Q learning is one of the most popular(Gao et al.,2020).Recently,the original DRL algorithm naive DQN and its improved algorithm DQN,which combines Q learning with deep neural network,have been introduced and applied into Atari games to achieve automatic control at or beyond the human level (Mnih et al.,2013,2015).However,these two algorithms lead to overoptimistic value estimates,so van Hasselt et al.(2016) proposed DDQN algorithm to alleviate this problem.DDQN algorithm has been successfully used for battery energy storage system,power system,and stock trading(Arulkumaran et al.,2017;Bui et al.,2019;Shi et al.,2021).These studies show that the DDQN algorithm can avoid the agent trapped into local optimization and is suitable for the environment with large state space.

    At present,only limited work about RL has been done in the oil and gas industry.Hourfar et al.(2019) used the DRL algorithm to optimize reservoir water injection.By allowing the agent to dynamically adjusting the water injection rate,higher NPV can be obtained than the traditional optimization method.Miftakhov et al.(2020) use reservoir pressure and water saturation distribution as direct observations to train agents to optimize injection and production parameters.Li and Misra (2020) transformed the history matching problem into a continuous decision problem,and adapted the reservoir permeability using a RL algorithm to achieve automatic history fitting.Guevara et al.(2018)used RL to optimize the gas injection strategy in steam-assisted gravity drainage process.The field test results show that the RL method increases the NPV by at least 30%and reduces the calculation cost by more than 60%.Unfortunately,although the RL has lots of advantages,few work about RL has been done in well test interpretation.

    In this study,the DDQN algorithm based agent learns to match the well test data in optimal number of steps by iteratively adjusting the parameters of well test model.This work represents,to our knowledge,the first application of DRL approach for automatically match the well test curves.The new proposed method has a fast speed and reliable results in automatic well test interpretation,which is of great significance to improve the repeatability of well test interpretation.

    2.Theory

    2.1.Reinforcement learning

    The reinforcement learning (RL) is an important paradigm of machine learning,whose goal is to find an optimal strategy to obtain the largest cumulative expected return by train an agent.Markov Decision Process(MDP)is a basic theoretical framework to solve the problem of RL.Within this framework,machines that learn and implement decisions are called agent.Anything outside the agent that interacts with it is called the environment.In the interaction process,the agent observes the statesof the current environment and chooses the actionan under a certain policy π,the environment responds to the action,and the new states′and rewardrwill be fed back to the agent.Therefore,assuming that starting from the initial states0,executing the MDP will result in a sequence,s0,a0,r0,s1,a1,r1,...,sn,an,rn.

    The agent's job is to optimize the policy for taking action to maximize the cumulative expected return(Sutton and Barto,2018).The return at steptis the sum of the discount rewardswhere γ∈[0,1 is discount rate,which determines the present value of future rewards.In RL,it is the most important method to train agents for solving MDP problem based on action value functionQπ(s,a).Qπ(s,a)represents the expected return on the actionataken in accordance with policy π at states,as in Eq.(1).

    Qπ(s,a)calculates the value of the action in a certain state.Simply speaking,Qπ(s,a)expresses how good it is for an agent to be in a certain state (Sun,2020).Therefore,the optimal strategy is based on the optimal value of the action.Specifically,when the optimal action value functionQ*(s,a)=is obtained,the optimal policy π*(s)=(s,a)is to pick the action corresponding to the maximumQ*(s,a)in each state.In general,Q*(s,a)can be solved by Bellman optimality equation (Sutton and Barto,2018),as Eq.(2),which shows the relationship between the current optimal action value function and the subsequent optimal action value function.

    wheres′is the next state reached after taking actiona,anda′is the action taken in the next state.As long as iterates on Eq.(2),it will eventually converge to the optimal action value functionQ*(s,a)(Watkins and Dayan,1992).

    However,the action value function described above can only represent discrete states.For a continuous state space,like the well test curve,the function approximator,denoted asQ(s,a;θ)≈Q*(s,a),must be used to estimate the action value function.Q(s,a;θ)is usually designed as a neural network (NN) with weights θ,as shown in Fig.1,which method is called deep reinforcement learning(DRL).This study uses the DDQN algorithm,a value based DRL algorithm,proposed by van Hasselt et al.(2016) to train the agent for curve matching,and the details will be discussed in Section 3.

    Fig.1.Schematic diagram of NN function approximator.Wherein,the input layer accepts the state s consisting of the target and the predicted derivative pressure curve,and the output layer returns the action value Q(s,a).

    2.2.Well test model

    In this work,three classical well test models are used to construct an environment for training agents respectively,and the performance of agents in different environments is evaluated to verify the generalization ability of the proposed method.The three well test models are:homogeneous model,dual-porosity model,and radial composite model.To facilitate analysis,the parameters are dimensionless and are defined in Appendix 1.This section briefly introduces these three well test models.

    2.2.1.Homogeneous model

    It is assumed that a production well in homogeneous infinite formation is affected by wellbore storage effect and skin effect.According to the solution of Lee et al.(2003),the dimensionless bottom hole pressurepwDis a function oftD,CD,andS,which can be denoted as follows:

    whereCDis the dimensionless wellbore storage,Sis the skin factor,tDis the dimensionless time.Therefore,the required inversion parameters areCD,S.A typical log-log plots of bottomhole pressure and its derivative curve is shown in Fig.2.The derivative curve type used in this work is Bourdet pressure derivative curve (Bourdet et al.,1984).

    Fig.2.The typical well test curves for homogeneous model.

    2.2.2.Dual-porosity model

    It is assumed that there are two types of pore media in the formation:the fracture system,which is the fluid flow channel,and the matrix rock system,which is the fluid reservoir space.According to the solution of Lee et al.(2003),thepwDis a function oftD,CD,S,ω,and λ,which can be expressed as follows:

    where ω is the storativity ratio,which represents the storage capacity of the fracture;λ is the inter-porosity flow factor,which represents the communication between the fracture and the matrix;φ is porosity;Ctis the total compressibility,MPa-1;ais the parameter characteristic of the system geometry;rwis the well radius,m;kis permeability,mD;Subscript m represents matrix;Subscript f represents fracture.Therefore,the parameters need to be inverted areCD,S,ω,and λ.A typical log-log plots of pressure and its derivative curve is shown in Fig.3.

    2.2.3.Radial composite model

    It is assumed that there are two seepage zones with different physical parameters in the formation and there is no additional pressure drop at the interface of the two zones.According to the solution of Chu and Shank(1993),ThepwDis a function oftD,CD,S,M,andRi,which can be denoted as follows:

    with

    Fig.3.The typical well test curves for dual-porosity model.

    Fig.4.The typical well test curves for radial composite model.

    whereMis mobility ratio;Riis inner zone radius,m;μ is viscosity,mPa·s;Subscript ir represents inner zone;Subscript er represents outer zone.Therefore,the parameters to be inverted areCD,S,M,andRi.A typical log-log plots of pressure and its derivative curve is shown in Fig.4.

    3.Method

    Based on concepts of DRL,the well test curve matching process can be regarded as an MDP.Therefore,the automated well test curve matching process can be achieved by applying DRL algorithm.Fig.5 is a schematic diagram of automatic curve matching process based on MDP.In this process,the target and predicted pressure derivative curve are considered as statestat first.Here,the target curvecomes from the measured curve,and the predicted curveis calculated from the well test model.Following that,the agent provides the actionatof parameter adjustment and updates the policy based on the rewardrt.Next,the environment based on the well test (WT) model receives actionatand then updates the parameters of the WT model.Finally,the environment outputs new statesst+1and rewardsrt+1back to the agent.In this way,the agent continuously interacts with the environment to obtain more rewards,which will make the prediction parameters continuously update to the target parameters to complete the curve matching.

    In the remainder of this section,we will detail the agent and environment of the study in turn.

    Fig.5.Schematic diagram of agent interacting with the environment in MDP.The environment is encapsulated by the well test model (WT model).

    3.1.Agent

    This section will introduce the double deep Q-network(DDQN)algorithm used to train the agent,the policy used when the agent interacts with the environment,and the discrete mode of action space to avoid the curse of dimensionality in the parameter adjustment process.

    3.1.1.DDQN algorithm

    As noted earlier,in this work the double deep Q-network(DDQN)(van Hasselt et al.,2016)algorithm is considered.DDQN is a value-based DRL algorithm that can be used with continuous state spaces.DDQN uses the policy based on the online Q-networkQ(s,a;θ)to select actions that interact directly with the environment,and the target Q-networkQ(s,a;θ′)to evaluate the selected actions separately,as shown in Fig.6.Separating the action selection from the evaluation can alleviate the overly optimistic estimate of the value of the action and improves the accuracy of the action value estimation.In addition,to improve the utilization efficiency of data,the transition tuples at timestept,et=(st,at,rt,)generated by the interaction between the agent and the environment is stored in a buffer called experience replay memory Dt={e1,e2,...,et} (Mnih et al.,2015),as Fig.6.Therefore,by extracting experiments randomly in D,it is possible to use minibatch samples to update these two Q-networks.More precisely,the loss function,as Eq.(6),is used to updates the weights θ of online Q-networks,while the weights θ’of target Q-network is a delayed copy of the online Qnetworks,which is to copy after a certain number of training.The complete training process is discussed in Section 3.3.

    where θ is the weight of online Q-networks,θ′is the weight of target Q-networks,γ is discount factor.The notation(s,a,r,s′)U(D)represents a minibatch samples are uniformly sampled from the replay memory.

    Once the training of the agent is completed,the online Q-networks is used for prediction.For a particular state,the online Qnetworks will provide a Q-value for each of the possible actions for that state.Finally,the action is determined by the policy in terms of Q-value,which will be discussed in detail later.

    3.1.2.Policy

    Fig.6.Schema of the DDQN model in training process.

    When agents interact with the environment,an important problem is to trade-off exploration and exploitation.Exploitation means that the agent takes the most valuable action it estimates every time,and exploration means that the agent randomly takes actions to explore the environment.Proper exploration will help the agent find better policy and improve the accuracy of the action value estimation.However,the agent cannot simultaneously exploration and exploitation in the same action selection.Therefore,the selection of action needs to obey a certain probability distribution to make exploration and exploitation alternate.In Section 5.1,we discussed the influence of different policies on the training process.The results show that Boltzmann exploration policy(Derhami et al.,2008)can better balance the exploration and exploitation of agents in the curve matching task.In the Boltzmann exploration policy,the probability distribution of the agent's action selection is determined by Eq.(7).

    where τ ∈(0,1 is the temperature parameter,which controls the certainty of the action.

    3.1.3.Curse of dimension

    The DRL algorithm based on value functionQ(s,a)needs to clarify the actions that an agent can perform,which means that the action space of the agent is discrete.Therefore,when it is applied to a high-dimensional continuous action space,such as well test curve matching problem,the continuous action needs to be discrete.However,this will lead to huge combinatorial increase in the number of actions with the number of well-test parameter dimensions.For example,for the homogeneous model,it is assumed that the parameters to be inverted are dimensionless well storage coefficientCDand skin coefficientS.If the parameter space is discretized into 100 intervals,the action space of the two parametersCDandSwill reach 1002.For the agent,the value of this 1002action needs to be calculated,that is,the output dimension of the Q network is 1002.Generally,if the number of parameters to be inverted isNand the number of parameter discrete spaces isnd,then the total action space to be considered isWhen the parameter precision requirement is high or the number of parameters is large,the agent will not be able to handle the resulting huge action space.To alleviate the problem,an asynchronous parameter tuning method is proposed.To be more specific,the agent only adjusts one well test parameter at a time,and the action for each parameter is only set to increase and decrease the value of the parameter,and the step length of the increase or decrease is set to a fixed value.Obviously,the total action space that needs to be considered is only 2Nbased on this approach.This allows the current discrete action DRL algorithm to be applied to well test curve matching task.In this work,three typical well test models are used to verify the proposed method.The parameter range and action space of each well test model are shown in Table 1.In addition,the reward weights are given for different parameters to make the agent pay more attention to those parameters that are important in reflecting the reservoir characteristics,as shown in Table 1.In the next section,we will discuss how these weights are used.For the action step length,it can be set as follows.

    Table 1 Action spaces in three environments based on well test models.

    a) The well test model parameters ψ are uniformly distributed in the interval:

    b)The sensitivity of the double log curve to the parameter values in different ranges is different.We hope to have smaller step length in more sensitive areas.Therefore,some well test model parameters ψ are exponentially distributed in the interval:

    In this work,theSis set to uniformly distributed,and other parameters are set to exponentially distributed.ndis set to 100.

    3.2.Environment

    After the agent takes an action,its state will change and the environment will generate feedback (reward).The environment is constructed based on the test well model and is designed for the agents to interact with it for curve matching.In an environment,the state and reward provided to the agent are two basic elements.For the state,we define it as in Eq.(10).

    wherew(i)is the reward weight of theith well test model parameter,and its value is shown in Table 1.The interpretation results of important or insensitive parameters can be improved by setting different reward weights.δ is the maximum error when curve matching is completed.The definition oferris as follows:

    Equation(12)indicates that when the agent takes an action,the parameter error becomes smaller and the reward will be positive.Otherwise,the reward will be negative.In addition,negative rewards need to be greater than positive rewards,which helps the agent reach its goal faster and reduce unnecessary actions(Wiewiora,2003).The quality of reward design significantly affects the performance of the agent.Therefore,in Section 5.2,we discuss the influence of different reward design methods on the results.

    3.3.Training process and implementation details

    Algorithm 1 is the training process for the proposed method to accomplish automated pressure derivative curve matching.The process starts with an initial target pressure derivative curve at the beginning of each episode.We trained the agent to perform a total of 200 episodes of curve matching.The termination condition of each episode is that the curve matching is completed or the number of timesteps exceeds 300.To allow the agent to fully explore the environment,the agent's action in the first ten episodes are randomly sampled.After completing the exploration,the agent enters the training state.

    Algorithm 1.The training process of an agent in this work

    Fig.8.The result of 100 times curve matching of the agent on the homogeneous model.

    Fig.9.The result of 100 times curve matching of the agent on the dual porosity model.

    In this work,the Adam algorithm (Kingma and Ba,2014) was used to updates the weight θ of the online Q-network by minimizing the loss function in Eq.(6) with a learning rate of 0.0001.The weight θ′of target Q-network was updated every 1000 timesteps by the delayed copy.The discount factor γ in Eq.(6) is set to 0.99.The capacity of the replay memory is 105,and the minibatch size is set to 128.The temperature parameter τ in Eq.(7)is 1 during the training and 10 during the inference.

    The Q-network used in this work is a neural network composed of three hidden layers,and the number of neurons in each layer is 500,500,300,respectively.The hidden layer activation function is ReLU,while the output layer is not activated.The input layer takes the state with a feature number of 80 as input,and the output layer gets the Q-value for each action.The output dimension is determined by the environment.In this work,the output dimension based on the homogeneous model environment is 4,and others environment is 8.The agent is implemented based on PyTorch and trained on NVIDIA 1060Ti graphic processing unit.The average training time was 26 min,and the average time for each curve matching was only 2.86 s.In addition,the raw data,due to the data noise from human factors,measuring instruments,and production conditions,which will affect the results of automatic parameter inversion.In this work,wavelet threshold denoising method(Valencia et al.,2016) is used to denoise the original pressure and derivative data.

    4.Method verification

    4.1.Result verification

    To show the reliability of DDQN,Fig.7 compares the performance of DDQN on different well test models.To avoid the influence of randomness on the results (Mania et al.,2018;Henderson et al.,2018),we perform 5 random simulations and draw the training curve with 95%confidence interval.The results show that the agent trained based on the DDQN algorithm takes the fewest timesteps to complete the curve matching and obtains the highest reward in the homogeneous model.However,for the dual porosity and radial composite model,the agent has similar performance in both.This indicates that the number of parameters to be inverted,that is,the size of the action space,has a greater impact on the performance of the agent than the well test model itself.

    Fig.10.The result of 100 times curve matching of the agent on the radial composite model.

    In order to verify the reliability of the proposed method,the trained agent based on DDQN algorithm was asked to complete 100 times of curve matching on different well testing models,and the error of the finally obtained parameters was analyzed.In the inference process,the agent only selects actions and does not update the parameters of its action value network Q.The parameter inversion results on the three well test models are shown in Figs.8-10.The results show that the predicted parameters have a pronounced correlation with the actual parameters (consideringR2),which proves that the trained agent are able to invert the curve parameters accurately.Furthermore,we can observe that the accuracy of parameter inversion on the homogeneous model is higher than the other two models,which indicates that the parameter inversion error increases correspondingly with its number.

    We performed further statistical analysis of the errors in the parameter inversions.Table 2 is the statistical indicators used and their calculation formulas.The statistic results of parameter errors are reported in Table 3.Concretely,the mean of relative error(MRE)of the parameters is 7.58%for the homogeneous model,10.66%for the radial composite model,and 12.79% for the dual porosity model.In addition,the median is substantially lower than the mean,as Table 3,which indicates that the accuracy of the parameter inversion is quite high without considering the effect of extreme values.Standard deviation of relative error reflects the effect of parameter sensitivity on the inversion results.We can observe that the standard deviation of relative error of the skin factors (S) is generally larger due to its low sensitivity.Moreover,mean absolute error(MAE)and root mean square error(RMSE)are calculated to visualize the magnitude of the error,as shown in Table 3.Finally,coefficient of determinationR2are calculated to synthetically evaluate the performance of the parameter inversion.The results show that the agents perform well in these three models,and the error of parameter inversion and the parameter inversion results balance accuracy and stability.

    Table 2 The calculation formula of evaluation index. K is the number of inferences.

    Table 3 Error statistical of parameter inversion results.

    Table 4 Performance scores for all models.

    4.2.Result comparison

    In this section,to further show the advantages of DDQN,we compare its results with other two DRL algorithms and three classicmachine learning (ML) algorithms.

    The other two DRL algorithms are Naive DQN(Mnih et al.,2013),DQN (Mnih et al.,2015).Fig.11 shows the single-step rewards obtained by the Naive DQN (Mnih et al.,2013),DQN (Mnih et al.,2015),and DDQN algorithms when performing curve-fitting training on the homogeneous,dual porosity,and radial composite well test model.Fig.11 indicates that DDQN algorithm finally converges to a stable reward value on all well test models and the reward is the highest with the smallest variance.The Na?ve DQN single-step reward has been hovering at 0,indicating that a good strategy cannot be learned.The performance of DQN lies between these two algorithms.The training results show that the DDQN algorithm is more suitable for the problem of automatic curve matching.

    Three alternative supervised ML algorithms are:artificial neural network (ANN),random forest (RF),support vector regression(SVR)with multiple output.In this work,the input of ML algorithm is set to the measured pressure derivative curve with a feature number of 40,and the output is set to the unknown parameters ofthe well test model.The sample size of training set and validation set are 900 and 100,respectively.

    Fig.11.Performance curves of different agents on different well test models.

    Fig.12.Performance curves of different policies.(a) Cumulative reward curve for each episode,and (b) single-step reward curve.

    Fig.13.Histograms of steps required to complete curve matching 100 times by different policies.(a) Boltzmann exploration policy,and (b) ε-greedy policy.

    Fig.14.Performance curves of different reward designs.(a) Cumulative reward curve for each episode,and (b) single-step reward curve.

    Table 4 shows the comparison between the inference results of DDQN and the prediction results of ML algorithm in the validation set.The indicator in Table 4 are averages of all parameters for each well test model.We can observe that,for the homogeneous model,RF has obtained the best parameter inversion results (consideringR2),and DDQN has obtained the suboptimal results.For dual porosity model and radial composite model,DDQN achieved the optimal results,followed by RF,ANN and SVR.When the unknown parameters of the well test model are few (homogeneous model),the ML algorithm generally performs well.However,when the well test model becomes complex,the parameter inversion results of DDQN will be more robust,which can be observed from the fluctuation of Mean of Relative Error in different well test models.

    Table 6 Case 2 parameter inversion result.

    5.Results and discussion

    5.1.Impact of policy

    Fig.15.Histograms of steps required to complete curve matching 100 times by different reward designs.(a) Parameter-based reward design,and (b) curve-based reward design.

    Fig.16.Visualization of the step-by-step parameter adjustment process in Example 1. A:action index.

    In this section,we compare the effects of ε-greedy policy and bBoltzmann exploration policy,two commonly used exploration and balance methods,on the performance of the agent's well test curve matching.Among them,the action probability distribution of the ε-greedy policy is shown in Eq.(13).

    Fig.12 presents the training process under different policies.The results show that the Boltzmann exploration policy will enable the agent to obtain higher cumulative rewards and single-step rewards.In addition,the agent under Boltzmann exploration policy hasfewer steps when completing curve matching,as shown in Fig.13.

    This is because Boltzmann exploration policy is more suitable for a highly certain environment such as well test curve matching tasks.Besides,the use of Boltzmann exploration policy allows the agent to be biased towards exploration in the early stages of training and towards exploitation in the later stages.Therefore,the action certainty of the agent using Boltzmann exploration policy will improve with training,and converge to a better strategy.whereA*is the optimal action,a∈A(s),|A(s)|is the number of actions,ε is the probability of taking a random action.

    Fig.17.Visualization of the step-by-step parameter adjustment process in Example 2. A:action index.

    5.2.Reward design

    In this section,we discuss the impact of two different reward design methods on the agent performance in the task of automatic curve matching.In RL,the design of reward function will significantly affect the training effect of the agent (Ng et al.,1999;Laud,2004).Therefore,it is very important to find a reward function suitable for the current environment for the agent.For curve matching problems,an intuitive reward function design method is based on the error between the target curve and the prediction curve,as in Eq.(14).However,the reward design based on curve error is not a good practice.In fact,when the agent takes the action of parameter adjustment,the reduction of the error between the target curve and the prediction curve sometimes does not mean that the parameter error is reduced.As a result,this ambiguous reward design causes the agent's training to oscillate or diverge.Correspondingly,a more accurate reward design method is to directly use the error between the target curve parameters and the prediction curve parameters.So,we adopt this method to design the reward function,as shown in Eq.(12).

    Fig.14 compares the training process of the two reward design methods.The training process is based on Algorithm 1 and the reward value is still as shown in Eq.(11).It is shown that when the agent uses the curve error-based reward design,the cumulative reward is more,as in Fig.14(a),but the single-step reward is lower,as in Fig.14(b).This indicates that the agent has taken an action that can be rewarded but does not help accomplish the goal.Moreover,Fig.15 compares the histograms of steps required to complete curve matching 100 times by these two methods.The results indicate that setting the reward based on the parameter error will complete the curve matching task faster.

    6.Field application

    To further show the practicality of the proposed automatic matching method,case studies were carried out with 3 typical wells as examples.To make the agent suitable for reservoirs and wellbore under different conditions,the data used is dimensionless.

    6.1.Example 1

    Fig.18.Visualization of the step-by-step parameter adjustment process in Example 3. A:action index.

    Example 1 is an exploration well from the Dagang Oilfield.The permeability is interpreted as 5.4 mD.The well was produced 222 min before the test at a rate of 13.8 m3/d,shut in 621 min during the test,and a good build-up pressure curve was measured.The pressure derivative curve of this well shows the characteristics of a homogeneous reservoir.After dimensionless,the agent trained on the homogeneous model is used to automatically fit the curve.After 26 steps of parameter adjustments,the agent completed the matching process.To understand the curve matching process of the agent,the parameter adjustment action is shown in Fig.16 step by step.Finally,the automatic interpretation result isS=2.64,CD=79.In curve matching process,the agent prefers to adjustCDwith obvious features first,and then adjustS.Table 5 compares the results of automatic matching and computer-assisted manual matching,and the average relative error between them is 4.9%.It is shown that the automatic fitting method has the potential to reach the level of manual interpretation.

    6.2.Example 2

    Example 2 is a well test data from the Sichuan Basin.The well is a high-pressure gas well,and the reservoir has the characteristics of dual porosity.The permeability is interpreted as 0.23 mD.Due to the extremely low permeability of the reservoir,radial flow characteristics still did not appear after 934 h of shut-in testing.Afterdimensionless,the agent trained on dual porosity model is used to automatically fit the curve.The agent completed the curve matching by adjusting the parameters 27 times,as shown in Fig.17.The result of automatic interpretation isS=0.86,CD=2032,ω=0.13,λ=6.4 × 10-7.The results show that in the case of insufficient test time,the agent can also fit the curve well.Table 6 compares the results of automatic matching and computerassisted manual matching,and the average relative error between them is 13.92%.The error in Case 2 is higher than in Case 1,but still acceptable.

    Table 7 Case 3 parameter inversion result.

    6.3.Example 3

    Example 3 is a gas well in the Tarim Basin.The well was tested for 72 h of shut-in pressure build-up,and a pressure build-up curve has been obtained.The pressure derivative curve dropped at the end of the test,showing the characteristics of a composite formation.After dimensionless,the agent trained on radio composite model is used to automatically fit the curve.The agent completed the curve matching by gradually adjusting the parameters,as shown in Fig.18.The result of automatic interpretation isS=8,CD=300,M=0.23,Ri=40 m.The interpretation results show that the skin of the well is large.It turns out that there is pollution around the well and there is serious sanding problem.4.3.Table 7 compares the results of automatic matching and computerassisted manual matching,and the average relative error between them is 4.30%.The result of automatic interpretation in this case is very close to the result of manual interpretation,which proves the practical value of the proposed method.

    7.Conclusions

    In this work,we successfully applied DRL to the task of automatically interpreting well test data.In the automatic interpretation process,the agent interacts in an environment encapsulated based on the well test model to learn how to adjust the parameters to match the well test curve.By testing the performance of the DRL algorithms on different well test models,the following key conclusions are drawn:

    (1) By making the agent adjust the curve parameters asynchronously,the dimensioning disaster was alleviated,and DDQN algorithm was successfully used in the automatic curve matching task on different well test models.Using DDQN algorithm to perform 100 curve matching tests on three well test models,the results show that the mean relative error of the parameters is 7.58% for the homogeneous model,10.66% for the radial composite model,and 12.79% for the dual porosity model.

    (2) Comparing the performance of Na?ve DQN,DQN,DDQN algorithms on the homogeneous model,radial composite model,and dual porosity model,it is shown that the agent based on the DDQN algorithm obtains the highest cumulative reward on these three well test models.In addition,compared with the supervised ML algorithm,DDQN has the least fluctuation of evaluation index on different well test models,which reflects its robustness in curve matching.

    (3) The experimental results show that the use of parameterbased reward design can achieve better training results.In addition,Boltzmann exploration policy is more suitable for agents to balance exploration and exploitation on curve matching tasks.

    (4) In the three field case tests,the agent completed the curve matching within 30 steps.By visualizing the process of the step-by-step parameter tuning,it was verified that the agent learned the correct strategy.

    Acknowledgements

    This work received funding support from National Natural Science Foundation of China (52074322),Beijing Natural Science Foundation (3204052),Science Foundation of China University of Petroleum,Beijing (No.2462018YJRC032),and National Major Project of China (2017ZX05030002-005).The authors sincerely thank the colleagues at State Key Laboratory of Petroleum Resources for their helpful support.

    Appendix 1.Dimensionless definitions

    宅男免费午夜| 国产xxxxx性猛交| 亚洲成色77777| 国产极品天堂在线| 精品午夜福利在线看| 高清不卡的av网站| 热re99久久精品国产66热6| 亚洲三级黄色毛片| 亚洲国产精品一区二区三区在线| 国产男人的电影天堂91| 国语对白做爰xxxⅹ性视频网站| 久久久久久人人人人人| 久久久久国产网址| 国产黄色免费在线视频| 男女高潮啪啪啪动态图| 一级a爱视频在线免费观看| 亚洲第一av免费看| 欧美最新免费一区二区三区| 午夜福利乱码中文字幕| 一级片免费观看大全| 亚洲一级一片aⅴ在线观看| videossex国产| 中文字幕亚洲精品专区| 免费在线观看完整版高清| 日韩av不卡免费在线播放| 亚洲国产欧美在线一区| 老司机亚洲免费影院| 国产亚洲av片在线观看秒播厂| 久久精品人人爽人人爽视色| 观看av在线不卡| 久久久久久久久免费视频了| 欧美 日韩 精品 国产| 亚洲男人天堂网一区| 亚洲精品日韩在线中文字幕| 美女xxoo啪啪120秒动态图| 少妇 在线观看| 在线观看免费视频网站a站| 最近最新中文字幕大全免费视频 | 国产毛片在线视频| 美女大奶头黄色视频| 中国三级夫妇交换| 一区二区三区乱码不卡18| 国产在线一区二区三区精| 九草在线视频观看| 又大又黄又爽视频免费| 国产精品久久久久久av不卡| 久久久国产欧美日韩av| 亚洲国产精品一区二区三区在线| 欧美另类一区| 麻豆精品久久久久久蜜桃| 欧美97在线视频| 中文字幕人妻丝袜制服| xxx大片免费视频| 99久久精品国产国产毛片| 国产亚洲av片在线观看秒播厂| 丝袜美腿诱惑在线| 久久久久国产一级毛片高清牌| 2021少妇久久久久久久久久久| 免费黄频网站在线观看国产| 成年美女黄网站色视频大全免费| 国产男女内射视频| 美女国产高潮福利片在线看| 精品国产一区二区三区四区第35| 视频在线观看一区二区三区| 我的亚洲天堂| 国产精品久久久久久精品古装| 国产毛片在线视频| 日韩伦理黄色片| 日日爽夜夜爽网站| 男女午夜视频在线观看| 国产成人精品久久久久久| 少妇的丰满在线观看| 人人妻人人添人人爽欧美一区卜| 天堂俺去俺来也www色官网| 18禁国产床啪视频网站| 日韩av在线免费看完整版不卡| 2018国产大陆天天弄谢| 韩国av在线不卡| 日韩在线高清观看一区二区三区| 美女主播在线视频| 国产成人a∨麻豆精品| 日韩制服丝袜自拍偷拍| 国产精品一区二区在线观看99| 少妇人妻 视频| 大片免费播放器 马上看| 国产深夜福利视频在线观看| 亚洲成人av在线免费| 色婷婷av一区二区三区视频| 天天躁日日躁夜夜躁夜夜| 新久久久久国产一级毛片| 精品国产露脸久久av麻豆| 日本vs欧美在线观看视频| 一区二区三区乱码不卡18| 亚洲欧美色中文字幕在线| 成年美女黄网站色视频大全免费| 在线观看国产h片| 国产精品国产三级国产专区5o| 久久av网站| 国产 一区精品| 国产成人免费无遮挡视频| 在线亚洲精品国产二区图片欧美| 欧美激情高清一区二区三区 | 日韩欧美精品免费久久| av在线老鸭窝| 国产黄色免费在线视频| 黄色毛片三级朝国网站| 日本91视频免费播放| 制服丝袜香蕉在线| 午夜久久久在线观看| 狂野欧美激情性bbbbbb| 亚洲精品美女久久av网站| 成人国产麻豆网| 美国免费a级毛片| 免费少妇av软件| 亚洲精品自拍成人| 国产免费视频播放在线视频| 黑人巨大精品欧美一区二区蜜桃| 精品国产一区二区三区久久久樱花| 美女国产视频在线观看| 曰老女人黄片| 国产精品二区激情视频| 性少妇av在线| 在线免费观看不下载黄p国产| 亚洲精华国产精华液的使用体验| 亚洲熟女精品中文字幕| 一级毛片我不卡| 日本欧美视频一区| av在线观看视频网站免费| 久久99热这里只频精品6学生| 精品99又大又爽又粗少妇毛片| 色94色欧美一区二区| 一级毛片 在线播放| 不卡av一区二区三区| 一区福利在线观看| 久久精品亚洲av国产电影网| 午夜91福利影院| 午夜影院在线不卡| 91午夜精品亚洲一区二区三区| 黑人欧美特级aaaaaa片| 精品人妻一区二区三区麻豆| 丝袜人妻中文字幕| 久久精品国产综合久久久| 一级爰片在线观看| 国产精品偷伦视频观看了| 久久精品久久精品一区二区三区| 老女人水多毛片| 美女大奶头黄色视频| 国产成人av激情在线播放| 久久久久久人人人人人| 亚洲精品aⅴ在线观看| 久久久久人妻精品一区果冻| 乱人伦中国视频| 大话2 男鬼变身卡| 国产一区二区 视频在线| 亚洲美女搞黄在线观看| 国产男女内射视频| 中文字幕制服av| 国产有黄有色有爽视频| 黄片播放在线免费| 成年av动漫网址| av.在线天堂| 欧美另类一区| 丁香六月天网| 两性夫妻黄色片| 9191精品国产免费久久| 亚洲人成网站在线观看播放| 性少妇av在线| 亚洲精品中文字幕在线视频| 日本-黄色视频高清免费观看| 欧美中文综合在线视频| 狠狠精品人妻久久久久久综合| 久久精品人人爽人人爽视色| www.精华液| 欧美日韩精品网址| 1024香蕉在线观看| 观看av在线不卡| 国产免费一区二区三区四区乱码| h视频一区二区三区| freevideosex欧美| 国产成人免费无遮挡视频| 新久久久久国产一级毛片| 午夜福利在线观看免费完整高清在| av网站在线播放免费| 亚洲五月色婷婷综合| 国产精品二区激情视频| 中国三级夫妇交换| 国产日韩欧美视频二区| 岛国毛片在线播放| 一本—道久久a久久精品蜜桃钙片| 精品国产国语对白av| 欧美在线黄色| 超色免费av| 91在线精品国自产拍蜜月| 成人二区视频| 国产有黄有色有爽视频| av一本久久久久| 亚洲欧美日韩另类电影网站| 国产在线免费精品| 亚洲综合色网址| 99热全是精品| 大陆偷拍与自拍| 欧美精品国产亚洲| 狠狠精品人妻久久久久久综合| 免费观看av网站的网址| 久久综合国产亚洲精品| 黑人巨大精品欧美一区二区蜜桃| 国产女主播在线喷水免费视频网站| av天堂久久9| 18禁动态无遮挡网站| 亚洲欧美一区二区三区黑人 | kizo精华| 日日撸夜夜添| 韩国高清视频一区二区三区| 大香蕉久久网| 精品一区在线观看国产| 日日啪夜夜爽| 视频在线观看一区二区三区| 日韩欧美精品免费久久| 亚洲男人天堂网一区| 极品人妻少妇av视频| 国产成人aa在线观看| 美女高潮到喷水免费观看| 国产亚洲欧美精品永久| 999久久久国产精品视频| 中文天堂在线官网| 国产1区2区3区精品| 日韩欧美一区视频在线观看| 国产精品秋霞免费鲁丝片| 少妇 在线观看| 大陆偷拍与自拍| 亚洲情色 制服丝袜| 精品亚洲成国产av| 久久久久精品久久久久真实原创| 日韩中文字幕欧美一区二区 | 久久久久国产网址| 青春草亚洲视频在线观看| 午夜福利视频在线观看免费| 免费在线观看黄色视频的| 下体分泌物呈黄色| 精品国产一区二区三区久久久樱花| 亚洲av电影在线观看一区二区三区| 日本91视频免费播放| 一区在线观看完整版| 免费日韩欧美在线观看| 性高湖久久久久久久久免费观看| 国产一区二区 视频在线| 亚洲第一av免费看| 桃花免费在线播放| 日韩制服丝袜自拍偷拍| av免费观看日本| 男人爽女人下面视频在线观看| www.av在线官网国产| 国产精品欧美亚洲77777| 丝瓜视频免费看黄片| 视频在线观看一区二区三区| 9热在线视频观看99| videosex国产| 国精品久久久久久国模美| 午夜福利一区二区在线看| 极品人妻少妇av视频| 妹子高潮喷水视频| 伊人亚洲综合成人网| 99久久综合免费| 欧美老熟妇乱子伦牲交| 午夜久久久在线观看| 新久久久久国产一级毛片| 日本色播在线视频| 亚洲欧美精品综合一区二区三区 | 国产av国产精品国产| 午夜福利一区二区在线看| 中国三级夫妇交换| 天天躁夜夜躁狠狠躁躁| 亚洲熟女精品中文字幕| 丰满饥渴人妻一区二区三| 久久精品国产亚洲av高清一级| 91精品伊人久久大香线蕉| 国产成人一区二区在线| 91久久精品国产一区二区三区| 在线观看人妻少妇| 少妇人妻久久综合中文| av线在线观看网站| 99热国产这里只有精品6| 在线免费观看不下载黄p国产| 亚洲欧美一区二区三区黑人 | 美女大奶头黄色视频| 一本久久精品| 国产高清不卡午夜福利| 色播在线永久视频| av有码第一页| 啦啦啦中文免费视频观看日本| kizo精华| 搡老乐熟女国产| 日本欧美视频一区| 亚洲国产欧美在线一区| 国产精品久久久av美女十八| 国产97色在线日韩免费| 国产野战对白在线观看| 极品人妻少妇av视频| 18禁动态无遮挡网站| 久久精品亚洲av国产电影网| 韩国高清视频一区二区三区| 一区二区三区精品91| 9热在线视频观看99| 中文字幕另类日韩欧美亚洲嫩草| 好男人视频免费观看在线| 十八禁网站网址无遮挡| xxxhd国产人妻xxx| 亚洲国产日韩一区二区| av在线播放精品| 亚洲精品一二三| 男女边吃奶边做爰视频| 久热这里只有精品99| 美女中出高潮动态图| 街头女战士在线观看网站| 亚洲欧美精品综合一区二区三区 | 亚洲成人av在线免费| 国产av国产精品国产| 亚洲男人天堂网一区| 伦理电影免费视频| 蜜桃在线观看..| 一边亲一边摸免费视频| 亚洲精品久久成人aⅴ小说| 久久久久国产一级毛片高清牌| 新久久久久国产一级毛片| 搡女人真爽免费视频火全软件| 丝袜在线中文字幕| 黄片无遮挡物在线观看| 蜜桃国产av成人99| 欧美日韩精品成人综合77777| 另类精品久久| 国产熟女午夜一区二区三区| 18禁动态无遮挡网站| 国产精品无大码| 亚洲精品成人av观看孕妇| 午夜福利在线观看免费完整高清在| 国语对白做爰xxxⅹ性视频网站| 人人妻人人澡人人爽人人夜夜| 精品少妇久久久久久888优播| 欧美黄色片欧美黄色片| 国产一区二区 视频在线| 午夜免费男女啪啪视频观看| 精品一区在线观看国产| a级片在线免费高清观看视频| √禁漫天堂资源中文www| 亚洲第一av免费看| 国产激情久久老熟女| 久久久久久久亚洲中文字幕| 又粗又硬又长又爽又黄的视频| 在线观看人妻少妇| 精品午夜福利在线看| 国产一区有黄有色的免费视频| 中国三级夫妇交换| 国产免费又黄又爽又色| 国产成人精品婷婷| 80岁老熟妇乱子伦牲交| 国产日韩一区二区三区精品不卡| 99re6热这里在线精品视频| 青春草亚洲视频在线观看| 国产精品不卡视频一区二区| 国产成人av激情在线播放| 尾随美女入室| 久久青草综合色| 国产一区二区三区综合在线观看| 久久精品久久精品一区二区三区| 五月开心婷婷网| 久久国产亚洲av麻豆专区| 免费大片黄手机在线观看| 校园人妻丝袜中文字幕| 免费大片黄手机在线观看| 国精品久久久久久国模美| 国产欧美亚洲国产| 在线观看免费高清a一片| 欧美日本中文国产一区发布| 久久久久久久精品精品| 国产欧美亚洲国产| 9191精品国产免费久久| 狠狠婷婷综合久久久久久88av| 午夜老司机福利剧场| 精品国产露脸久久av麻豆| 亚洲精品国产av成人精品| 中文字幕人妻丝袜制服| 下体分泌物呈黄色| 日产精品乱码卡一卡2卡三| 秋霞在线观看毛片| 亚洲av国产av综合av卡| 国产精品一二三区在线看| 两性夫妻黄色片| 丝袜在线中文字幕| 日韩制服骚丝袜av| 一个人免费看片子| 国产亚洲欧美精品永久| 蜜桃在线观看..| 黑人欧美特级aaaaaa片| 国产一区二区激情短视频 | 交换朋友夫妻互换小说| 秋霞在线观看毛片| 欧美人与性动交α欧美精品济南到 | 啦啦啦在线免费观看视频4| 97在线人人人人妻| 天堂中文最新版在线下载| 久久精品熟女亚洲av麻豆精品| a级毛片在线看网站| 日韩不卡一区二区三区视频在线| 久久久久精品性色| 国产成人精品在线电影| 免费久久久久久久精品成人欧美视频| 久久久久久免费高清国产稀缺| 黄频高清免费视频| 久久精品国产综合久久久| 久久精品熟女亚洲av麻豆精品| 欧美精品一区二区大全| 精品国产国语对白av| 美女高潮到喷水免费观看| 国产综合精华液| 久久狼人影院| 亚洲欧美成人综合另类久久久| 国产精品久久久久久精品古装| 免费高清在线观看视频在线观看| 狂野欧美激情性bbbbbb| 肉色欧美久久久久久久蜜桃| 中文字幕人妻熟女乱码| 老司机亚洲免费影院| www.av在线官网国产| 中文字幕最新亚洲高清| 国产色婷婷99| 久久久精品国产亚洲av高清涩受| 欧美日韩av久久| 看十八女毛片水多多多| 久久热在线av| 精品国产一区二区三区四区第35| 日韩制服骚丝袜av| 如何舔出高潮| 最新的欧美精品一区二区| 日本欧美国产在线视频| 黑丝袜美女国产一区| 亚洲色图综合在线观看| 午夜免费男女啪啪视频观看| 亚洲综合色网址| 丁香六月天网| 国产极品粉嫩免费观看在线| 啦啦啦啦在线视频资源| 2022亚洲国产成人精品| 男人爽女人下面视频在线观看| 色网站视频免费| 男男h啪啪无遮挡| kizo精华| av国产精品久久久久影院| 女人久久www免费人成看片| 亚洲欧美一区二区三区久久| 成人毛片a级毛片在线播放| 青春草国产在线视频| 哪个播放器可以免费观看大片| 精品99又大又爽又粗少妇毛片| 中国三级夫妇交换| 免费看不卡的av| 精品少妇久久久久久888优播| 夫妻性生交免费视频一级片| 久久久久网色| 国产 一区精品| 伊人久久国产一区二区| 美国免费a级毛片| 欧美亚洲 丝袜 人妻 在线| 夜夜骑夜夜射夜夜干| 中文字幕人妻丝袜制服| 人妻人人澡人人爽人人| 色网站视频免费| 亚洲欧美一区二区三区久久| 波野结衣二区三区在线| 丝袜美足系列| 成人亚洲精品一区在线观看| 日本爱情动作片www.在线观看| 精品酒店卫生间| 纵有疾风起免费观看全集完整版| 国产精品熟女久久久久浪| 男女免费视频国产| 国产 精品1| 国产成人精品久久二区二区91 | 大香蕉久久成人网| 丰满饥渴人妻一区二区三| 欧美日韩成人在线一区二区| 国产精品国产三级国产专区5o| 亚洲精品日韩在线中文字幕| 侵犯人妻中文字幕一二三四区| 亚洲精品自拍成人| 亚洲三区欧美一区| 婷婷色麻豆天堂久久| 一二三四在线观看免费中文在| 天堂中文最新版在线下载| 婷婷色综合www| 成年美女黄网站色视频大全免费| 女性被躁到高潮视频| 在线看a的网站| 午夜激情av网站| 精品人妻偷拍中文字幕| av女优亚洲男人天堂| 中文天堂在线官网| 日韩 亚洲 欧美在线| 日韩av在线免费看完整版不卡| 男人爽女人下面视频在线观看| 丝袜美腿诱惑在线| videossex国产| 欧美最新免费一区二区三区| 嫩草影院入口| videosex国产| 免费观看性生交大片5| 不卡视频在线观看欧美| 黄色怎么调成土黄色| 亚洲,欧美,日韩| 美女xxoo啪啪120秒动态图| 亚洲视频免费观看视频| 国产精品 欧美亚洲| 一区二区日韩欧美中文字幕| 在线观看三级黄色| 国产免费又黄又爽又色| av又黄又爽大尺度在线免费看| 国产色婷婷99| 男女边吃奶边做爰视频| av视频免费观看在线观看| 国产亚洲午夜精品一区二区久久| 久久精品熟女亚洲av麻豆精品| 亚洲精品国产一区二区精华液| 青春草视频在线免费观看| 777米奇影视久久| h视频一区二区三区| 多毛熟女@视频| 国产精品国产av在线观看| 精品一区在线观看国产| 亚洲人成网站在线观看播放| www.熟女人妻精品国产| 久久精品熟女亚洲av麻豆精品| 大片电影免费在线观看免费| 国产又爽黄色视频| 国产精品人妻久久久影院| 天天躁狠狠躁夜夜躁狠狠躁| 七月丁香在线播放| 欧美在线黄色| 久久久久国产一级毛片高清牌| 十八禁高潮呻吟视频| 女人精品久久久久毛片| 国产亚洲精品第一综合不卡| 久久久精品国产亚洲av高清涩受| av线在线观看网站| 成人国产av品久久久| 五月开心婷婷网| 91精品国产国语对白视频| 免费高清在线观看日韩| 精品久久久精品久久久| 国产熟女欧美一区二区| 夜夜骑夜夜射夜夜干| 国产成人精品久久久久久| 亚洲综合精品二区| 亚洲欧美精品自产自拍| 我要看黄色一级片免费的| 欧美xxⅹ黑人| av女优亚洲男人天堂| 亚洲精品自拍成人| 三级国产精品片| 欧美亚洲 丝袜 人妻 在线| 亚洲av综合色区一区| 如日韩欧美国产精品一区二区三区| 亚洲精品在线美女| 国产精品秋霞免费鲁丝片| 一级爰片在线观看| 欧美精品av麻豆av| 国产精品.久久久| 亚洲一区中文字幕在线| 只有这里有精品99| 18禁观看日本| 啦啦啦在线观看免费高清www| 亚洲精品一二三| 久久国产亚洲av麻豆专区| 中文字幕人妻丝袜制服| 欧美精品亚洲一区二区| 中文欧美无线码| 欧美成人午夜免费资源| 精品午夜福利在线看| 国产不卡av网站在线观看| 久久精品久久精品一区二区三区| 欧美变态另类bdsm刘玥| 国产av国产精品国产| 久久鲁丝午夜福利片| 免费在线观看视频国产中文字幕亚洲 | 国产日韩欧美在线精品| 母亲3免费完整高清在线观看 | 欧美日韩亚洲高清精品| 亚洲av国产av综合av卡| 999精品在线视频| 免费播放大片免费观看视频在线观看| 欧美精品高潮呻吟av久久| 亚洲国产精品一区三区| 波野结衣二区三区在线| 下体分泌物呈黄色| 波多野结衣一区麻豆| 亚洲少妇的诱惑av| 中文精品一卡2卡3卡4更新| 国产不卡av网站在线观看| 中文字幕av电影在线播放| 亚洲av电影在线进入| 老司机影院成人| 婷婷色麻豆天堂久久| 国产欧美日韩一区二区三区在线| 99久久中文字幕三级久久日本| 美女午夜性视频免费| 国产一区二区 视频在线| 欧美激情极品国产一区二区三区| 夜夜骑夜夜射夜夜干| 亚洲av中文av极速乱| 狠狠精品人妻久久久久久综合| 可以免费在线观看a视频的电影网站 | 精品国产超薄肉色丝袜足j| 久久这里只有精品19| 午夜福利乱码中文字幕| 久久青草综合色| 丝袜美腿诱惑在线| 韩国av在线不卡|