• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Power law decay of stored pattern stability in sparse Hopfield neural networks

    2021-04-26 03:19:54FeiFangZhouYangandShengJunWang
    Communications in Theoretical Physics 2021年2期

    Fei Fang,Zhou Yang and Sheng-Jun Wang

    School of Physics and Information Technology,Shaanxi Normal University,Xi’an 710119,China

    Abstract Hopfield neural networks on scale-free networks display the power law relation between the stability of patterns and the number of patterns.The stability is measured by the overlap between the output state and the stored pattern which is presented to a neural network.In simulations the overlap declines to a constant by a power law decay.Here we provide the explanation for the power law behavior through the signal-to-noise ratio analysis.We show that on sparse networks storing a plenty of patterns the stability of stored patterns can be approached by a power law function with the exponent -0.5.There is a difference between analytic and simulation results that the analytic results of overlap decay to 0.The difference exists because the signal and noise term of nodes diverge from the mean-field approach in the sparse finite size networks.

    Keywords: Hopfield neural network,attractor neural networks,associative memory

    1.Introducti on

    The Hopfield attractor neural network [1,2] is a well-known Ising-like model of associative memory [3-15].Trained by the Hebbian rule,a number of patterns can be stored and become attractors of the neural network.As a pattern is presented to the trained attractor neural network,the network state converges to the nearest stored pattern which most closely resembles the input pattern.Therefore Hopfield neural networks can function as a form of associative memory [1].When only a few of patterns are stored,these patterns are stable attractors in Hopfield neural networks.With the number of patterns increases,the quality of recognition decreases[16].As a stored pattern is presented to the network,retrieval errors occur because of the crosstalk between stored patterns,that is,the stability of stored patterns decreases [3].

    The performance of the Hopfield model on scale-free(SF) networks has been studied.In SF networks,the connection degrees,i.e.the number of connections of each node,are heterogeneous.The distribution of degree is power law.The heterogeneous degree affects the behavior of the Hopfield model.For example,retrieval errors are not distributed uniformly among nodes [3].Especially,the power law relation between the performance of the network and the number of patterns has been found on SF networks [16].

    The power law relation has an important influence on the network’s property,that is,the stability of patterns declines slowly as the number of stored patterns increases.Therefore,this is a beneficial feature for the dynamics of associative memory.In fact,power law behaviors pervade neural activities [17] and experiments which has shown that memory exhibits power law forgetting [18-20].Using a mean-field neural network model with slow plasticity,an elegant analysis has been proposed that neural networks can exhibit power law forgetting,i.e.universal power law relaxation of the mean synaptic strength along critical manifold or at the tricritical point [18].However,it is still an open question that how the power law behavior emerges in Hopfield neural networks.

    In this work we present the mechanism of the power law relation through analytical analysis.We simulate the Hopfield network on SF networks and homogeneous Erd?s-Rényi(ER) random networks,and obtain that the overlap obeys power law decay to a constant which depends on network size and connection density.Using signal-to-noise analysis,we obtain that as the network is sparse,the relation between the overlap and the number of stored patterns can be described by a power law function,while the overlap decays to 0.

    2.Model

    In the Hopfield neural network model [1,2],node states si=±1,i ∈ {1,2,…,N} are binary.The dynamical equation of the system is [1,3]

    where Jijis the entry of the coupling matrix.Using the Hebbian rule:

    When the number n of stored patterns is small,these patterns are stable attractors of the Hopfield neural network.When a stored pattern or a blurred pattern is presented to the neural network,the output of the Hopfield neural network is the same as the attractor,xi=When the number n of stored patterns is large,stored patterns cannot be completely recovered [1].Some nodes in the network flipped their states siand the network state is different from the attractor.However the network state can be close to one of stored patterns.So the network still recognize stored patterns [16].To quantify the computational performance of Hopfield attractor neural networks,we employ the stability of stored patterns[3].A stored pattern is presented to a neural network,then the overlap between the output state and the initial state is computed as the measurement of the stability.The overlap is defined like this:

    where xi=±1 denotes the output of the ith node,i.e.the node state of the network in which the dynamic has converged into the stable state.

    We use ER random network model and SF network model in simulations.ER random networks are generated as follows:starting with N disconnected nodes,we connect each of possible edges with a probability 0 < p < 1.The degree distribution is a Poisson distribution [11]

    where 〈k〉 denotes the average degree in the network.The average degree of the model is equal to pN.

    SF networks are generated as follows: starting with m0fully connected nodes,at each time step,a new node with m ≤ m0edges is added by linking the new node j to m different nodes already present in the system.The probability that the new node j is linked to an existing node i is linearly proportional to the actual degree of i:

    The average degree of SF model approaches to 〈k〉=2m in large networks.The degree distribution of SF model is like:

    We take m=m0in our work and the network is connected for any value of average degree.

    3.Results

    On SF networks,the power law relation between the overlap and the number of stored patterns was obtained using computer simulations in[16].So we first analyze the stability of patterns on SF networks.Then we also analyze the behavior of the Hopfield neural network model on ER random networks.

    3.1.SF networks

    The relation between the overlap and the number of stored patterns has been derived in [11,21].First,we review the formula of overlap on SF networks.

    If Hopfield neural networks consist of N nodes and memorized n patterns,the condition for a patternto be stable can be expressed by N inequalities,one for each node i:

    Substitute the Hebbian rule (i.e.equation (2)) into the inequality (7) gives

    Break the left hand side of equation into two parts

    The first term on the left is called the signal term,because it is equal to the connection degree kiof the node i which is positive and tends to reinforce the state stable.This term is denoted by Tsignal.The double sum of the second part includes(n - 1)kiterms equal to +1 or -1.Because patterns are uncorrelated,this sum is 0 on the average.Its standard deviation is the square root of the number of terms(n - 1)ki.This term is called noise term which tends to make the state unstable,denoted by Tnoise.When patterns are uncorrelated,the value of noise term follows a Gaussian distribution

    where the mean value μ=0,the standard deviationThe noise term Tnoise< -k,the sum of the signal term and the noise term less than zero,then the state of node is unstable.So the probability that nodes of degree k are unstable is:

    For simplicity,we transform the unstable probability into the formula represented by error function:

    where erf(x) =For the given number of patterns n,U(k) is a decreasing function.

    The stability of stored patterns in networks can be predicted by the probability U(k).For the network which has converged into the stable state,the overlap order parameter can be rewritten asNflipis the number of flipped nodes[11].For SF networks,assuming a sharp cutoff of the distribution at low k,the distribution is normalized as [12]

    The number of flipped nodes is

    Then the overlap order parameter is:

    Based on the result of overlap,we take further analytic treatment to interpret the power law relation between the overlap and the number of patterns.The power series expansion of error function has the form

    When x is fairly small,only a few terms of this series are necessary for obtaining a good approximation.

    For very a small average degree k and a large number of patterns n,is small.We use the expansion with only the first term in computing the overlap on SF networks

    Then

    We use y1to represent the coefficient

    For SF networks,the coefficient y1can be numerically computed,when the parameters γ andkmin=mare given.And the overlap is

    When n ?1,an approximate formula reads as:

    Therefore,we obtain that the relation between the overlap and the number of stored patterns follows a power law function when the network is sparse and the number of stored patterns is large.Theoretical analysis shows that the exponent is -1/2.

    It was obtained in simulations in the[16]that the overlap φ varies roughly as (φ - φ0) ∝ n-τ.Following the [16],we measure the overlap difference (φ - φ0) for SF networks in simulations.The network is built using the SF network model[22,23].The size of networks is N=10 000.One of the stored patterns is used as the network’s initial state.Here we recover the relation reported in [16].The simulation results are shown in figures 1,2.With the average degree increases,the value of φ0decrease.When the average degree〈k〉=4,10,20 and 50,the final overlap φ0respectively is 0.24,0.23,0.21 and 0.19.In figure 1,for networks with〈k〉=4,the relation between the overlap difference and the number of patterns is power law.By fitting the simulation results into a power law function,we obtain the exponent of the relation is -0.541 ± 0.001,which is close to the theoretical results of the exponent -1/2.

    The limit of the number of patterns that can be stored with small error is called the storage capacity [9].We compute the storage capacity using the threshold of overlap 0.95.Below the capacity,the relation between overlap and the number of patterns cannot be a power decay.For network with N=10 000,the capacity of SF networks is n=2 with〈k〉=4.Therefore the whole curve shown in the figure 1 can exhibit a power law decay.The value of φ0is 0.24.

    When the average degree of the network is large,the capacity becomes large,the relation is not power law for small values of the number of patterns.However,our analysis shows that as the the ratio k/2(n - 1) becomes small,the relation turns to a power law.The results of networks with the average degree 〈k〉=50 are shown in figure 2.The capacity of SF networks is n=10 with 〈k〉=50.As the number of patterns is larger than 10,the relation changes into a power law.For networks with 〈k〉=50,the final overlap φ0decreased to 0.19,as shown in the inset of figure 2.The exponent is -0.689 ± 0.003.

    The change of the relation can be explained by the signalto-noise ratio(SNR)analysis.The power law relation is based on the expansion of the error function under the condition that the connection degree is relatively small compared with the number of patterns k/n ?1.As the network become dense,the condition k/n ?1 is not satisfied.The relation φ - φ0versus n cannot be expressed as a power law function,as seen in figure 2.As the number of stored patterns increases,the neural network satisfies the condition,then the relation turn into a power law function.

    Figure 1.The overlap difference φ - φ0 versus the number of patterns n on scale-free networks on a log-log scale.The blue dashed line is the analytic results.Squares represent the simulation results.The red solid line is the fitted line.Data points from n=2 to 2500 are used in fitting.The parameters are N=10 000,average degree〈k〉=4.Inset:the overlap φ versus the number of patterns n on scale-free networks on a log-log scale.The final overlap φ0 is 0.24.Each data point is the average over 1000 realizations.

    Figure 2.The overlap difference φ - φ0 versus the number of patterns n on scale-free networks on a log-log scale.The blue dashed line is the analytic results.Squares represent the simulation results.The red solid line is the fitted line.Data points from n=40 to 2500 are used in fitting.The parameters are N=10 000,average degree〈k〉=50.Inset:the overlap φ versus the number of patterns n on scale-free networks on a linear coordination.The final overlap φ0 is 0.19.Each data point is the average over 1000 realizations.

    There is a difference between the above simulation and the analytic results,as shown in the inset of figure 1.In analytic results,the overlap decays to 0,i.e.φ ∝ n-0.5.The φ0cannot be obtained in the analysis.The difference between simulation and analytic results may due to two facts.Firstly,in the network with finite size,the distribution of the noise term of nodes deviates from the exact Gaussian distribution.We calculate the distribution of the noise term of nodes in a realization of the SF network,as shown in figure 3.However,the distribution of the noise term of nodes is not the exact Gaussian distribution.In the region Tnoise< -〈k〉 the probability obtained by simulation is lower than the analytic results.The extreme values of the noise term appear rarely.We compute the unstable probability of nodes using both simulations and analysis and shown them in figure 4.In a network with finite size,the difference from the exact Gaussian distribution makes the unstable probability smaller.

    Figure 3.The distribution of the noise term Tnoise in scale-free networks.The dashed line is the exact Gaussian distribution in the analytic results.The curve is the numerical simulation results.The parameters are N=10 000,pattern number n=100,and average degree 〈k〉=50.

    Secondly,the simulation is performed in the way that nodes are updated in a serial order.As one node is flipped by updating,the local field of its neighbor is changed.We use the initial network state to calculate the distribution of the noise term Tnoiseof nodes in SF networks.For a given network realization,nodes i=5001,...,10 000 are used in computing the distribution.The distribution is shown by the dashed line in figure 5.The center of the curve is at Tnoise=0.In simulation,we update a part of nodes which are i=1,…,5000.Then we use the nodes i=5001,...,10 000 to compute the distribution again.The results is shown by the solid line in figure 5.The distribution of the noise term Tnoisehas shifted to the right.The noise term of un-updated nodes becomes larger,so the unstable probability becomes smaller.

    Figure 4.The unstable probability function U(k)versus the degree k in scale-free networks.The red dashed line is the analytic results.The black solid line is the numerical simulation results.The parameters are N=10 000,pattern number n=100,average degree 〈k〉=4.Each data point is 10 000 realizations.

    Figure 5.The distribution of the noise term Tnoise of nodes in a realization of the scale-free network.The dashed line is the results obtained at the initial state.The solid line is the results obtained when 5000 nodes are updated.The parameters are N=10 000,pattern number n=10 000,and average degree 〈k〉=50.The nodes i=5001,...,10 000 are used in computing the two results.

    We provide an understanding of the change of noise.For node A which flips,the noise is negative,like:

    Figure 6.The fraction R of flipped nodes.The fraction R is computed in a moving window of 500 nodes.The dashed line is the results obtained at the initial state.The solid line is the results obtained by comparing the signal and noisy term for each node in the simulation of updating process.The parameters are N=10 000,pattern number n=10 000,and average degree 〈k〉=50.

    where kAis the connection degree of node A.On average,for j satisfying aAj=1.For the node B which is a neighbor of the node A,the noise term TBat the initial state is:

    The second term is a part of TA,and its mean value is negative.When node A filps,ΔTB=on average.So the noise term of nodes gets bigger.When average degree 〈k〉 is larger,the contribution of the term becomes smaller.

    Moreover,we compare the fraction R of flipped nodes predicted by the analysis and the simulation.The fraction R of flipped nodes is computed using a moving window of 500 nodes.Using noise obtained at the initial state of a given realization,the fraction R is close to 0.47,which is shown by the dashed line in figure 6.In the simulation of the same realization,we update node state in the numerical order,i.e.from node i=1 to node i=N.As one node is flipped by updating,the local field of its neighbor is changed.When node i is updated,we compute the fraction R of flipped nodes in nodes from node i - 499 to node i.The fraction R of flipped nodes in updated nodes decreases in the updating process as shown by the solid line in figure 6.Because the local field is changed in simulation,the noise term of unupdated nodes becomes larger,so the unstable probability of un-updated nodes becomes smaller,the fraction R of flipped nodes decreases.Therefore,the simulation results deviate from the theoretical results.

    The difference between the local field and the mean-field also induces that simulation results diverge from the theoretical results.Both of these two effects can be weakened by enlarging the network size and the network connection density.

    Figure 7.The final overlap φ0 in scale-free networks.The density of connections 〈k〉/N=0.01.Parameters are the average degree〈k〉=50,100,200,300,400 and 800.The size of networks respectively is N=5000,10 000,20 000,30 000,40 000 and 80 000.

    We also simulated different sizes of SF networks with the average degree 〈k〉=50.When the network sizes of N=10 000,20 000 and 40 000 nodes,the final overlap φ0respectively is 0.19,0.18 and 0.18.With the network size N gets larger,the final ovarlap φ0decrease slightly.Figure 8 shows results in different sizes of networks.

    Consider the effect of the finite size of the network and the density of connections on the final overlap separately,we keep the density of connections 〈k〉/N is the constant,and checked whether φ0approaches to zero when N gets large.We let 〈k〉/N=0.01,when the network sizes of N=5000,10 000,20 000,30 000,40 000 and 80 000 nodes,the final overlap φ0respectively is 0.202,0.182,0.167,0.161,0.155 and 0.146.When the density of connections is the constant,the final overlap φ0decreases with network size,as shown in figure 7.

    We fitted the simulation results of the relation between φ - φ0and n shown in figure 8.When the network sizes of N=10 000,20 000 and 40 000 nodes,the slope respectively is -0.689 ± 0.003,-0.679 ± 0.005 and -0.659 ± 0.001.As the network sizes increases,the slope close to -0.5.

    3.2.ER random networks

    For ER random networks [22,23],the degree of nodes follows a Poisson distribution.The number of flipped node is

    Figure 8.The overlap difference φ - φ0 versus the number of patterns n on scale-free networks on a log-log scale with different network sizes.Parameters are the average degree〈k〉=50.The size of networks respectively is N=10 000(squares),20 000(circles)and 40 000(upper triangles).Inset: the overlap φ versus the number of patterns n on scale-free networks on a linear coordination.Each data point is averaged over 1000 realizations.

    The overlap between the output state and the attractor is

    That is

    In ER random networks,the coefficientcan be numerically computed when the parameter 〈k〉 is given.The overlap is

    When n ?1,an approximate formula reads as

    Under the condition that the value ofis small,the relation between the stability and the number of stored patterns on ER random networks is the same as that on SF networks.

    Figure 9.The overlap difference φ - φ0 versus the number of patterns n on ER random networks on a log-log scale.The blue dashed line is the analytic results.Squares represent the simulation results.The red solid line is the fitted line.Data points from n=2 to 6500 are used in fitting.The parameters are N=10 000,average degree〈k〉=4.Inset:the overlap φ versus the number of patterns n on random networks on a linear coordination.The final overlap φ0 is 0.27.Each data point is the average over 1000 realizations.

    We also compute the relation between the stability and the number of stored patterns on ER random networks using computer simulations.The size of networks is N=10 000.One of the stored patterns is used as the networks initial state.The simulation results are shown in figures 9,10.The overlap φ varies roughly as (φ - φ0) ∝ n-τ.As the same as SF network,there is φ0in the relation.With the average degree increases,the final overlap φ0decreases.When the average degree 〈k〉=4,10,20 and 50,the final overlap φ0respectively is 0.27,0.24,0.22 and 0.19.

    Our simulation results show that the power law behavior does not only appears on the SF networks but also on the homogeneous ER random networks.For network with N=10 000,the capacity of random networks is n=2 with〈k〉=4.In figure 9,the power law behavior agrees with the theoretical prediction in the whole parameter region of n.The final overlap φ0is 0.27,as shown in the inset of figure 9.The slope is τ=-0.527 ± 0.004 and is close to the theoretical result.

    As the average degree increases,the capacity increases.The capacity of random networks is n=13 with〈k〉=50.The relation also changes from φ=1 to the power law relation when n is larger than the capacity,as seen in figure 10.The final overlap φ0decreased to 0.19.The slope of the right hand side of the curve is -0.687 ± 0.002.According to SNR analysis,as the network become dense,the condition k/n ?1 is not satisfied.Then the relation between φ - φ0and n cannot be expressed as a power law function.As the number of stored patterns increases,the neural network satisfies the condition.Then the relation turn into the power law function.But the exponent is smaller than that of sparser networks.

    This discrepancy is related to the finite size (i.e.N=10 000)of the simulated network.We simulate larger ER random networks with the average degree〈k〉=50.The inset of figure 11 shows results without subtracting φ0in different sizes of networks.When the network sizes of N=10 000,20 000,40 000,80 000 and 160 000 nodes,the final overlap φ0respectively is 0.19,0.19,0.18,0.18 and 0.18.Figure 11 shows the results of φ - φ0in different sizes of networks.

    Figure 10.The overlap difference φ - φ0 versus the number of patterns n on ER random networks on a log-log scale.The blue dashed line is the analytic results.Squares represent the simulation results.The red solid line is the fitted line.Data points from n=40 to 6500 are used in fitting.The parameters are N=10 000,average degree〈k〉=50.Inset:the overlap φ versus the number of patterns n on random networks on a linear coordination.The final overlap φ0 is 0.19.Each data point is the average over 1000 realizations.

    Figure 11.The overlap difference φ - φ0 versus the number of patterns n on ER random networks on a log-log scale.Parameters are the average degree〈k〉=50.The size of networks respectively is N=10 000(squares),20 000(circles),40 000(upper triangles),80 000(lower triangles) and 16 000(diamonds).Inset: the overlap φ versus the number of patterns n on random networks on a linear coordination.Each data point is averaged over 1000 realizations.

    For ER random network,we let〈k〉/N=0.01,when the network sizes of N=5000,10 000,20 000,40 000,80 000 and 160 000 nodes,the final overlap φ0respectively is 0.207,0.187,0.172,0.159,0.149 and 0.140.When the density of connection is the constant,the final overlap φ0decreases with network size,which is similar to the results of the SF networks.

    We fitted the φ - φ0curve in figure 11.When the network sizes is N=10 000,20 000,40 000,80 000 and 160 000,the slope respectively is -0.687 ± 0.002,-0.670 ± 0.008,-0.563 ± 0.006,-0.555 ± 0.005 and-0.525 ± 0.008.With the network size N gets larger,the exponent close to the theoretical result -0.5.

    3.3.Network with randomly degree distribution

    The SNR analysis suggests that the appearance of power-law decay of overlap does not depend on the network degree distribution.We use P(k) to represent an arbitrary degree distribution.The number of flipped nodes is

    The overlap φ between stable state and initial pattern can be written as

    When we substitute equation (31) into equation (32),we get

    where,

    Hence the overlap φ is

    The overlap φ is

    When n ?1,an approximate formula reads as

    under the condition that the value ofis small.

    We consider the network with the distribution that the node degree k follows a randomly generated histogram whose mean value 〈k〉 is controlled.The histogram are generated as follows:for a given value of average degree〈k〉,the range of degree k is

    Figure 12.The overlap difference φ - φ0 versus the number of patterns n on randomly degree distribution networks on a log-log scale for two average degree 〈k〉=25(squares) and 〈k〉=50(circles).Data points from n=21 to 10 000 are used in fitting(squares).Data points from n=59 to 10 000 are used in fitting(circles).The size of networks is N=10000.Inset: the overlap φ versus the number of patterns n on randomly degree distribution networks on a linear coordination.The final overlap φ0 is 0.21(squares) and 0.19(circles).Each data point is averaged over 1000 realizations.

    where 2Δ + 1 is the width the range of k.The probability that a node has the degree k is

    where εkis a random number uniformly distributed in [0,1].

    By computer simulation,we obtain the relation between the overlap and the number of stored patterns for two average degrees 〈k〉=25 or 50.Figure 12 shows that overlap turns from a constant to the power law decay as the number of stored patterns increase.For the average degree〈k〉=25,the exponent of the power law decay is -0.658 ± 0.003.As the average degree increases,the capacity of the network increases.The crossover from a constant to the power law decay shifts to right as the average increases from 25 to 50.For the average degree 〈k〉=50,the exponent of the power law decay is -0.683 ± 0.004.These results are qualitatively consistent with the above simulation results.Therefore,the SNR analysis results also are applicable to the network with randomly degree distribution.These results suggest that under the condition of sparse connection density,the power law decay can occurs on various kinds of networks.

    4.Conclusions

    We analytically studied the relation between the stability and the number of stored patterns in Hopfield neural networks.Using the SNR method,we analyzed the stability of patterns on SF networks and ER random networks.We show that the stability can be expressed by the error function.Under the condition that the network is sparse and the number of stored patterns is large,the stability can be approximated by a power law function with the exponent -0.5.

    We performed computer simulations on both SF networks and ER random networks.In previous work the power law relation between the stability and the number of stored patterns was obtained in SF networks.Here we show that as the network is sparse enough both the SF and random topology can exhibit the power law relation with the exponent close to -0.5,which agrees with the theoretical result.On dense networks,the relation has a power law tail in the simulation result.

    There is a difference between analytic results and simulation results.First,the overlap in simulations decay to a constant which decreases with the connection density,while the analytic results of overlap decay to 0.The difference may due to that the noise term in simulation deviates from the exact Gaussian distribution which is used in SNR analysis.Furthermore,node state updates in a serial order,the local field of nodes changes as neighbors flips.The difference manifests itself in finite size networks.Second,the exponent of the power law decay is smaller than -0.5 on dense networks with finite size.As the network size increases,the exponent close to -0.5.

    We also numerically check the network with randomly degree distribution besides ER and SF networks.We showed that the power law decay with an exponent closes to -0.5 in the network,which implies the universality of this phenomenon,regardless of the network architecture.

    Different from the forgetting[18-20],here the power law behavior is not about time.When consider memory accumulation in neural network,it is natural to assume that memory is steadily filled up [10].Then the relation between the overlap and the number of stored patterns is potentially helpful for understanding power law forgetting.

    Acknowledgments

    This work was supported by NSFC (Grant No.11 675 096),and FPALAB-SNNU (Grant No.16QNGG007).

    亚洲精品粉嫩美女一区| 欧美乱妇无乱码| a在线观看视频网站| 国产精品美女特级片免费视频播放器 | 亚洲成人久久性| 日韩 欧美 亚洲 中文字幕| 欧美性猛交黑人性爽| 久久久久九九精品影院| 观看免费一级毛片| 不卡一级毛片| 99在线人妻在线中文字幕| 免费人成视频x8x8入口观看| 色综合欧美亚洲国产小说| 女人爽到高潮嗷嗷叫在线视频| 午夜福利高清视频| 国产精品爽爽va在线观看网站 | 99国产精品一区二区三区| 久久久久亚洲av毛片大全| 国产一区二区在线av高清观看| 亚洲五月天丁香| 久久草成人影院| 9191精品国产免费久久| 久久国产乱子伦精品免费另类| 国产一区二区三区在线臀色熟女| 两人在一起打扑克的视频| 久久国产亚洲av麻豆专区| 国产99白浆流出| 宅男免费午夜| 一本久久中文字幕| 亚洲欧美日韩高清在线视频| 久久精品91蜜桃| 欧美午夜高清在线| 香蕉国产在线看| xxx96com| 精品电影一区二区在线| 国产精品精品国产色婷婷| 久久这里只有精品19| 精品久久久久久久人妻蜜臀av| 少妇粗大呻吟视频| 黑丝袜美女国产一区| 亚洲一卡2卡3卡4卡5卡精品中文| 中文字幕高清在线视频| 国产亚洲精品第一综合不卡| 男女之事视频高清在线观看| 国产久久久一区二区三区| 色老头精品视频在线观看| 丁香欧美五月| cao死你这个sao货| 国产成人啪精品午夜网站| 欧美日韩瑟瑟在线播放| 色综合亚洲欧美另类图片| 欧美黑人巨大hd| 精品久久久久久成人av| 亚洲国产精品合色在线| 欧美午夜高清在线| 在线观看免费午夜福利视频| 91九色精品人成在线观看| 午夜免费成人在线视频| 亚洲七黄色美女视频| 国产单亲对白刺激| 婷婷精品国产亚洲av| 久久精品夜夜夜夜夜久久蜜豆 | 成在线人永久免费视频| 这个男人来自地球电影免费观看| 色尼玛亚洲综合影院| 又黄又粗又硬又大视频| 一二三四在线观看免费中文在| 国产午夜福利久久久久久| 精品人妻1区二区| 国产蜜桃级精品一区二区三区| or卡值多少钱| 一边摸一边抽搐一进一小说| 欧美久久黑人一区二区| 中文亚洲av片在线观看爽| 国产高清激情床上av| 国产成人精品久久二区二区91| 精品午夜福利视频在线观看一区| 国产精品一区二区精品视频观看| 国产私拍福利视频在线观看| 日韩高清综合在线| 久久青草综合色| 亚洲电影在线观看av| 亚洲精品在线观看二区| 女性被躁到高潮视频| 成年免费大片在线观看| www.熟女人妻精品国产| 国产精品1区2区在线观看.| 国产av一区二区精品久久| 色播亚洲综合网| 国产精品永久免费网站| 99久久久亚洲精品蜜臀av| 黄色视频不卡| 国内精品久久久久久久电影| 亚洲av五月六月丁香网| 欧美性猛交╳xxx乱大交人| 久久久国产成人免费| 99精品在免费线老司机午夜| 波多野结衣av一区二区av| 18禁美女被吸乳视频| 色综合欧美亚洲国产小说| 亚洲欧美精品综合一区二区三区| 香蕉av资源在线| 亚洲国产欧美一区二区综合| 亚洲国产精品久久男人天堂| 999久久久精品免费观看国产| 波多野结衣高清作品| 在线观看日韩欧美| 黄色丝袜av网址大全| 十分钟在线观看高清视频www| 哪里可以看免费的av片| 亚洲精品色激情综合| or卡值多少钱| 成年人黄色毛片网站| 亚洲精品国产一区二区精华液| 淫妇啪啪啪对白视频| 激情在线观看视频在线高清| 少妇粗大呻吟视频| 国产黄色小视频在线观看| 色av中文字幕| 身体一侧抽搐| 嫩草影院精品99| 精品欧美国产一区二区三| 日韩大尺度精品在线看网址| 亚洲成人久久性| 国产精品自产拍在线观看55亚洲| 欧洲精品卡2卡3卡4卡5卡区| 欧美+亚洲+日韩+国产| 欧美性猛交╳xxx乱大交人| 精品午夜福利视频在线观看一区| 一边摸一边做爽爽视频免费| 90打野战视频偷拍视频| 久久久久精品国产欧美久久久| 日本黄色视频三级网站网址| 叶爱在线成人免费视频播放| xxx96com| 亚洲国产精品999在线| 亚洲黑人精品在线| 亚洲av中文字字幕乱码综合 | 亚洲人成电影免费在线| 色婷婷久久久亚洲欧美| 亚洲久久久国产精品| 一进一出抽搐动态| 亚洲色图av天堂| 国产精品,欧美在线| 别揉我奶头~嗯~啊~动态视频| 久久久久免费精品人妻一区二区 | 十分钟在线观看高清视频www| 午夜激情av网站| 美女高潮喷水抽搐中文字幕| 在线看三级毛片| 欧美 亚洲 国产 日韩一| 日韩欧美一区二区三区在线观看| 好男人在线观看高清免费视频 | 麻豆av在线久日| 欧美最黄视频在线播放免费| 亚洲第一av免费看| 麻豆国产av国片精品| 精品国产国语对白av| 久久人人精品亚洲av| 少妇裸体淫交视频免费看高清 | 夜夜躁狠狠躁天天躁| 精品电影一区二区在线| 午夜亚洲福利在线播放| 啦啦啦观看免费观看视频高清| 欧美三级亚洲精品| 黄频高清免费视频| 一级a爱视频在线免费观看| 免费电影在线观看免费观看| 狂野欧美激情性xxxx| 久热爱精品视频在线9| 天堂√8在线中文| 国产97色在线日韩免费| 亚洲精品在线美女| 最近最新中文字幕大全免费视频| 免费在线观看日本一区| 男人舔女人下体高潮全视频| 免费在线观看亚洲国产| 一区二区三区激情视频| 国产黄色小视频在线观看| 亚洲第一电影网av| 99久久久亚洲精品蜜臀av| 国产精品九九99| 欧美人与性动交α欧美精品济南到| 又黄又粗又硬又大视频| 国产精品九九99| 欧美另类亚洲清纯唯美| 精品人妻1区二区| 麻豆av在线久日| 久久久久久久久中文| 精品国产一区二区三区四区第35| 一级作爱视频免费观看| 国产黄片美女视频| 久久久久久大精品| 久久国产精品人妻蜜桃| 国产亚洲精品第一综合不卡| 在线观看日韩欧美| 久久香蕉精品热| 视频区欧美日本亚洲| 熟女少妇亚洲综合色aaa.| 999精品在线视频| 亚洲av第一区精品v没综合| 我的亚洲天堂| 久久精品aⅴ一区二区三区四区| 免费在线观看黄色视频的| 757午夜福利合集在线观看| 美女高潮喷水抽搐中文字幕| 精品日产1卡2卡| 日韩欧美 国产精品| 最近最新中文字幕大全免费视频| 精品卡一卡二卡四卡免费| 久热这里只有精品99| 国产国语露脸激情在线看| 女性被躁到高潮视频| 国产一区在线观看成人免费| 欧美色视频一区免费| 熟妇人妻久久中文字幕3abv| 国产91精品成人一区二区三区| 午夜福利视频1000在线观看| 婷婷亚洲欧美| 国产精品免费一区二区三区在线| 亚洲五月婷婷丁香| 一进一出好大好爽视频| 老汉色av国产亚洲站长工具| 亚洲成av人片免费观看| 怎么达到女性高潮| 精品少妇一区二区三区视频日本电影| 亚洲片人在线观看| 日本 欧美在线| 国产av一区在线观看免费| 欧美精品亚洲一区二区| 亚洲精品久久成人aⅴ小说| 69av精品久久久久久| www国产在线视频色| 国产亚洲精品一区二区www| 国产三级在线视频| 丝袜美腿诱惑在线| 久久久国产成人精品二区| 亚洲最大成人中文| 久久天堂一区二区三区四区| 午夜福利视频1000在线观看| 手机成人av网站| 大型av网站在线播放| 亚洲精品av麻豆狂野| 久久婷婷人人爽人人干人人爱| 中文字幕精品亚洲无线码一区 | www.999成人在线观看| 亚洲人成伊人成综合网2020| 亚洲人成网站高清观看| 国产视频内射| 国产高清videossex| 成人国产综合亚洲| 午夜福利一区二区在线看| 亚洲精华国产精华精| 国产片内射在线| 亚洲中文字幕日韩| 男女床上黄色一级片免费看| 99精品久久久久人妻精品| 精品卡一卡二卡四卡免费| 久久久国产欧美日韩av| 国产高清有码在线观看视频 | 日日干狠狠操夜夜爽| 99久久综合精品五月天人人| 国产精品国产高清国产av| 99在线人妻在线中文字幕| 欧美日韩福利视频一区二区| 久久久久国产精品人妻aⅴ院| 亚洲精品粉嫩美女一区| 成年女人毛片免费观看观看9| www日本在线高清视频| 国产一区二区激情短视频| 国产成人啪精品午夜网站| 99国产极品粉嫩在线观看| 久久久久国内视频| 国产伦在线观看视频一区| 亚洲av成人av| 丰满人妻熟妇乱又伦精品不卡| 免费看a级黄色片| 男人的好看免费观看在线视频 | 久久婷婷成人综合色麻豆| 亚洲国产欧美一区二区综合| 曰老女人黄片| 亚洲精品中文字幕一二三四区| 成人精品一区二区免费| 女生性感内裤真人,穿戴方法视频| 亚洲欧洲精品一区二区精品久久久| 久久久国产欧美日韩av| 亚洲成a人片在线一区二区| 悠悠久久av| 国产精品久久视频播放| 欧美不卡视频在线免费观看 | 婷婷丁香在线五月| 伊人久久大香线蕉亚洲五| 午夜亚洲福利在线播放| 国产精品久久视频播放| 男女视频在线观看网站免费 | 黄色片一级片一级黄色片| 午夜两性在线视频| 日韩大码丰满熟妇| 国产亚洲精品久久久久久毛片| 亚洲七黄色美女视频| 欧美日本视频| 久久中文字幕一级| 国产伦一二天堂av在线观看| 叶爱在线成人免费视频播放| 亚洲一区二区三区不卡视频| or卡值多少钱| 韩国av一区二区三区四区| 99在线人妻在线中文字幕| 久久人妻av系列| 久久婷婷人人爽人人干人人爱| 99久久综合精品五月天人人| 美女午夜性视频免费| 成人国语在线视频| 村上凉子中文字幕在线| 两人在一起打扑克的视频| 日韩免费av在线播放| 1024手机看黄色片| 亚洲精品国产区一区二| 成年人黄色毛片网站| 国产精品亚洲一级av第二区| 精品久久久久久久末码| 亚洲精品一卡2卡三卡4卡5卡| 国产成人精品无人区| 搡老熟女国产l中国老女人| 国产午夜精品久久久久久| 九色国产91popny在线| 丰满的人妻完整版| 国产精品永久免费网站| 少妇被粗大的猛进出69影院| 老司机福利观看| 亚洲国产毛片av蜜桃av| 性欧美人与动物交配| 亚洲av成人不卡在线观看播放网| 国产麻豆成人av免费视频| 国产一区二区三区视频了| 十八禁人妻一区二区| 国语自产精品视频在线第100页| 午夜亚洲福利在线播放| 色婷婷久久久亚洲欧美| 色播亚洲综合网| 成人国产综合亚洲| 精品久久久久久久末码| 亚洲人成网站在线播放欧美日韩| 这个男人来自地球电影免费观看| 一区二区三区激情视频| 好男人在线观看高清免费视频 | 十分钟在线观看高清视频www| 亚洲精品在线观看二区| 最新在线观看一区二区三区| 婷婷精品国产亚洲av在线| 国产极品粉嫩免费观看在线| 国产精品影院久久| 女人被狂操c到高潮| 精品人妻1区二区| 免费在线观看亚洲国产| 精品久久久久久,| 在线av久久热| 国产在线观看jvid| 国产精品久久电影中文字幕| 一区二区日韩欧美中文字幕| 日本三级黄在线观看| 男人舔女人下体高潮全视频| www.熟女人妻精品国产| 欧美黄色片欧美黄色片| 高潮久久久久久久久久久不卡| 亚洲,欧美精品.| 日韩欧美国产在线观看| 亚洲第一青青草原| 看免费av毛片| 99在线人妻在线中文字幕| 成人一区二区视频在线观看| 国产三级黄色录像| 久久久久久久久中文| 久久久久久久午夜电影| 亚洲av成人一区二区三| 一个人免费在线观看的高清视频| 亚洲专区字幕在线| 亚洲av熟女| 在线十欧美十亚洲十日本专区| 欧美日韩乱码在线| 久久久久国产精品人妻aⅴ院| 51午夜福利影视在线观看| 中文字幕精品免费在线观看视频| 亚洲国产精品成人综合色| www日本在线高清视频| 狠狠狠狠99中文字幕| 国产精品亚洲美女久久久| 免费高清在线观看日韩| 日本熟妇午夜| 久久亚洲真实| 亚洲五月色婷婷综合| 一区二区三区精品91| 夜夜看夜夜爽夜夜摸| 黄片大片在线免费观看| 老汉色∧v一级毛片| 欧美日韩福利视频一区二区| 国产男靠女视频免费网站| 免费在线观看成人毛片| 久久国产精品男人的天堂亚洲| 91麻豆av在线| 香蕉av资源在线| 亚洲人成网站高清观看| 在线天堂中文资源库| 波多野结衣巨乳人妻| 一个人免费在线观看的高清视频| 欧美性长视频在线观看| 1024视频免费在线观看| 精品国产超薄肉色丝袜足j| 最近在线观看免费完整版| 一区二区三区高清视频在线| 99久久精品国产亚洲精品| 亚洲真实伦在线观看| 十分钟在线观看高清视频www| 国产又爽黄色视频| 午夜久久久久精精品| 岛国视频午夜一区免费看| 精品久久久久久久久久久久久 | 亚洲国产中文字幕在线视频| 精品乱码久久久久久99久播| 人妻丰满熟妇av一区二区三区| 美女高潮到喷水免费观看| 人妻丰满熟妇av一区二区三区| 国产一卡二卡三卡精品| 法律面前人人平等表现在哪些方面| 一卡2卡三卡四卡精品乱码亚洲| 精品一区二区三区视频在线观看免费| 精品国产乱码久久久久久男人| 人人澡人人妻人| 国产成人系列免费观看| 嫁个100分男人电影在线观看| 亚洲午夜精品一区,二区,三区| 精品不卡国产一区二区三区| 欧美黑人欧美精品刺激| 美女扒开内裤让男人捅视频| www.自偷自拍.com| 午夜福利视频1000在线观看| 欧美不卡视频在线免费观看 | 国产不卡一卡二| 久久香蕉激情| 日韩精品中文字幕看吧| 国产一区二区激情短视频| 亚洲天堂国产精品一区在线| 免费高清视频大片| 精品午夜福利视频在线观看一区| 亚洲第一av免费看| 最近最新中文字幕大全电影3 | 欧美中文综合在线视频| 亚洲欧美日韩无卡精品| 亚洲人成伊人成综合网2020| 丝袜在线中文字幕| www.www免费av| 麻豆成人av在线观看| 日韩av在线大香蕉| 国产亚洲精品综合一区在线观看 | 国产精品av久久久久免费| 最新美女视频免费是黄的| 亚洲成人精品中文字幕电影| 99热6这里只有精品| 丝袜在线中文字幕| 青草久久国产| 一边摸一边做爽爽视频免费| av视频在线观看入口| 国产精品亚洲美女久久久| videosex国产| xxx96com| 日韩欧美一区二区三区在线观看| 国产成人欧美在线观看| 亚洲成人久久性| 中国美女看黄片| 精品卡一卡二卡四卡免费| 久久久久久久午夜电影| 日日夜夜操网爽| 天堂动漫精品| 久久国产精品人妻蜜桃| 国产免费av片在线观看野外av| 亚洲成国产人片在线观看| 国产精品久久久久久亚洲av鲁大| 欧美日韩一级在线毛片| 日本成人三级电影网站| 日韩av在线大香蕉| 天天一区二区日本电影三级| 国产亚洲精品久久久久5区| 韩国av一区二区三区四区| 女生性感内裤真人,穿戴方法视频| 最近最新中文字幕大全电影3 | 麻豆av在线久日| 国产精品一区二区精品视频观看| 久久中文字幕一级| 亚洲国产日韩欧美精品在线观看 | 日韩大码丰满熟妇| 一级a爱视频在线免费观看| 亚洲欧美一区二区三区黑人| 久久精品国产综合久久久| 久久久久久国产a免费观看| 久久久久久久久免费视频了| 两个人免费观看高清视频| 特大巨黑吊av在线直播 | av片东京热男人的天堂| 精品电影一区二区在线| 丝袜人妻中文字幕| 一本综合久久免费| 中文字幕另类日韩欧美亚洲嫩草| 国产亚洲精品综合一区在线观看 | 黄频高清免费视频| 色综合婷婷激情| 黄色视频不卡| 国产1区2区3区精品| 19禁男女啪啪无遮挡网站| 我的亚洲天堂| 男女床上黄色一级片免费看| 亚洲天堂国产精品一区在线| av欧美777| 国产亚洲欧美精品永久| 亚洲第一av免费看| 婷婷丁香在线五月| 搡老岳熟女国产| 少妇的丰满在线观看| 在线视频色国产色| 成年免费大片在线观看| 波多野结衣av一区二区av| 精品欧美一区二区三区在线| 精华霜和精华液先用哪个| 欧美色欧美亚洲另类二区| 国产av在哪里看| 桃色一区二区三区在线观看| 伊人久久大香线蕉亚洲五| 最近在线观看免费完整版| 午夜两性在线视频| 欧美成人一区二区免费高清观看 | 51午夜福利影视在线观看| 色综合婷婷激情| 精品久久久久久久毛片微露脸| 精品高清国产在线一区| 欧美黑人欧美精品刺激| 国产又黄又爽又无遮挡在线| 男女做爰动态图高潮gif福利片| 婷婷丁香在线五月| 日韩成人在线观看一区二区三区| 亚洲成国产人片在线观看| 亚洲欧美日韩无卡精品| 成人三级做爰电影| 欧美黑人精品巨大| 日韩欧美在线二视频| 国产精品,欧美在线| 高潮久久久久久久久久久不卡| 女生性感内裤真人,穿戴方法视频| 精品高清国产在线一区| 香蕉丝袜av| 日韩国内少妇激情av| av视频在线观看入口| 午夜激情av网站| 丝袜人妻中文字幕| 久久婷婷成人综合色麻豆| 久9热在线精品视频| 色哟哟哟哟哟哟| 国产av又大| 老熟妇仑乱视频hdxx| 日韩一卡2卡3卡4卡2021年| 国产精品久久久久久亚洲av鲁大| 国产区一区二久久| 国产一区在线观看成人免费| 黄色丝袜av网址大全| 国产91精品成人一区二区三区| 中文在线观看免费www的网站 | 午夜福利一区二区在线看| 熟女少妇亚洲综合色aaa.| 欧美黑人精品巨大| 伦理电影免费视频| 亚洲久久久国产精品| 亚洲中文日韩欧美视频| 日韩视频一区二区在线观看| 亚洲一区二区三区不卡视频| 亚洲人成77777在线视频| a在线观看视频网站| 成人特级黄色片久久久久久久| 久久午夜亚洲精品久久| 精品电影一区二区在线| 免费在线观看亚洲国产| 男女午夜视频在线观看| 国产区一区二久久| 精品国产一区二区三区四区第35| www.熟女人妻精品国产| 日本精品一区二区三区蜜桃| 成人一区二区视频在线观看| 91成年电影在线观看| 欧美性长视频在线观看| 国产精品久久视频播放| 亚洲成人免费电影在线观看| 亚洲成av人片免费观看| 天天躁夜夜躁狠狠躁躁| 日韩一卡2卡3卡4卡2021年| 亚洲国产精品999在线| 欧洲精品卡2卡3卡4卡5卡区| 美女高潮喷水抽搐中文字幕| 色综合婷婷激情| 满18在线观看网站| 一边摸一边做爽爽视频免费| 中文字幕人妻熟女乱码| 色在线成人网| 午夜成年电影在线免费观看| 一进一出好大好爽视频| 中文字幕久久专区| av在线天堂中文字幕| www日本在线高清视频| 最近最新免费中文字幕在线| 亚洲中文字幕日韩| 一级作爱视频免费观看| 国产三级在线视频| 欧洲精品卡2卡3卡4卡5卡区| 欧美午夜高清在线| 国产激情偷乱视频一区二区| 欧美日韩乱码在线| 精品福利观看| 国产亚洲av嫩草精品影院| www.自偷自拍.com|