• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Deep Q-Learning Based Optimal Query Routing Approach for Unstructured P2P Network

    2022-03-14 09:26:30MohammadShoabandAbdullahShawanAlotaibi
    Computers Materials&Continua 2022年3期

    Mohammad Shoab and Abdullah Shawan Alotaibi

    Department of Computer Science,Faculty of Science at Al Dawadmi,Shaqra University,Shaqra,Saudi Arabia

    Abstract:Deep Reinforcement Learning(DRL)is a class of Machine Learning (ML) that combines Deep Learning with Reinforcement Learning and provides a framework by which a system can learn from its previous actions in an environment to select its efforts in the future efficiently.DRL has been used in many application fields, including games, robots, networks, etc.for creating autonomous systems that improve themselves with experience.It is well acknowledged that DRL is well suited to solve optimization problems in distributed systems in general and network routing especially.Therefore,a novel query routing approach called Deep Reinforcement Learning based Route Selection(DRLRS)is proposed for unstructured P2P networks based on a Deep Q-Learning algorithm.The main objective of this approach is to achieve better retrieval effectiveness with reduced searching cost by less number of connected peers,exchanged messages,and reduced time.The simulation results shows a significantly improve searching a resource with compression to k-Random Walker and Directed BFS.Here, retrieval effectiveness, search cost in terms of connected peers, and average overhead are 1.28, 106, 149,respectively.

    Keywords: Reinforcement learning; deep q-learning; unstructured p2p network; query routing

    1 Introduction

    Machine Learning(ML) is widely used to analyze the data and to create or assisting in the formulation of predictions with the help of some algorithms and methods in the discipline of computer science and statistics [1-3].In today’s computerized societies, machine learning plays a key role, and ML-based components will undoubtedly be included in almost every gadget and machines to better control the operations and accustom to their environment.Machine learning is an excellent technique for resolving complex issues and has proven its efficacy in image and speech recognition, robot guidance, autonomous automobile guidance, telecommunications, and various other fields.ML is recognized to give (often) better outcomes than humans for tasks such as categorization and optimization [4].

    Based on how learning is carried out, machine learning approaches are divided into four categories [1,2]: supervised, unsupervised, semi-supervised, and reinforcement.In supervised learning,input and output variables are used to learn the mapping function from input to output; the goal is to approximate the mapping function to the point where an output (also known as the label)can be accurately predicted from its associated input.Unsupervised learning, also known as learning without teacher, only input is used; the goal is to model the structure or distribution of data(e.g., data clustering) in order to understand specific features about data.Semi-supervised learning is similar to supervised learning, with the exception that not all observations are labeled (outputs).Finally, reinforcement learning is a behavioral psychology-inspired technique that enables system modeling based on agents interacting with their environments [5,6].In the sequel, the paper only focuses on deep reinforcement learning (DRL) application to route the query in unstructured P2P networks.

    The two types of Reinforcement Learning algorithms are model-free and model-based RL algorithms.Model-free RL algorithms predict future states and rewards without learning a model of their environment’s transition function.The best examples of model-free learning are Policy Gradient, Deep Q-Networks, and Q learning, as they don’t create a model of the environments transition function [6].The existence of two things distinguishes reinforcement learning from supervised or unsupervised learning:

    ? An environment: This may be a maze, a video game, the financial market, or something else.

    ? An agent: This is the AI that learns how to operate and succeed in a given environment.

    An iterative feedback loop is used to teach the agent how to operate in the environment.The state will alter as a result of the action taken by the agent, depending on the rewards received either won or lost.Here Fig.1 shows a visual representation of this iterative feedback loop of actions, states, and rewards.The agent can learn which actions are beneficial in a given state by taking actions and obtaining rewards from the environment.

    Figure 1: Iterative feedback loop of Deep Q Learning (Source: Sutton and Barto (2018))

    DRL (Deep Reinforcement Learning) is a rapidly developing field that combines Reinforcement Learning and Deep Learning.It’s also the most popular sort of Machine Learning since it can handle a wide range of complicated decision-making tasks previously unsolvable by a machine with human-like intelligence.The “deep” part of reinforcement learning refers to artificial neural networks with several (deep) layers that mimic the structure of the human brain [7].Deep learning necessitates a lot of training data and a lot of computer power.The Deep Q-Learning algorithm is one of the most important principles in DRL.Instead of explicitly computing Q-values in Q learning through value iterations, a function approximator could be employed to estimate the effective Q-function, with neural networks currently being the method of choice [8].Deep QLearning incorporates artificial neural networks into the Q-learning process, and a network that uses neural networks has been used to approximate Q-functions [9].

    The complexity and variability of modern P2P networks, as well as end-user quality of service(QoS) and security requirements, service provider economics, and social inter-networking, have all expanded dramatically since the early P2P networks.P2P network technology has progressed from wired and manually configured networks to highly dynamic and autonomous networks,particularly unstructured P2P networks [4].The majority of today’s networks have evolved beyond human administration and configuration.As a result, machine learning approaches have been used to handle concerns and challenges in the networking area, such as traffic classification and prediction, fault management, configuration management, congestion control, QoS monitoring,energy efficiency, and security management [10-12].The goal of ML applications to networks is to automatically learn the dynamics of P2P networks, including new flow arrivals, congestion points,topology changes, link quality, and energy consumption, in order to improve the service quality provided to end-users while optimizing network resources and providers’revenues.

    In P2P networks, query routing is the problem of selecting paths to search the requested resource, while meeting QoS requirements, if any.Searching in P2P systems has been achieved by two techniques Blind search and Informed search.These search techniques relied on various query mechanisms in unstructured P2P systems, such as gossiping [13-15], random walk [16,17],k-walker [16,18], controlled flooding, and pure flooding [19].Among these, gossiping is an attractive and widely adopted mechanism for modern query routing approaches [15,20].Rather than selectingkneighbors at random, as k-walker, random walk, and controlled flooding do, gossiping routes the query to thekneighbors with the highest probability of holding the requested resource [14,15,20,21].The current gossiping mechanism introduces the K-Neighbor Selection (KNS) problem, which uses a particular scoring function to select thekrelevant neighbors with the maximum score.To overcome this problem, two query routing approaches have been proposed:content-oriented and query-oriented routing methods [20].An index about neighbor’s collection has been created on each peer by previously collected data.This index with a specific scoring function assigns the neighbors’weights according to their query content and shared resources.As a result, the query has been routed to the firstkhigh-scoring neighbors.The Query-oriented routing approach selects the suitablekanticipated neighbors for future queries based on data acquired from previously sent queries.In order to define or develop scoring functions, several machine learning and data mining techniques have been applied [20,21].

    Query-oriented methods are more efficient than content-oriented methods as they only use data collected from previous queries, so no additional network communication is required to create and maintain the index [14,15,20,21].In this respect, a novel query routing algorithm, ‘Deep Reinforcement Learning Route Selection (DRLRS)’has been introduced for efficient neighbor selection in unstructured P2P networks using the Deep Q-Learning algorithm.Deep Q-Learning is a part of reinforcement learning where intelligent agents take actions in an uncertain and potentially complex environment to maximize cumulative reward.The primary objective of the DRLRS algorithm is to reduce the number of connected peers and overhead in order to achieve high retrieval effectiveness with lower communication costs.This study’s three key contributions are listed below, along with their novelty.

    (1) This study introduced a fully distributed query routing algorithm DRLRS for unstructured P2P network that depends on the locally gathered data for learning on every peer.

    (2) The K-NS problem has been provided a new formulation, and deep reinforcement learning has been considered to formalized for natural model for this issue in which a peer learns through performing actions that result in a reward after completion of every search.As a result, a peer’s aim is to discover a neighbor selection policy that maximizes the total reward.To achieve this, a Deep Q-Learning algorithm is used at every forwarder peer to select thekneighbors for each search query.Every selected neighbor generates a binomial reward that indicates its capacity to locate the query’s relevant resource.To optimize the cumulative reward, a peer must learn a K-NS strategy to generate greater user satisfaction.

    (3) The cold start problem, which is a significant flaw in query routing methods, is also addressed.This issue arises when a new peer enters the network, and traditional queryoriented approaches presume that the relevant peer has already submitted a specific number of queries and replies, which are saved in the sender log file as training data.However, for the recently connected peer, this assumption is incorrect.As a result, existing approaches for generating training data required the deployment of the k-random walker strategy for a given number of queries, which surely resulted in poor performance at first.To address this issue, a Deep Q-Learning-based route selection algorithm was developed, which learns about neighbors and chooses the neighbor with the largest estimated reward based on prior queries.

    2 Existing Systems and Algorithms

    In this section, a discussion has been made about some unstructured P2P systems’routing methods.The main objective of these systems and algorithms is to find those peers sharing related resources for the queries with decreasing number of connected peers and network traffic.Therefore, these systems and algorithms have been divided in to the following three categories:

    ? Basic Searching Algorithms:The most common and foremost algorithms to search relevant resources in unstructured P2P systems are Breadth-First Search (BFS) or flooding [22,23].In this algorithm, the forwarder peer which initiates or receives a query first searches the related resource in the local collection of resources.When a resource is found in the local collection, it will reply to the requesting peer with a message containing a list of related resources that have been retrieved.The message is forwarded through the reverse request path to the requesting peer.This is where the query propagation process stops until the Time To Live (TTL) reaches a specific predefined value [22].BFS tries to find the maximum number of results.However, this generates a large number of messages and increases the number of connected peers with heavy network traffic compared to other approaches [23].Napster and Gnutella are the suitable examples who implemented this routing method [24,25].

    Although, this method is robust, it is prolonged and consumes network resources excessively.Hence, several other methods such as Depth First Search (DFS), controlled flooding,k-walker, and gossiping were introduced as an improvement of BFS.In contrast to BFS,which sends queries to all neighbors, DFS allows each peer to choose a candidate neighbor to whom the query should be sent.If the query forwarder doesn’t get a response within a specific amount of time (TTL), the peer chooses another neighbor to deliver the query [23].This is an iterative process that is performed until the query is answered or until all of the neighbors have been queried.FreeNet is an example of a P2P system that uses the DFS approach [26,27].Instead of sending the query to all neighboring peers as BFS does, the controlled flooding forwards the query to the arbitrarily selectedkneighbors.When DFS and controlled flooding are combined, another method called K-Walker is created.The peer sends the query tokrandomly chosen neighbors, who then send the request to the next random neighbor, and so on, until the relevant peer is discovered or the TTL value is reached [22].Another exciting technique widely used in intelligent query routing methods is Directed BFS.The primary purpose of this searching technique is to direct the query tokneighbors that may have appropriate query resources [14,15,21], rather than arbitrarily selecting neighbors.Therefore, a specific scoring algorithm is employed to identify thekhighest scoring relevant neighbors.The particular scoring function utilizes the previously collected meta-data about neighbors to rank neighbors concerning the query content and shared resource.Therefore, the query is directed to the topkhigh-scoring neighbors [28,29].

    ? Content-Oriented Algorithms:These algorithms utilize information extracted from each peers’shared content and create a local index with global knowledge.Further, this index provides an approximate view of the entire network shared content with peers’profiles.Therefore, a query forwarder peer will be able to route the query with better retrieval effectiveness efficiently.An algorithm based of notation of semantic communities is Improved Niche Genetic Algorithm (INGA) [30].Each peer is expected to play a specific role in the network, such as recommender, content provider, and so on.INGA can determine the most suitable peer to redirect the query by the function associated with it.Moreover, each peer collects and manages information and facts locally that creates a topical knowledge of the peer.Scalable Query Routing (SQR) is another algorithm that aim to achieve low bandwidth [31].This algorithm keeps a routing table on each peer that, based on previous experience, advises the position of items in the network.Furthermore, the Exponentially Decaying Bloom Filter (EDBF) data structure compresses probabilistic routing tables and enables for efficient query propagation [23].

    Another algorithm similar to directed BFS and intelligent search is the routing index-based search algorithm.The entire search process has been guided by three approaches using neighbor information, but the information collection and utilization is differ.In directed BFS, neighbors’information has been utilized by only the query issuing peer; the rest of the peers uses BFS to forward the query if they don’t find the requested resource in their local collection.Whereas, Intelligent search makes use of information from earlier queries that have been replied by neighbors.However, the routing-index based search method stores information about the number of documents and subjects of documents available at neighbor peers.This information helps to select the best neighbor peer to forward the query [32].For query routing, specific P2P systems use a classification problem.The classifier in this issue tries to classify an item based on particular features.The Semantic Overlay Model adepts the classification problem to locate appropriate peers to answer a specific query.This model routes the query to the semantically similar peers instead of broadcasting the query.To classify into categories, semantic vectors have been used that represents the peers’semantic similarity and uses meta-information to classify peers by interests.As a result, it increases the recall rate and, at the same time, reduce the hops and messages [33].

    ? Query-Oriented Algorithms:The Query-oriented routing methods relies on previously collected data in previously sent queries in order to selectkneighbors for the upcoming queries.To attain this, different statistical and intelligent methods have been utilized to define or learn the scoring function.In this respect, Alanazi and Yeferny, (2019) proposed a reinforcement learning (RL) based query routing approach, which is based on a classical RL problem called Multi-Armed Bandit (MAB) for the K-NS problem.Initially, to select one relevant neighbor for each search query, three MAB-based learning algorithms are used.These three algorithms are Epsilon-Greedy (EG), Upper Confidence Bound (UCB), and Thompson Sampling (TS) and called as 1-neighbour-Selection (1-NS) algorithms.Further,one of the three algorithms run for K times in the K-NS algorithm to selectkrelevant neighbors.Shamshirband and Soleimani (2021) introduced a novel and efficient query routing algorithm using reinforcement learning with learning automata (LA) called “l(fā)earning automata adaptive probabilistic search (LAAPS) algorithm” and performs a keyword-based search on the routing tables with scores.In this method, the LAAPS estimation of the current state of the P2P system is related to the Markov process of the discrete-parameters of the discrete-state, and depends on the revising action and the selection of the next update stage of each action.Therefore, the goal is to update the action based on the information obtained by interacting with the environment and finding the correct peer for the route.Each peer requesting for routing has an LA with certain operations, and each adjacent peer is considered to be the neighborhood of that node [34].Kalogeraki et al.(2002)introduced a query-oriented method called Intelligent Search (IS).This method manages a routing table in which neighbor configuration files are stored and provides a simple vector representation of each neighbor configuration file, including the recently processed queries by the neighbor and the number of queries received.Every time, a peer initiates or receives a query, it conducts an online evaluation of its neighbors about its configuration and the content of the query.Then send the query to the firstkrelevant neighbors.Formal Concept Analysis (FCA) theory [35] has been used by Arour et al.[20].for efficient query routing in P2P information retrieval systems .Typically, It is a method that extracts interested relational data sets, implicitly extracts user interests from previously submitted queries and associated query-hits, and stores the user interests in a local per peer knowledge base.To route the query, the learning query routing algorithm first calculates the similarity between the user’s interest and the keywords in the query, and then identifies the peers that are closest to the query contained in the user’s interest askrelated peers, and these peers are selected as forwarding peer for the query Alanazi et al.[22].(2009) presented a supervised machine learning algorithm, called “Route Learning”, to solve the K-NS problem.In the learning phase of this routing algorithm, the data about sent queries and related queries is accumulated.In addition, adjacent peers and forthcoming queries are regarded as classes and new objects to be classified in the supervised multi-classification problem.Therefore,neighbors (i.e., classes) to which the query is stipulated are considered as thekrelevant neighbors [21,22].

    Every peer have the same responsibility in pure P2P unstructured and decentralized network.However, some query methods in this type of P2P network uses a notation of super-peer like in Backpressure algorithm [36].In this algorithm, super-peer resolves the query or forward the query to other super-peers for their underling peers.Whenever superpeers received a query first they check the desired resource they have, as well as check their underling group of peers.This algorithm is a query-oriented algorithm that utilize the previously collected information to decide the route of a query.Another algorithm called Self Learning Query Routing Algorithm that aims to improve it’s knowledge by learning the peer’s interest based on the peer’s previous search history [37].Moreover, the rank of friendship between two peers is determined by the number of shared files.So, only friend peers are involved in query routing initially.If the friend peers doesn’t have the desired resource, a broadcast search is performed.Previous results of searching allows peers to learn gradually about other peers in the network that share the same interest [23].

    3 Problem Formulation

    According to the literature, it has been observed that most of the searching techniques in the unstructured P2P network are based on the gossiping method.In this method, the query forwarder peer selectskneighbor from the set of N neighbors according to the relevance for the query.Moreover, there are some other approaches relies on various specific scoring functions to selectkneighbors with the highest score.The formulation to selectkneighbors in this study is different from existing methods.Thus, a natural model called “Deep Reinforcement Learning”for this issue has been considered, in which an agent learns from its previous actions that produce rewards.The agent’s aim is to find a selection criteria that maximize the collective reward.In this way,the forwarder peer selectskneighbors using the deep q learning algorithm for each search query.Every selected neighbor provides a reward value that explicit its ability to produce a relevant resource for the query.In fact, for higher user satisfaction, the agent should learn neighbor selection criteria to maximize the collective reward.However, most existing RL algorithms are designed to select only one neighbor (action) from a set of various neighbors (actions), whereas thekneighbor selection algorithm must select several neighbors for each query.To address this issue, a generic neighbor selection algorithm has been introduced that uses a Deep Q-Learning algorithm to selectkneighbors.

    4 Deep Q-Learning

    Learning Q-values-the value of taking specific action in a given condition-is the foundation of Q-Learning.Deep Q-Networks (DQNs) are similar to tabular Q-learning in principle, but instead of keeping all of Q-values in a look-up table, these have been represented as a neural network in Deep Q-Learning.This allows for more generalization and a more diverse representation [38,39].DQN is a combination of reinforcement learning and deep learning algorithm.This is motivated by the fact that the storage space of a Q-table in the traditional reinforcement learning algorithm Q-Learning is limited, whereas the state in the real world or even the virtual world is nearly infinite, making it impossible to build a Q-table that can store an ample state space.However, in machine learning, there is a method that is particularly good at this, namely, the neural network, which can take the state and action as inputs and then obtain the Q-value of the action after neural network analysis, eliminating the need to record the Q-value in the table and instead of using the neural network to predict the Q-value directly [40].

    In Deep Q-Learning, the user stores all past experiences in memory, as well as the future action determined by the network’s Q-output.Thus, Q-network gains the Q-value at stateSt, and at the same time target network (Neural Network) calculates the Q-value for the stateSt+1(next state) to make the training stabilized and blocks the abruptly increments in Q-value count by copying it as training data on each iterated Q-value of the Q-network.It has already been demonstrated that using a random batch of previous data boosts the stability of neural network training [39].So, to increase agents’performance, Deep Q-Learning employs yet another concept:experience replay, which is nothing more than the stocking of previous experiences.The target network uses experience replay for training and by the Q-network for calculating the Q-value.

    4.1 Target Network and Experience Replay

    Switching to a DQN representation for Q-learning brings a few challenges that the tabular versions do not have to deal with.This is induced by the function approximator’s nonlinear deep neural network and data sequence-dependent correlations and frequent updates to Q approximation.The correlations arise as a result of taking actionaat each time steptwhile looking at an episode.The reward received at the timet(Rt)is strongly correlated to the state and action at the timet-1, which is linked to the state and action at the timet-2 and so on.Because of the path dependence, training a DQN is difficult [9].

    This is handled with experience replay, which is a memory bank comprising different states,actions, and rewards that have been randomly sampled.When the learning has been updated by network smoothing, the random sampling breaks any sequence dependency in the data.There’s also a neuroscience component here, as experience replay is thought to occur in the brain to encode long-term memories [38].

    Estimation of the next state is the second source of instability.The following equation could be used to update values in tabular Q-learning.

    With the tabular method, onlyQ(st,at)is updating, however, using the DQN technique, this update is slightly different, and the entire network is updated at each step.When the whole network is updated, the optimum action is estimated at the next state(maxaQ(st+1)).Because the target is constantly moving, the backpropagating error for the identical states and actions varies from one update to the next, making it more difficult for the network to learn [9,38].

    Here,αis the learning rate,γis the discount factor andmaxαis the maximum reward attainable in the state.The values of bothαandγare generally set between 0 and 1.Settingαto 0 means that the Q-values are never updated, and nothing is learned.However, settingαto a high value such as 0.9 means that learning can occur quickly.Furthermore, ifγis closer to 0, the agent will tend to consider only immediate rewards.Ifγis closer to 1, the agent will consider future rewards with greater weight and be willing to delay the reward.Considering previous studies, the values ofαandγare set to 0.5 and 0.9 in the current study.

    To deal with this, a target network is built that is a clone of the training neural network but only copies it every N time steps.This means that the error will stay steady for a while, enabling the network to learn, before the target network is updated to a better approximation, allowing the network to learn all over again.

    4.2 Loss Function

    Because a deep neural network represents the Q-function, an alteration to the rule is required to make it relevant to backpropagation.This loss function must also be differentiable in terms of the network’s(θ)parameters.The DQN loss function works in the same way as the tabular Q-learning update rule: an action is performed using the Q-function, and the reward received is compared to the best action estimate in the new state [38,39].

    Still having rewards and Q-value estimations for states and actions, but now there are two networks designated asθandθt, which are current network and target network, respectively.Errors are squared here to penalize massive errors far more severely than minor ones.To update the network, the backpropagation algorithm, which takes the derivative of all the values in the layers and modifies the related layers, might be utilized [38,39].

    ?

    (Continued)

    ?

    5 Proposed Algorithm

    A generic Deep Reinforcement Learning Route Selection (DRLRS) algorithm has been proposed to select thekneighbors that rely on the Deep Q-Learning algorithm to forward the query.Thus, every peer can locally determine other peer’s connections.Initially, every peer selects it’s neighbor arbitrarily.After each query is given within the network, the choice ofkneighbors has been made by the Q-value for the forthcoming query.In detail, each peer holds a Deep Q-Learning algorithm for every state and action related to its neighbor.

    In the proposed DRLRS algorithm, the inputs are the set of all neighborsN=n1,n2,...,nm,the search queryqtat timet, and the number of neighbours to be selectedk.The result is the set ofSwhich is relevant neighbors to the forward queryqt.On every iterationi=1,...,k, DRLRS depends on the Deep Q-Learning algorithm to select from the setN=n1,n2,...,nmthe neighbornjwith the highest expected probability of success according to Deep Q-Learning algorithm.The selected neighbornjat iterationiis than added to the set of relevant neighborsSand removed from the set of neighborsN.Algorithm 2 illustrates the proposed DRLRS algorithm.

    Algorithm 2: Deep reinforcement learning based route selection algorithm Input:qt: search query at time t N=n1,n2,...,nm: set of all neighbours k: number of neighbours to be selected for qt Output:S: set of neighbours to be selected S=?:

    (Continued)

    While i=1.......k do select neighbours nj using Deep Q-Learning algorithm S=S ∪nj N = N{nj}End

    6 Performance Evaluation

    To evaluate the performance of the proposed algorithm, first, a network simulator has been developed in python language using pyvis library, which is meant for quick generation of visual network graphs and Fast Network Simulation Setup (FNSS) toolchain to setup a network experiment scenario.Moreover, Apache STORM has been configured and used by system libraries for distributed real-time computations.Fig.2 illustrates the developed unstructured P2P network for simulation.An Apple MAC book pro with Intel Core i7 Quad-core 2.9 GHz CPU and 16.0 GB RAM was used to perform all experiments.A well-known document collections acquired from TREC has been used as a dataset that contains around 25,000 documents.In the experimental topology, 5000 queries were distributed among 1000 peers.The TTL of the query is set to 10,and the number of neighbors to be selected by the forwarder peer is set to 4.In this simulation,the system is considered in fix mode, where significant improvements are observed, and the issue of ‘churn’is not explicitly considered [41].

    Figure 2: Topological representation of developed unstructured P2P network

    In the developed network simulator, to experiment with Deep Q-Learning, it is necessary to create a custom environment.An environment in reinforcement learning must have four traits in general: (i) a state representation, (ii) an action representation, (iii) a reward function, and (iv)a time step function.The developed custom environment contains all the above four traits to select thekrelevant neighbors for the current query.An environment-class called “EnvDrlrs” was created with the help of the Open AI Gym toolkit, and the different functions were implemented from the base gym.Env class.

    6.1 Complexity Analysis

    The time complexity of the DRLRS algorithm was the same as that of Apache STORM,which isO(kT|E), wherekrepresents the number of training epochs andTrepresents the time.For space complexity, the DRLRS algorithm does not store any Q-table, so the space complexity could be calculated as |S||A|, whereSwas the explored network state, andA=4, a set of action selections for this experiment.The DRLRS algorithm does not occupy too much space, as it stores only variables.

    6.2 Efficiency Measures

    Alanazi and Yeferny (2019) proposed efficiency measures as retrieval effectiveness and the search cost of their study about query routing based on reinforcement learning in the P2P system.In this study, the same evaluation measures have been considered.The retrieval effectiveness of DRLRS and its competitors has been evaluated using a Recall metric.The recallR(qi)for the queryqiis defined as follows:

    where, RRR is number of relevant resources retrieved and RLR is the number of relevant resources.The cumulative average recall up tonqueriesCARnis defined as follows:

    For search cost evaluation, two metrics have been used:

    I.Connected Peers (CP): for queryqi,CP(qi)is the number of connected peers.The average number of connected peers for n sent queries (ACPn) is defined as:

    II.Average Overhead: This is measured by the number of exchanged messages for the queryqi.The average overhead up tonsent queries(AOn)is defined as:

    7 Results and Discussion

    This section discussed the results based on the retrieval effectiveness and search cost compression for a single peer executing 5000 queries.Furthermore, ten peers were selected randomly that executed 5000 different queries per peer, and the efficiency was measured in terms of number of average connected peers and the average time taken.

    7.1 Retrieval Effectiveness Compression

    Fig.3 represents the average recall of k-Random Walker, Directed BFS, and DRLRS.As shown in figure k-Random Walker has a low but stable recall of around 0.45.Whereas, initially,the recall of Directed BFS is low as k-Random Walker than it shows significant improvement after each calculation of approximation.Moreover, it has been observed that the recall of the proposed DRLRS algorithm rapidly increased after sending few queries.The average recall of k-Random Walker, Directed BFS, and DRLRS for all sent queries is 0.4, 0.85, and 1.28, respectively.These outcome shows that the balance between exploration and exploitation has been efficiently achieved by DRLRS, which relies on the Deep Q learning algorithm.Moreover, it has been observed that the routing performance of DRLRS improves continuously and, therefore, finishes the cold start phase in significant time.

    Figure 3: Evolution of average recall

    7.2 Search Cost Compression

    The search cost of the proposed routing algorithm DRLRS has been illustrated in Figs.4 and 5 with the comparison of k-Random Walker and Directed BFS.During the simulation period, the number of connected peers and average overhead per query of k-Random Walker were found to be high but constant at around 135 and 240, respectively.In addition, the number of connected peers and average overhead per peer query of Directed BFS initially increased, but after around half number of queries, it decreased.Indeed, the Directed BFS starts with Zero knowledge and relies on approximation function, which leads to achieving lower performance initially but increases the performance after some time.However, Directed BFS improved after each peer’s value approximation and shows the number of connected peers and average overhead around 123 and 224, respectively.Moreover, the search cost of the proposed DRLRS algorithm is going down rapidly and outperform k-Random Walker and Directed BFS.The connected peers and average overhead per query are 106 and 149, respectively, that shows the DRLRS algorithm locates more peers holding relevant resources for the query.

    Figure 4: Evolution of number of connected peers

    Figure 5: Evolution of average overhead

    Previous results have been obtained by one peer executing 5000 queries and selecting suitable neighbors for the desired resource.Further, ten peers have been selected randomly, and each peer executed 5000 different queries to evaluate the performance of DRLRS and it’s competitors.Fig.6 shows the average connected peers on each selected peer.Moreover, the average time taken in executing queries by each selected peer has been shown in Fig.7.The results shows that DRLRS outperforms it’s competitors and set fewer peers to find the requested resource in less time.

    Figure 6: Evolution of average connected peers for randomly selected peers

    Figure 7: Evolution of average time taken to search by randomly selected peers

    8 Conclusion

    In this paper, a novel approach has been proposed to route the query efficiently to find the pertinent resource.In this regard, the query routing problem has been described as a deep reinforcement learning problem, and a fully distributed strategy to solving it has been developed.Therefore, the Deep Q-Learning based query routing algorithm DRLRS is introduced to select the neighbors intelligently in order to improve the performance of the P2P network.Results shows that the DRLRS systematically learns from the previously sent queries and efficiently selects the best neighbors holding the relevant resource for the current query.The proposed algorithm improves the retrieval effectiveness and search cost continuously and outperforms the k-Random Walker and Directed BFS.However, in the future, the churn problem can be included in the study by taking into the account that at any time peers could join and leave the network that leads the frequent link changes between the peers, which may affect the performance of neighbors selection.

    Funding Statement:Authors would like to thank the Deanship of Scientific Research at Shaqra University for supporting this work under Project No.g01/n04.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    热re99久久精品国产66热6| 青春草亚洲视频在线观看| 交换朋友夫妻互换小说| 大片电影免费在线观看免费| 国产男人的电影天堂91| 午夜福利影视在线免费观看| 热re99久久国产66热| 中文字幕高清在线视频| 99国产极品粉嫩在线观看| 国产真人三级小视频在线观看| 国产精品久久久av美女十八| 亚洲国产看品久久| 中文字幕另类日韩欧美亚洲嫩草| 男女下面插进去视频免费观看| 亚洲精华国产精华精| 成年女人毛片免费观看观看9 | 国产免费现黄频在线看| 视频区图区小说| 色视频在线一区二区三区| 国产精品99久久99久久久不卡| 亚洲成人免费av在线播放| 国产一区二区激情短视频 | 亚洲全国av大片| 男女床上黄色一级片免费看| 中文欧美无线码| 日本猛色少妇xxxxx猛交久久| 天天躁日日躁夜夜躁夜夜| 久久人人97超碰香蕉20202| 中文字幕人妻丝袜一区二区| 最黄视频免费看| av天堂久久9| 水蜜桃什么品种好| 精品国产一区二区三区四区第35| 欧美在线一区亚洲| 男女高潮啪啪啪动态图| 建设人人有责人人尽责人人享有的| 一区二区日韩欧美中文字幕| 午夜福利在线观看吧| 亚洲综合色网址| 午夜91福利影院| 好男人电影高清在线观看| 日本黄色日本黄色录像| 后天国语完整版免费观看| 亚洲情色 制服丝袜| 侵犯人妻中文字幕一二三四区| 美女中出高潮动态图| 51午夜福利影视在线观看| 在线av久久热| 下体分泌物呈黄色| 黄片小视频在线播放| 18在线观看网站| 男女高潮啪啪啪动态图| 丝袜在线中文字幕| 一个人免费在线观看的高清视频 | 亚洲精品成人av观看孕妇| 午夜福利乱码中文字幕| a在线观看视频网站| 久久久久精品国产欧美久久久 | 成年美女黄网站色视频大全免费| 成人三级做爰电影| 久久久久国产精品人妻一区二区| 深夜精品福利| 成年人黄色毛片网站| 两个人看的免费小视频| 12—13女人毛片做爰片一| 亚洲成av片中文字幕在线观看| 久久免费观看电影| 搡老乐熟女国产| 秋霞在线观看毛片| 亚洲久久久国产精品| 欧美精品一区二区免费开放| 国产免费视频播放在线视频| 国产亚洲av片在线观看秒播厂| 日韩视频一区二区在线观看| 自拍欧美九色日韩亚洲蝌蚪91| 我要看黄色一级片免费的| 免费少妇av软件| 在线看a的网站| 亚洲综合色网址| 曰老女人黄片| 亚洲av美国av| 精品国产一区二区久久| 一本久久精品| av在线老鸭窝| 欧美精品一区二区大全| 精品第一国产精品| 嫩草影视91久久| 91成年电影在线观看| 极品人妻少妇av视频| 王馨瑶露胸无遮挡在线观看| 极品少妇高潮喷水抽搐| 亚洲av欧美aⅴ国产| 亚洲欧洲日产国产| 不卡一级毛片| 天天添夜夜摸| 美女扒开内裤让男人捅视频| 热re99久久精品国产66热6| 国产欧美日韩精品亚洲av| 美女主播在线视频| 欧美国产精品va在线观看不卡| 啦啦啦免费观看视频1| cao死你这个sao货| 亚洲国产精品一区三区| 黑人巨大精品欧美一区二区蜜桃| 国产免费福利视频在线观看| 成人18禁高潮啪啪吃奶动态图| 亚洲三区欧美一区| 91成年电影在线观看| 国产精品亚洲av一区麻豆| 亚洲va日本ⅴa欧美va伊人久久 | 欧美日韩亚洲国产一区二区在线观看 | 精品乱码久久久久久99久播| tocl精华| 搡老熟女国产l中国老女人| 亚洲国产毛片av蜜桃av| 嫁个100分男人电影在线观看| 人人澡人人妻人| 狠狠精品人妻久久久久久综合| 久久久水蜜桃国产精品网| 久久精品久久久久久噜噜老黄| 咕卡用的链子| 91麻豆精品激情在线观看国产 | 精品亚洲乱码少妇综合久久| 永久免费av网站大全| 91麻豆av在线| 午夜两性在线视频| 国产精品av久久久久免费| 色综合欧美亚洲国产小说| 水蜜桃什么品种好| 午夜福利免费观看在线| 多毛熟女@视频| 999久久久精品免费观看国产| 亚洲精品粉嫩美女一区| 婷婷丁香在线五月| 一级a爱视频在线免费观看| 精品乱码久久久久久99久播| 成人国语在线视频| 激情视频va一区二区三区| 日韩三级视频一区二区三区| 中文精品一卡2卡3卡4更新| 水蜜桃什么品种好| 女人久久www免费人成看片| 亚洲av男天堂| 国产精品国产三级国产专区5o| 亚洲精品美女久久av网站| 操美女的视频在线观看| 9热在线视频观看99| 免费在线观看完整版高清| 国产国语露脸激情在线看| av电影中文网址| 国产精品香港三级国产av潘金莲| 99精品久久久久人妻精品| 欧美乱码精品一区二区三区| 一本一本久久a久久精品综合妖精| 操出白浆在线播放| 老司机在亚洲福利影院| 成人免费观看视频高清| 精品一区在线观看国产| 日本五十路高清| 母亲3免费完整高清在线观看| 91大片在线观看| 中文字幕制服av| 亚洲性夜色夜夜综合| 国产精品免费大片| 天天躁夜夜躁狠狠躁躁| 亚洲精品自拍成人| 在线观看舔阴道视频| av欧美777| 久久九九热精品免费| 国产免费视频播放在线视频| 国产又色又爽无遮挡免| 一本色道久久久久久精品综合| 亚洲精品一卡2卡三卡4卡5卡 | 亚洲精品国产色婷婷电影| 欧美 日韩 精品 国产| 中文字幕另类日韩欧美亚洲嫩草| 国产日韩一区二区三区精品不卡| 亚洲伊人色综图| 天天操日日干夜夜撸| 高潮久久久久久久久久久不卡| 国产在线一区二区三区精| 国产成人精品久久二区二区91| 亚洲国产精品一区三区| 日本五十路高清| 色精品久久人妻99蜜桃| 一本大道久久a久久精品| 少妇被粗大的猛进出69影院| 国产一区二区在线观看av| a级毛片黄视频| 日本精品一区二区三区蜜桃| 99热网站在线观看| 999久久久精品免费观看国产| 丝袜美腿诱惑在线| 国产精品一区二区免费欧美 | www.熟女人妻精品国产| 午夜免费成人在线视频| 国产精品成人在线| 欧美 亚洲 国产 日韩一| 国产又色又爽无遮挡免| 老司机午夜十八禁免费视频| 精品国产超薄肉色丝袜足j| 成在线人永久免费视频| av一本久久久久| 亚洲全国av大片| av网站免费在线观看视频| 日本av手机在线免费观看| tube8黄色片| 成人手机av| 丰满少妇做爰视频| 国产一卡二卡三卡精品| 一本色道久久久久久精品综合| 丰满饥渴人妻一区二区三| 久久国产精品大桥未久av| 免费av中文字幕在线| 91字幕亚洲| 欧美精品一区二区大全| 亚洲精品国产av成人精品| 国产一卡二卡三卡精品| 成人国产av品久久久| 国产日韩欧美在线精品| 免费在线观看黄色视频的| 在线 av 中文字幕| 午夜福利免费观看在线| 精品国产乱子伦一区二区三区 | 国产日韩一区二区三区精品不卡| 亚洲国产看品久久| 国产区一区二久久| 啦啦啦啦在线视频资源| 老鸭窝网址在线观看| 亚洲精品av麻豆狂野| 日韩精品免费视频一区二区三区| 夜夜夜夜夜久久久久| 韩国高清视频一区二区三区| 电影成人av| 亚洲中文日韩欧美视频| 亚洲久久久国产精品| av网站在线播放免费| 久久人人97超碰香蕉20202| 亚洲精品自拍成人| 美女国产高潮福利片在线看| 欧美激情高清一区二区三区| 50天的宝宝边吃奶边哭怎么回事| 国产主播在线观看一区二区| 最新在线观看一区二区三区| 韩国精品一区二区三区| 日本av免费视频播放| 欧美性长视频在线观看| 午夜福利在线免费观看网站| 精品国产乱子伦一区二区三区 | 啦啦啦中文免费视频观看日本| 男人添女人高潮全过程视频| 免费在线观看黄色视频的| 少妇 在线观看| 天天添夜夜摸| 99国产精品99久久久久| 在线精品无人区一区二区三| www.熟女人妻精品国产| av欧美777| 日韩一卡2卡3卡4卡2021年| 国产精品一二三区在线看| 国产亚洲精品久久久久5区| 一级毛片精品| 亚洲男人天堂网一区| 在线十欧美十亚洲十日本专区| 啦啦啦 在线观看视频| 欧美另类一区| 精品视频人人做人人爽| 亚洲精品乱久久久久久| 久久精品国产a三级三级三级| 嫁个100分男人电影在线观看| 不卡av一区二区三区| 欧美激情极品国产一区二区三区| 亚洲av电影在线观看一区二区三区| www.自偷自拍.com| 波多野结衣一区麻豆| 久热爱精品视频在线9| 欧美日韩亚洲综合一区二区三区_| 亚洲情色 制服丝袜| 欧美97在线视频| 极品人妻少妇av视频| 亚洲精品国产区一区二| a级毛片在线看网站| 午夜福利影视在线免费观看| 成在线人永久免费视频| 99久久99久久久精品蜜桃| 精品乱码久久久久久99久播| 国产精品一区二区在线不卡| av欧美777| 成人国语在线视频| 自线自在国产av| 十八禁人妻一区二区| 亚洲国产精品成人久久小说| 国产精品av久久久久免费| 一区二区三区四区激情视频| kizo精华| 欧美另类一区| 考比视频在线观看| 亚洲精品国产一区二区精华液| av线在线观看网站| 亚洲欧美成人综合另类久久久| 妹子高潮喷水视频| 黄片大片在线免费观看| 亚洲欧美激情在线| 亚洲精品自拍成人| 少妇 在线观看| 99国产精品一区二区蜜桃av | 亚洲色图综合在线观看| 日本黄色日本黄色录像| 久久影院123| 精品一区二区三卡| 自拍欧美九色日韩亚洲蝌蚪91| 97在线人人人人妻| 一进一出抽搐动态| 午夜视频精品福利| 一区在线观看完整版| 久久国产精品影院| 18在线观看网站| 欧美日韩精品网址| 国产成人精品久久二区二区91| 日韩免费高清中文字幕av| 亚洲中文av在线| 亚洲欧美日韩高清在线视频 | 国产黄色免费在线视频| 亚洲欧美日韩另类电影网站| 亚洲av日韩在线播放| 欧美日韩国产mv在线观看视频| 欧美激情极品国产一区二区三区| 免费一级毛片在线播放高清视频 | 99精品久久久久人妻精品| 精品国产超薄肉色丝袜足j| 久久精品成人免费网站| av免费在线观看网站| 亚洲欧美一区二区三区黑人| 国产成人精品久久二区二区免费| 亚洲中文av在线| 国产99久久九九免费精品| 久久亚洲国产成人精品v| 伦理电影免费视频| 久久亚洲国产成人精品v| av一本久久久久| 国产主播在线观看一区二区| 久久久精品94久久精品| 日韩精品免费视频一区二区三区| 中文字幕色久视频| 在线观看一区二区三区激情| 日韩中文字幕视频在线看片| 欧美 日韩 精品 国产| 国产精品 国内视频| 国产精品熟女久久久久浪| 俄罗斯特黄特色一大片| 美国免费a级毛片| 黑丝袜美女国产一区| 午夜日韩欧美国产| 亚洲人成77777在线视频| 免费一级毛片在线播放高清视频 | 别揉我奶头~嗯~啊~动态视频 | 桃花免费在线播放| 亚洲伊人色综图| 亚洲国产av新网站| 久久久国产欧美日韩av| 亚洲精品乱久久久久久| 精品国产一区二区三区四区第35| 亚洲欧洲日产国产| 1024香蕉在线观看| 亚洲欧美日韩高清在线视频 | 国产av又大| 亚洲欧美日韩另类电影网站| 亚洲av国产av综合av卡| 啦啦啦啦在线视频资源| 99精品久久久久人妻精品| 国产高清视频在线播放一区 | 国产日韩欧美在线精品| 亚洲精品国产一区二区精华液| 精品视频人人做人人爽| 久久久久精品国产欧美久久久 | 成人三级做爰电影| 精品人妻在线不人妻| 欧美另类亚洲清纯唯美| 日韩欧美一区视频在线观看| 欧美黑人精品巨大| 各种免费的搞黄视频| 丁香六月天网| 日韩中文字幕视频在线看片| 一区福利在线观看| 一级毛片电影观看| 亚洲第一青青草原| 中国国产av一级| 国产伦理片在线播放av一区| 人人妻人人添人人爽欧美一区卜| 亚洲精品美女久久久久99蜜臀| 久久99热这里只频精品6学生| 精品少妇黑人巨大在线播放| 亚洲色图综合在线观看| 免费高清在线观看日韩| 国产在视频线精品| 久久99一区二区三区| www日本在线高清视频| 男女免费视频国产| 欧美精品亚洲一区二区| 汤姆久久久久久久影院中文字幕| 久久久久久免费高清国产稀缺| 亚洲天堂av无毛| 丝袜在线中文字幕| 极品人妻少妇av视频| videos熟女内射| 日本vs欧美在线观看视频| 十八禁高潮呻吟视频| 菩萨蛮人人尽说江南好唐韦庄| 国产精品影院久久| 亚洲国产欧美在线一区| 一区二区日韩欧美中文字幕| 久久人人爽av亚洲精品天堂| 国产欧美亚洲国产| 99久久99久久久精品蜜桃| 国内毛片毛片毛片毛片毛片| 在线 av 中文字幕| 汤姆久久久久久久影院中文字幕| 飞空精品影院首页| 操美女的视频在线观看| 久久狼人影院| 成人影院久久| 纯流量卡能插随身wifi吗| 黄色片一级片一级黄色片| 老熟女久久久| 色综合欧美亚洲国产小说| 久久久精品免费免费高清| 久久影院123| 老汉色av国产亚洲站长工具| 久久精品国产综合久久久| 国产有黄有色有爽视频| 国产精品一区二区精品视频观看| 欧美精品一区二区免费开放| 日韩制服丝袜自拍偷拍| 亚洲欧美成人综合另类久久久| 韩国高清视频一区二区三区| 人成视频在线观看免费观看| 国产精品1区2区在线观看. | 亚洲精品中文字幕在线视频| 日韩一卡2卡3卡4卡2021年| 欧美精品高潮呻吟av久久| 欧美日韩亚洲国产一区二区在线观看 | 久久影院123| 国产在线视频一区二区| 美国免费a级毛片| 99国产精品一区二区蜜桃av | a级毛片黄视频| 日韩大码丰满熟妇| 欧美亚洲日本最大视频资源| 少妇人妻久久综合中文| 美女午夜性视频免费| www.精华液| 蜜桃在线观看..| 精品一区二区三区av网在线观看 | 久久99一区二区三区| 亚洲精品第二区| 久久久久久久久久久久大奶| 欧美人与性动交α欧美软件| 一级片免费观看大全| 妹子高潮喷水视频| 日本撒尿小便嘘嘘汇集6| 国产在线观看jvid| 多毛熟女@视频| www.999成人在线观看| 午夜日韩欧美国产| 交换朋友夫妻互换小说| 欧美亚洲日本最大视频资源| 成人国产av品久久久| 久久精品人人爽人人爽视色| 在线 av 中文字幕| 如日韩欧美国产精品一区二区三区| 亚洲av成人不卡在线观看播放网 | 亚洲国产精品999| 啦啦啦中文免费视频观看日本| 国产精品av久久久久免费| 欧美激情久久久久久爽电影 | 国产高清videossex| av欧美777| 91老司机精品| 国产精品自产拍在线观看55亚洲 | 热re99久久国产66热| 满18在线观看网站| 婷婷色av中文字幕| 成人国语在线视频| 久久亚洲国产成人精品v| 男男h啪啪无遮挡| 最近最新中文字幕大全免费视频| 狂野欧美激情性bbbbbb| 国产av又大| 国产野战对白在线观看| 免费av中文字幕在线| 亚洲国产日韩一区二区| 欧美 亚洲 国产 日韩一| 最新的欧美精品一区二区| 伊人久久大香线蕉亚洲五| 各种免费的搞黄视频| av视频免费观看在线观看| 一级毛片女人18水好多| 肉色欧美久久久久久久蜜桃| 国产精品一二三区在线看| 国产一区二区在线观看av| 午夜福利乱码中文字幕| 一级毛片女人18水好多| 亚洲欧美一区二区三区黑人| 日韩 亚洲 欧美在线| 成人黄色视频免费在线看| 91成人精品电影| 日本猛色少妇xxxxx猛交久久| 人人妻人人爽人人添夜夜欢视频| av电影中文网址| 亚洲精品久久久久久婷婷小说| 飞空精品影院首页| 色综合欧美亚洲国产小说| 黄片小视频在线播放| 在线观看舔阴道视频| 美女大奶头黄色视频| 丝袜在线中文字幕| 自拍欧美九色日韩亚洲蝌蚪91| 日本精品一区二区三区蜜桃| 国产精品久久久久成人av| 大香蕉久久成人网| 黑人巨大精品欧美一区二区蜜桃| 夫妻午夜视频| 亚洲中文av在线| 老司机在亚洲福利影院| 国产国语露脸激情在线看| h视频一区二区三区| 日日爽夜夜爽网站| 亚洲av日韩在线播放| 丰满饥渴人妻一区二区三| 女人精品久久久久毛片| 亚洲av美国av| 超碰成人久久| 免费观看a级毛片全部| 久久国产精品男人的天堂亚洲| 9色porny在线观看| 中文字幕另类日韩欧美亚洲嫩草| 国产精品国产三级国产专区5o| 精品久久久久久电影网| 久久亚洲精品不卡| 免费一级毛片在线播放高清视频 | 熟女少妇亚洲综合色aaa.| 老汉色av国产亚洲站长工具| 欧美激情久久久久久爽电影 | 真人做人爱边吃奶动态| 欧美在线一区亚洲| 成年av动漫网址| 夫妻午夜视频| 久久国产精品大桥未久av| 国产日韩欧美视频二区| 18禁裸乳无遮挡动漫免费视频| 欧美黑人欧美精品刺激| 又黄又粗又硬又大视频| 国产色视频综合| 天天影视国产精品| 日本av手机在线免费观看| 久久99一区二区三区| 纯流量卡能插随身wifi吗| 亚洲欧洲日产国产| 精品久久久久久久毛片微露脸 | 男女下面插进去视频免费观看| 老熟妇仑乱视频hdxx| 国产亚洲欧美精品永久| 岛国毛片在线播放| 下体分泌物呈黄色| 亚洲国产av新网站| 亚洲,欧美精品.| 人妻人人澡人人爽人人| 亚洲五月婷婷丁香| tube8黄色片| 啦啦啦中文免费视频观看日本| 啪啪无遮挡十八禁网站| 99国产极品粉嫩在线观看| 欧美精品av麻豆av| 狠狠婷婷综合久久久久久88av| 老汉色∧v一级毛片| 美女扒开内裤让男人捅视频| 人人妻人人澡人人爽人人夜夜| 一边摸一边做爽爽视频免费| 爱豆传媒免费全集在线观看| 日韩,欧美,国产一区二区三区| 水蜜桃什么品种好| 桃花免费在线播放| 老汉色av国产亚洲站长工具| 国产精品99久久99久久久不卡| 日本vs欧美在线观看视频| 一本一本久久a久久精品综合妖精| www日本在线高清视频| 国产国语露脸激情在线看| 亚洲精品一二三| 夜夜骑夜夜射夜夜干| 美女主播在线视频| 久久毛片免费看一区二区三区| 久久精品国产综合久久久| 香蕉国产在线看| 男人添女人高潮全过程视频| 色播在线永久视频| 欧美精品一区二区免费开放| 成年动漫av网址| 精品国产国语对白av| 欧美成狂野欧美在线观看| 欧美在线黄色| 99久久综合免费| 亚洲人成电影观看| 搡老岳熟女国产| 精品高清国产在线一区| 久久性视频一级片| 精品一品国产午夜福利视频| 咕卡用的链子| 99香蕉大伊视频| 欧美变态另类bdsm刘玥| 精品久久久精品久久久| 欧美人与性动交α欧美精品济南到| 一区福利在线观看| 精品国内亚洲2022精品成人 | 亚洲第一青青草原|