• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An RGB-D Camera Based Visual Positioning System for Assistive Navigation by a Robotic Navigation Aid

    2021-07-26 07:23:06HeZhangLingqiuJinandCangYe
    IEEE/CAA Journal of Automatica Sinica 2021年8期
    關(guān)鍵詞:稠油油井規(guī)律

    He Zhang, Lingqiu Jin, and Cang Ye,

    Abstract—There are about 253 million people with visual impairment worldwide. Many of them use a white cane and/or a guide dog as the mobility tool for daily travel. Despite decades of efforts, electronic navigation aid that can replace white cane is still research in progress. In this paper, we propose an RGB-D camera based visual positioning system (VPS) for real-time localization of a robotic navigation aid (RNA) in an architectural floor plan for assistive navigation. The core of the system is the combination of a new 6-DOF depth-enhanced visual-inertial odometry (DVIO) method and a particle filter localization (PFL)method. DVIO estimates RNA’s pose by using the data from an RGB-D camera and an inertial measurement unit (IMU). It extracts the floor plane from the camera’s depth data and tightly couples the floor plane, the visual features (with and without depth data), and the IMU’s inertial data in a graph optimization framework to estimate the device’s 6-DOF pose. Due to the use of the floor plane and depth data from the RGB-D camera, DVIO has a better pose estimation accuracy than the conventional VIO method. To reduce the accumulated pose error of DVIO for navigation in a large indoor space, we developed the PFL method to locate RNA in the floor plan. PFL leverages geometric information of the architectural CAD drawing of an indoor space to further reduce the error of the DVIO-estimated pose. Based on VPS, an assistive navigation system is developed for the RNA prototype to assist a visually impaired person in navigating a large indoor space. Experimental results demonstrate that: 1)DVIO method achieves better pose estimation accuracy than the state-of-the-art VIO method and performs real-time pose estimation (18 Hz pose update rate) on a UP Board computer; 2)PFL reduces the DVIO-accrued pose error by 82.5% on average and allows for accurate wayfinding (endpoint position error ≤45 cm) in large indoor spaces.

    I. INTRODUCTION

    ACCORDING to Lancet Global Health [1], there are about 253 million people with visual impairment, of which 36 million are blind. Since age-related diseases (glaucoma,macular degeneration, diabetes, etc.) are the leading cause of vision loss and the world population is rapidly aging, more and more people become blind or visually impaired (BVI).Therefore, there is a crucial need for developing navigation aids to help the BVI with their daily mobility and independent lives. The problem of independent mobility of a BVI individual includes wayfinding and obstacle avoidance.Wayfinding is a global problem of planning and following a path towards the destination while obstacle avoidance is a local problem of taking steps without colliding, tripping, or falling.

    To provide wayfinding and obstacle avoidance functions to a BVI traveler at the same time, the location of the traveler and the 3D map of the surroundings must be accurately acquired. The technique to address the problem is called simultaneous localization and mapping (SLAM). In the literature, several robotic navigation aids (RNAs) [2]–[8] that use SLAM [9] for assistive navigation of the blind have been introduced. The performance of these RNAs relies on the pose estimation accuracy since the pose information is used to build a 3D map of the environment, locate the blind traveler in the environment, and guide the traveler to the destination.Stereo camera [2], [3] and RGB-D camera [4], [7] based visual SLAM methods have been developed to estimate the pose of RNA and detect surrounding obstacles from the generated 3D point cloud map. To improve the pose estimation accuracy, geometric features [5], [6], and inertial data [7], [8] have been utilized in existing SLAM methods.However, the pose error can still accrue over time and may become large enough (in case of a long trajectory) to break down RNA’s navigation function. To tackle this problem,visual maps [10], [11], Bluetooth low energy beacons [12],radio-frequency identifications [13], and near field communication tags [14] have been employed to correct accumulative pose estimation error. However, building a visual map ahead of time can be time-consuming as it requires extraction and storage of visual features point-by-point in a certain spatial interval to cover the entire navigational space, while the approach of using beacons or the alike requires re-engineering the environment and is thus not practical for assistive navigation.

    To address these disadvantages, we propose, in this paper, a vision position system (VPS) that uses an RGB-D (i.e., colordepth) camera and an inertial measurement unit (IMU) to estimate the pose of an RNA for wayfinding applications. The system uses the floor plan (i.e., architectural CAD drawing) of an indoor space to reduce the accumulative pose estimation errors by a 2-step scheme. First, a new visual-inertial odometry (VIO) method is used to estimate 6-DOF RNA poses along the path. Second, a 3D point cloud map (local map) is built (by using the estimated poses) and projected onto the floor plane to create a 2D map, which is then aligned with the floor plan by a particle filter localization (PFL) method(i.e., the 2D geometric features such as walls, doors, corners,and junctions of the two maps are aligned) to reduce RNA position and heading errors on the floor plan for wayfinding.

    The RNA prototype uses a sensor suite consisting of an RGB-D camera and an IMU for localization, making the device an RGB-D-camera-based visual-inertial system (RGBD VINS). A VINS employs a SLAM technique, known as VIO, to estimate the system’s motion variables by jointly using its visual-inertial data. In [15], three state-of-the-art VIO methods, namely, OKVIS [16], VINS-Mono [17], and VIORB[18], are compared in the context of RNA pose estimation.The results show that VINS-Mono outperforms the other two.However, to enable real-time computation of VINS-Mono on the UP Board computer used by the RNA, some modifications, such as using a constant inverse depth for each visual feature in the iterative optimization process, and reducing the size of the sliding window, must be made to the algorithm/implementation. These modifications, however,trade the method’s pose estimation accuracy for computational efficiency. To address the problem, we propose a socalled depth-enhanced visual-inertial odometry (DVIO) to estimate the RNA’s pose for assistive navigation. DVIO is developed based on the framework of VINS-Mono and it improves VINS-Mono’s pose estimation accuracy by 1) using the geometric feature (the floor plane extracted from the camera’s depth data) to create additional constraints between the graph nodes to reduce the accumulative pose error; 2)using the depth data from the RGB-D camera for visual feature initialization and update to avoid iterative computation of the visual features’ inverse depth in each step of the optimization process. Unlike VINS-Mono, DVIO does not need to estimate the metric scale, which is known from the depth data. As a result, it is free of pose estimation error induced by inaccurate scale estimation. Based on the DVIOestimated egomotion, a PFL method is employed to determine the RNA’s pose (3-DOF pose including position and heading)on the floor plan of the navigational space for wayfinding.The main contributions of this paper include:

    1) We propose a new VIO method, called DVIO, to estimate the 6-DOF pose of RNA. The method achieves better accuracy in pose estimation by using the depth data from an RGB-D camera.

    2) We introduce a VPS to estimate the RNA’s 3-DOF pose on the floor plan for wayfinding. VPS employs PFL to estimate the pose based on the DVIO-estimated egomotion.PFL helps to improve pose estimation accuracy.

    3) We develop an assistive navigation system based on VPS and validate its efficacy by experiments with the RNA prototype in the real world.

    II. RELATED WORk

    A. Related Work in VIO

    Existing VIO methods can be classified into two categories,namely loosely-coupled [19]–[21] and tightly-coupled [22],[16]–[18]. In this section, we provide an overview of the tightly-coupled methods as the proposed DVIO falls into the same category. MSCKF [22] is an extended Kalman filter(EKF) based visual-inertial SLAM method. It utilizes IMU measurements to predict the filter state and employs visual feature measurements to update the state vector. Unlike a traditional EKF, it simultaneously updates multiple camera poses (in the state vector) by using a novel measurement model for the visual features. This model estimates a visual feature’s 3D location by using its multi-view geometric constraint, computes the feature’s reprojection residuals on multiple images, and use them as innovation to update the state vector. The method adopts adelayedstate update strategy, i.e., a tracked visual feature is used to update the state vector only when it is no longer detected, to get the most from the multi-view constraint. In so doing, it uses much fewer visual features for state estimation as those features that are currently tracked are not used.

    On the contrary, the smoothing-based VIO methods[16]–[18] use all visual measurements of the related keyframes to estimate the current motion state and may achieve a more accurate result. OKVIS [16] is a smoothingbased VIO method that performs a nonlinear optimization by using a cost function consisting of the sensor measurements at several keyframes. Specifically, the cost function is formulated as the weighted sum of the residuals of the visual features’ reprojections and the inertial measurements. This formulation incorporates all visual features’ measurements and leads to better pose estimation accuracy than that of MSCKF [16]. OKVIS performs well on a stereo VINS.However, its performance may significantly degrade in the case of a monocular VINS. This is because it lacks a reliable approach to accurately estimating the initial values of the state variables (e.g., gyroscope bias, metric scale). Due to the nonconvexity of the cost function, a poor estimate of the initial state will likely cause the optimization process to be stuck at a local minimum and result in an incorrect pose estimation. To mitigate this issue, VIORB [18] implements a sophisticated sensor fusion procedure to bootstrap a monocular VINS with a more accurate estimate for the initial state, consisting of the pose, velocity, 3D feature locations, gravity vector, metric scale, gyroscope bias, and accelerometer bias. However,VIORB requires 15 seconds of visual-inertial data to obtain an accurate result. It is not suitable for our case that requires a scale estimation right at the beginning. ORB-SLAM3 [23]improves the initialization approach by using an inertial-only maximum-a-posterior (MAP) estimation step [24] to compute the values for the scale, velocities, gravity, and IMU biases.This step takes into account the IMU’s measurement uncertainty and the gravity magnitude in producing an estimate that is accurate enough for a joint visual-inertial bundle adjustment (BA). The output of the inertial-only MAP is used to initialize the values of the VINS’ state parameters to speed up the convergence of the visual-inertial BA. The approach allows the VINS to initialize itself in less than 4 seconds.

    Qin and Shen [25] discover that the metric scale error is linearly dependent upon the accelerometer bias and simultaneously estimating the scale and the accelerometer bias requires a long duration of sensor data collection. To overcome the problem, they propose the VINS-Mono [17]method, where the initialization process is simplified by ignoring the accelerometer bias. The method uses a two-step approach to initializing the VINS’ motion state. First, it builds a scale-dependent 3D structure by a visual-only structurefrom-motion method. Second, it aligns the IMU integration with the visual-only structure to recover the scale, gravity,velocity, and gyroscope bias. This initialization approach converges much faster (in ~100 ms) with negligible accuracy loss [25]. However, VINS-Mono [17] still falls short of realtime computing performance on a computer with limited computing power [26]. Using a smaller sliding window may speed up the computation. But it may result in unwanted loss of accuracy.

    Research efforts for the above state-of-the-art VIO methods have been mainly focused on monocular VINS [17], [18],[22], or stereo VINS [16], [27]. RGB-D-camera-based VIO is a less-explored area. In [28], an EKF based VIO method is introduced for pose estimation of an RGB-D VINS. The method uses the egomotion estimated by IMU preintegration to generate state prediction and treats the pose estimated by using the visual-depth data as the observation to compute the state update. Linget al.present a smoothing-based VIO method [29] for RGB-D VINS. The method determines the metric scale from the depth data and obtains the VINS’ initial motion state by simply aligning the visual-based pose with the IMU-preintegration-based pose. While it uses the standard VIO framework to estimate the VINS’ motion state, the visual features’ inverse depths are initialized by using the camera’s depth data and are kept as constant values in the optimization process. Shanet al.[30] proposed VINS-RGBD, a smoothingbased VIO method that exploits the depth information in the framework of VINS-Mono [17]. In the initialization process, it uses corner points [31] and employs a 3D-2D-PnP [32]method to build the visual structure. After the initialization, it estimates the VINS’ motion state and the inverse depths of the tracked visual features through a nonlinear optimization process. If a feature’s depth is provided by the RGB-D camera, the inverse depth value is treated as a constant.Otherwise, the depth value is estimated by triangulation, and the inverse depth value is iteratively updated in the later optimization process. The triangulated depths for far-range visual features are not accurate, and the depth estimation error can reduce the pose estimation accuracy during optimization.Instead, we avoid depth estimation for the far-range visual features, and we utilize the epipolar constraint to model their measurement residuals in the optimization step for pose estimation. Also, we exploit the geometric feature (the floor plane extracted from depth data) to reduce the accumulated pose estimation error. The proposed DVIO method improves the above smoothing-based VIO methods by incorporating visual features without depth and geometric feature into the graph for more accurate pose estimation. It achieves real-time computation (~18 Hz) with decent accuracy on a UP Board computer.

    B. Related Work in Localization

    As an incremental state estimation method, VIO accrues pose errors over time. When using VIO for navigation in a large space, a loop closure can be used to eliminate the accumulated pose error. However, if a loop closure cannot be detected or it is not detected in a timely fashion, the accrued pose error may become big enough to make the navigation system malfunction. To address the problem, the floor plans of the operating environments have been used to reduce accumulative pose errors in the robotics community. Boniardiet al.[33], [34] introduce a pose-graph method to track the 3-DOF pose of a robot in a floor plan (CAD drawing) by using a 2D LiDAR. A scan-to-map matching is first performed to align the LiDAR scans with the floor plan and determine the robot’s pose with respect to the floor plan. Then, this relative pose measurement is used to create additional edges (between the related nodes), which serve as prior constraints in the graph structure to incorporate the floor plan into the graph.The use of the floor plan in the graph SLAM process helps to reduce the accrued pose error. The method has a 50-ms runtime on an Intel Core i7 CPU (8-core, 4.0 GHz). It is difficult to achieve real-time computation on an Up Board computer (with a 4-core 1.92 GHz Intel ATOM CPU). In [35],Watanabeet al.present a method to localize a robot in indoor space by using an architectural floor plan and depth data of an RGB-D camera. The method first extracts a number of planes from the depth image at the robot’s current pose and projects the 3D points belonging to the planes onto the floor to produce a 2D source point cloud. It then uses a ray-tracing algorithm to generate a simulated 2D target point cloud from the floor plan. Finally, the robot pose (with respect to the floor plan) is determined by aligning the source point cloud with the target point cloud using the GICP algorithm [36]. However,the method can malfunction when the GICP algorithm is stuck to a local minimum or the scene is not geometrically featurerich.

    (6)在保持油井正常生產(chǎn)的情況下,繼續(xù)摸索稠油井出砂、排砂規(guī)律,制定和優(yōu)化工作制度,為今后的工作方向打下基礎(chǔ)。

    The multimodel PFL [37] based method is more robust for pose tracking. Winterhalteret al.employ a 6-DOF PFL approach [38] to track the camera pose for a Google Tango tablet in an indoor environment by using the data from the device’s RGB-D camera and IMU. The method utilizes the VIO-estimated motion to predict the pose for each particle. It computes an importance weight for each particle, which is proportional to the observation likelihood of the measurement given the particle’s state. The likelihood value is estimated by comparing the actual depth data with the expected depth data(from the floor plan) given the predicted pose. A particle survives with a probability proportional to its importance weight in the re-sampling step. To reliably track the device’s 6-DOF pose, 5000 particles are used. This results in a high computational cost. To achieve real-time computation, the PFL algorithm must run on a backend server. In this work, we simplify the method and employ a 3-DOF PFL method to estimate the RNA’s position and orientation on a 2D floor plan for real-time assistive navigation. Our method uses only 100 particles for pose tracking, resulting in real-time computation (~50-ms runtime) on an Up Board computer. The proposed method creates a local submap by registering several frames of depth data (instead of using just one frame of depth data [38]) and aligns this map with the floor plan to determine the device pose with respect to the floor plan. The multi-frame local submap is less likely to be geometrically featureless,making our method more robust to depth data noise.

    III. RNA PROTOTYPE AND NOTATIONS

    As depicted in Fig. 1, the RNA prototype uses an Intel Realsense D435 (RGB-D) Camera and an IMU (VN100 of VectorNav Technologies, LLC) for motion estimation. The D435 consists of a color camera that produces a color image of the scene and an IR stereo camera that generates the corresponding depth data. Their resolutions are set to 424×240 to produce a 20 fps data stream to the UP Board computer[39]. The D435 is mounted on the cane with a 25° tilt-up angle to keep the cane’s body out of the camera’s field of view. The VN100 is set to output the inertial data at 100 Hz.The prototype uses a mechanism called active rolling tip(ART) [40] to steer the cane to the desired direction of travel to guide the user. The ART consists of a rolling tip, a gear motor (with a built-in encoder), a motor drive, and a clutch. A custom control board is used to engage and disengage the clutch. When the clutch is engaged, RNA enters the robotcane mode and the motor drives the rolling tip and steers the cane into the desired direction. The slippage at the rolling tip is detected by comparing the encoder and gyro data. If the slippage is above a threshold, RNA switches itself into the white-cane mode temporarily until the slippage drops below the threshold. The details on this human-intent detection scheme for automatic mode switching are referred to [5].When the ART is disengaged, the rolling tip is disconnected from the gear motor, turning RNA into the white-cane mode,and the user can swing the RNA just like using a white cane.In this case, a coin vibrator (on the grip) vibrates to indicate the desired direction. The user can switch between the two modes by pressing a push button on the grip. The clutch controller and the motor drive are controlled by the Up Board via its general IO port and RS-232 port, respectively.

    Fig. 1. Top: RNA prototype. The body, camera, and world coordinate systems are denoted by {B} (or XbYbZb), {C} (or XcYcZc), {W} (or XwYwZw),respectively. The initial {B} is taken as the world coordinate system {W} after performing a rotation to make the Z-axis level and align the Y-axis with the gravity vector ← g-. In this paper, the superscripts b and c describe a variable in{B} and {C}, respectively. Bottom: Solidworks drawing of the ART.

    IV. DEPTH-ENHANCED VISUAL-INERTIAL ODOMETRY

    The motion state of RNA is estimated by the proposed DVIO method which consists of three parts: feature tracker,floor detector, and state estimator. The feature tracker extracts visual features from a color image and tracks them to the next image. It also selects keyframes based on the average parallax difference. If the average parallax of the tracked features between the current frame and the latest keyframe is larger than a threshold (10 pixels), this frame is treated as a keyframe. The tracked features in the keyframes are passed to the optimization process to estimate the VINS’ motion state.The features extracted from the non-keyframes are only used for tracking. The floor detector extracts the floor plane from the D435’s depth data. The state estimator estimates the state of the IMU by using the visual features, the floor plane, the depth data, and the IMU measurements. The details of each part are described below.

    A. Feature Tracker

    The feature tracker detects Harris corner features [41] at each image frame. To obtain a higher processing speed without compromising pose estimation accuracy, the image is evenly divided into 8 × 8 patches, within which at most 4 features are extracted and tracked. These features (at most 256) are tracked across image frames by the KLT tracker [42].A RANSAC process based on the fundamental matrix is devised to remove outliers that do not satisfy epipolar constraint. Inliers are passed to the state estimator for pose estimation.

    B. Floor Detector

    Fig. 2. Graph structure for DVIO.

    C. State Estimator

    Fig. 3. Characterization of the D435 camera: the linear motion table moves the camera from 400 mm to 2400 mm with a step-size of 100 mm. At each position, 300 frames of depth data were captured and used to compute the mean and RMS of the measurement errors. The method in [41] was employed to estimate the ground truth depth, which is then refined by using the known camera movement (100 mm) to obtain the ground truth depth. Given a camera pose, the wall plane is projected to the camera frame as the ground truth plane.

    V. VISUAL POSITIONING SYSTEM FOR ASSISTIVE NAVIGATION

    The DVIO-estimated pose is used to 1) generate a 3D point cloud map for obstacle avoidance, and 2) obtain a refined 2D pose by PFL on a floor plan map for wayfinding. DVIO and PFL form a visual positioning system, based on which an assistive navigation system is created as shown in Fig. 4. The system was developed based on the robot operating system(ROS) framework. Each ROS node is an independent functional module and it communicates with the others through a messaging mechanism. The Data Acquisition node acquires and publishes the camera’s and the IMU’s data,which are subscribed by the DVIO node for pose estimation.The Terrain Mapping node registers the depth data captured with different camera poses to form a 3D point cloud map,which is then reprojected onto the floor plane to create a 2D local grid map for obstacle avoidance and localization of RNA in the 2D floor plan. Based on the RNA’s location in the floor plan, the path planning module [45] determines the desired heading to direct RNA towards the next point of interest(POI). This information is passed to the Obstacle Avoidance module [46] to compute the desired direction of travel (DDT)that will move RNA towards the POI without colliding with the surrounding obstacle(s). Based on the DDT, the ART Controller steers RNA into the DDT, and the speech interface sends audio navigation messages to the blind traveler via the Bluetooth headset. Both the tactile and audio information will guide the blind traveler to move along the planned path. The details of the major modules such as PFL, path planning,obstacle avoidance, and ART control are described below.

    Fig. 4. Software pipeline for the assistive navigation software.

    A. Particle Filter Based Localization

    B. Path Planning

    We use our earlier POI graph method [45] for path planning. The graph’s nodes are the POIs (hallway junctions,elevators, etc.) and each edge between two nodes has a weight equal to the distance between them. The A* algorithm is used to find the shortest path from the starting point to the destination. At each POI along the path, a navigational message is generated based on the next POI. This message is conveyed to the user by the speech interface. In addition, at each junction POI where a turn is required, the needed heading angle change is computed as the difference between the current heading angle and the angle required to move towards the next POI.

    C. Obstacle Avoidance

    In this work, we employ the traversability field histogram(TFH) [46] method to determine an obstacle-free direction for RNA. First, a local terrain map surrounding RNA is converted into a traversability map (TM). Then, a polar traversability index (PTI) is computed for each 5° sector of the TM. The smaller the PTI, the more traversable the direction. The PTIs are structured in the form of a histogram. Consecutive sectors with a low PTI form a histogram valley, indicating a walkable direction to RNA. The valley closest to the RNA’s target direction is selected and the DDT for RNA is thus determined.The steering angle for RNA is calculated based on the DDT and the current RNA heading. The steering angle is then used to control the ART. In addition, a navigational message is generated based on the next POI. This message is conveyed to the user via the speech interface.

    D. ART Control

    Fig. 5. Left: RNA swings from A to B; Right: Computation of θ from the accelerometer data.

    body and it is known a priori.

    VI. EXPERIMENTS

    A. DVIO Accuracy: D435 + VN100

    The performance of DVIO was compared with that of VINS-Mono [17] and VINS-RGBD [30] by experiments.Eight datasets were collected by holding RNA and walking at a speed of ~0.7 m/s. During each data collection session, the user swung RNA just like using a white cane. The ground truth positions of the start point and endpoint are [0, 0, 0] and[0, 0, 20 m], respectively. We use the endpoint position error norm (EPEN) as the metric for pose estimation accuracy.DVIO’s pose estimation accuracy and computational cost can be tuned by adjusting the size of the sliding window. For the sake of real-time computation, we used a small window consisting of 4 pose-nodes for DVIO. For the fairness of comparison, VINS-Mono and VINS-RGBD also used a 4-node sliding window and their loop closure functions were disabled. To demonstrate that the use of the floor plane and the visual features with unknown depth improves pose estimation accuracy, we ran DVIO with three different conditions, denoted DVIO-DFV, DVIO-DF, and DVIO-D,representing the full DVIO implementation, DVIO that does not use visual features without depth, and DVIO that does not use visual features without depth and the floor plane,respectively. Their pose estimation accuracies are compared with that of VINS-Mono and VINS-RGBD in Table I. It can be seen that: 1) using the floor plane reduced the EPEN of DVIO-D by 21.5%; 2) using visual features without depth reduced the EPEN of DVIO-DF by 12.3%. Therefore, the full DVIO has the best accuracy. On average, it reduced the EPEN by 40.2% and 26.4% when compared with VINS-Mono and VINS-RGBD, respectively.

    B. DVIO Accuracy: Structure Core

    We collected eight more datasets from the most updated RGB-D camera with an integrated IMU-Occipital Structure Core (SC)-that can provide synchronized image, depth (0.7–7 meters), and inertial data and compared DVIO’s pose estimation performance with that of VINS-Mono and VINSRGBD by using these datasets. We characterized the SC by using the method in [43] and found that the depth measurement is of high accuracy (error < 2 cm) if the depth is no greater than 4.0 m (Fig. 6(b)).

    TABLE I COMPARISON OF EPENS (METERS) OF VINS-MONO,VINS-RGBD, AND DVIO

    Fig. 6. (a) SC sensor on a white cane for data collection. The coordinate systems of the body (IMU), color camera, and the LED-target are denoted by XbYbZb, XcYcZc, and XTYTZT, respectively. The LED-target will be tracked by the MoCap system to produce the ground truth poses. (b) Measurement error vs distance of the SC sensor.

    We installed the SC on a white cane in a way similar to D435 (see Fig. 6(a)) and collected eight datasets by swinging the cane and walking (~0.7 m/s) in our laboratory. Based on the ground truth poses provided by the OptiTrack motion capture (MoCap) system, we calculated the absolute pose error for each point on the trajectories generated by DVIO,VINS-Mono, and VINS-RGBD. Table II summarizes the results. It can be observed that DVIO has the smallest RMSE in seven of the eight experiments. Its RMSE is only slightly larger than that of VINS-RGBD in one experiment. This demonstrates that DVIO has a much more accurate pose estimation than the other methods. On average, it reduced the RMSE by 57.1% and 23.7% when compared with VINSMono and VINS-RGBD, respectively. The trajectories generated by the three methods for four of the experiments are compared in Fig. 7, which show that the trajectories generated by DVIO are more accurate than that of VINS-Mono/VINSRGBD.

    TABLE II RESULTS ON THE LAb DATASETS: RMSE OF THE ESTIMATED TRAjECTORY OF EACH VIO METHOD. TL - TRAjECTORY LENGTH

    C. Runtimes of DVIO and Other Modules

    Table III shows the runtimes of the major modules of the assistive navigation software system depicted in Fig. 4. The average runtimes of DVIO, terrain mapping, PFL, and obstacle avoidance are 55.6 ms, 19.9 ms, 17.5 ms, and 0.5 ms,respectively. Since each module runs as an independent thread on a different core of the CPU, the software system achieves real-time computation on the UP Board computer(~18 fps).

    TABLE III RUNTIME F OR MODULES OF RNA

    D. PFL Performance Evaluation

    To evaluate RNA’s localization performance, we carried out experiments by holding RNA and walking along several different paths on the second floor of the Engineering East Hall of Virginia Commonwealth University. We created the floor plan map (as shown in Fig. 8(a)) from the architectural floor plan drawing after performing necessary editing to the doors (to show the geometric shapes of the closed doors along the paths). The distinctive geometric shapes of the areas around the doors, junctions, and corners will be used by PFL for RNA localization in the floor plan. For each experiment,the target and actual endpoints of RNA were recorded and their difference is calculated as the EPEN for performance evaluation. Table IV summarizes the EPENs of the experiments. The trajectories estimated by PFL (i.e., DVIO +PF) and that by DVIO only are compared in Fig. 9 to demonstrate the improved localization accuracy. It can be seen that PFL has a smaller EPEN for each experiment. Its mean EPEN over all experiments is 0.58%, i.e., 82.5% smaller than that of DVIO, meaning that the particle filter reduces the DVIO-accrued pose error by 82.5% on average. It is noted that the use of EPEN in the percentage of path-length allows us to compute the mean value over experiments with different paths for overall performance comparison. In principle, PFL eliminates DVIO-accrued pose error whenever RNA “sees” a geometrically featured region. When RNA moves in a corridor (between two featured regions), PFL can eliminate the lateral but not the longitudinal position error. As a result,PFL’s pose error is the PF alignment error plus the uncorrected DVIO pose error since the last alignment(occurred at the last geometrically featured region). This means that the path-length does not affect the EPEN of the PFL method. One can see from Table IV that the EPEN of data sequence DS6/DS7 is much smaller than that of DS4 even if its path-length is much longer. This is because the endpoint of DS6 locates at junction 1 and the last concave wall of DS7 that RNA “saw” is very close to the endpoint while the elevator (endpoint for DS4) is much farther from junction 3 (the last-seen feature).

    From the trajectory plots (Fig. 9), it can be seen that the trajectories estimated by DVIO (blue lines) intersect with the walls or doors as the result of the accrued pose error. But the PFL method eliminated the pose errors from time to time and resulted in much more accurate trajectories (red lines).

    E. Wayfinding Experiments

    We tested the practicality of the visual positioning system by performing two navigation tasks in the Engineering East Hall. Task I is from RM 2264 to RM 2252 (path-length: ~35 meters) and task II is from RM 2264 to the elevator (pathlength: ~80 meters). Two sighted persons (blind-folded)performed these tasks. Each person conducted two experiments for each task and he/she stopped at the point when RNA indicated that the destination had been reached.The EPENs (in meters) for the experiments are tabulated in Table V. As the path-length does not affect a PFL-estimated trajectory, we use the absolute EPEN as the performance metric. The average EPEN for tasks I and II are 0.20 m and 0.45 m, respectively. Due to the small error, RNA successfully guided the users to get to the destinations in all experiments. In Table V, we also show the mean EPENs over persons and that over experiments for each task. Their values are close to the overall averaged value (0.20 m or 0.45 m),indicating a consistent localization performance.

    In these wayfinding experiments, we placed numerous obstacles along the paths to test the assistive navigation system’s obstacle avoidance function. The results show that the obstacle avoidance module functioned well and the ART successfully steered RNA into an obstacle-free direction toward the destination. As this is beyond the focus of this paper, we omit the details for simplicity. Successful obstacle avoidance reflects accurate pose estimation of PFL from a different aspect.

    VII. CONCLUSION AND FUTURE WORk

    Fig. 7. From left to right: trajectory comparison for datasets S1, S2, S5, and S8. The trajectories of the ground truth, VINS-Mono, VINS-RGBD, and DVIO are plotted in black, blue, green, and red, respectively. o indicates the start point and the end point of a trajectory.

    Fig. 8. Experimental settings for localization/wayfinding experiments.

    This paper presents a new VIO method, called DVIO, for 6-DOF pose estimation of an RGB-D-camera-based VINS. The method achieves better accuracy by using the geometric feature (the floor plane extracted from the camera’s depth data) to add constraints between the graph nodes to reduce the accumulative pose error. Specifically, it tightly couples the floor plane, the visual features, and the IMU’s inertial data in a graph optimization framework for pose estimation. Based on the characterization of the camera’s depth measurements,visual features are classified into ones with a near-range depth and ones with a far-range depth. For near-range visual features, the depth values are initialized and updated by directly using the camera’s depth measurements because these measurements are accurate. For far-range visual features, thedepths are regarded as unknown values because the camera’s depth measurements are less accurate and therefore, the epipolar plane model is used to create constraints between the related nodes in the graph. The use of the floor plane and the inclusion of both visual features with and without a depth value improved the pose estimation accuracy. To support wayfinding application in a large indoor space, a PFL method is devised to limit the accumulative pose error of DVIO by using the information of the operating environment’s floor plan. The PFL method builds a 2D local grid map by using the DVIO-estimated egomotion and aligns this map with the floor plan map to minimize the pose error. PFL and DVIO form a VPS for accurate device localization on the 2D floor plan map.

    TABLE IV COMPARISON OF EPENS: METERS (% OF PATH-LENGTH)

    We validated the VPS’ localization function in the context of assistive navigation RNA in a large indoor space. To extend VPS into a full navigation system, we developed other essential software modules, including data acquisition, path planning, obstacle avoidance, and ART control. The ART mechanism can steer RNA into the desired direction of travel to guide the visually impaired user to avoid obstacles and move towards the destination. Experimental results validate that: 1) DVIO has better pose estimation accuracy than stateof-the-art VIO and it achieves real-time computation on a UP Board computer; 2) PFL can substantially reduce DVIO’s accumulative pose error for localization in a floor plan; and 3)VPS can be effectively used for assistive navigation in a large indoor space for both wayfinding and obstacle avoidance.

    In terms of future work, we will recruit visually impaired human subjects to conduct experiments in various indoor environments to validate the assistive navigation function of the RNA prototype.[1]R. R A Bourne, S. R. Flaxman, T. Braithwaite, M.V. Cicinelli,et al.,“Magnitude, temporal trends, and projections of the global prevalence of blindness and distance and near vision impairment: A systematic review and meta-analysis,”Lancet Glob Healthm, vol. 5, no. 9,pp. 888–897, 2017.

    Fig. 9. Trajectories estimated by DVIO and PFL (DVIO + PF) on the floor plan of the Engineering East Hall. The start and endpoint for DS6/DS7 are the same.

    TABLE V EPENS OF WAYFINDING EXPERIMENTS

    猜你喜歡
    稠油油井規(guī)律
    相變換熱技術(shù)在油田稠油開采中應(yīng)用
    化工管理(2022年14期)2022-12-02 11:42:50
    規(guī)律睡眠中醫(yī)有妙招
    稠油不愁
    找規(guī)律 畫一畫 填一填
    找排列規(guī)律
    新型油井水泥消泡劑的研制
    一種油井水泥用抗分散絮凝劑
    巧解規(guī)律
    生化微生物技術(shù)在稠油采出水處理中的應(yīng)用
    遼河油田破解稠油高溫調(diào)剖難題
    国产精品久久久久久久电影| 国产免费一区二区三区四区乱码| 一边亲一边摸免费视频| 久久久国产一区二区| 久久久久精品久久久久真实原创| 亚洲欧美日韩无卡精品| 午夜福利在线观看免费完整高清在| 亚洲第一av免费看| 国产伦精品一区二区三区四那| 久久这里有精品视频免费| 成年人午夜在线观看视频| 欧美一区二区亚洲| 日韩强制内射视频| xxx大片免费视频| 91久久精品国产一区二区三区| 欧美日韩视频高清一区二区三区二| 国产av国产精品国产| 大码成人一级视频| 国产视频首页在线观看| 大片免费播放器 马上看| 亚洲精品日本国产第一区| 少妇猛男粗大的猛烈进出视频| 国产亚洲最大av| 在线观看免费高清a一片| av一本久久久久| 99久久中文字幕三级久久日本| 黑丝袜美女国产一区| 直男gayav资源| 高清在线视频一区二区三区| 国产视频内射| 99热6这里只有精品| 97精品久久久久久久久久精品| 亚洲色图综合在线观看| 青青草视频在线视频观看| 久久99蜜桃精品久久| 永久网站在线| 麻豆成人av视频| 精品亚洲乱码少妇综合久久| 丝瓜视频免费看黄片| av在线app专区| 亚洲av成人精品一区久久| 熟女av电影| 人妻制服诱惑在线中文字幕| 少妇高潮的动态图| 日韩国内少妇激情av| 国产伦精品一区二区三区四那| 97超碰精品成人国产| .国产精品久久| 午夜福利在线在线| 国产av精品麻豆| 免费观看在线日韩| 人体艺术视频欧美日本| 2022亚洲国产成人精品| 成人毛片a级毛片在线播放| 成人毛片60女人毛片免费| 欧美xxⅹ黑人| 日本av免费视频播放| 大码成人一级视频| 一级毛片久久久久久久久女| 欧美xxxx黑人xx丫x性爽| 男人爽女人下面视频在线观看| 香蕉精品网在线| 亚洲欧美精品自产自拍| 国产亚洲最大av| 成人国产av品久久久| 免费在线观看成人毛片| 在线免费观看不下载黄p国产| 一本一本综合久久| 国产高清国产精品国产三级 | 亚洲精品中文字幕在线视频 | 亚洲欧洲日产国产| 免费观看无遮挡的男女| 91在线精品国自产拍蜜月| 国语对白做爰xxxⅹ性视频网站| 视频中文字幕在线观看| 国产免费视频播放在线视频| 国产 精品1| 国产一级毛片在线| 国产色婷婷99| 国产成人一区二区在线| 五月开心婷婷网| 久久国产精品大桥未久av | 精品熟女少妇av免费看| 日韩av在线免费看完整版不卡| 国产精品无大码| 蜜臀久久99精品久久宅男| 精品久久久久久久久亚洲| 男女边摸边吃奶| 高清日韩中文字幕在线| 黑人猛操日本美女一级片| 亚洲精品国产成人久久av| 亚洲国产精品一区三区| 少妇裸体淫交视频免费看高清| 一区二区三区四区激情视频| 国产精品久久久久久精品电影小说 | 大片免费播放器 马上看| 在线 av 中文字幕| 欧美 日韩 精品 国产| 亚洲三级黄色毛片| 日韩国内少妇激情av| 精品人妻偷拍中文字幕| 亚洲av男天堂| 免费观看性生交大片5| 狂野欧美激情性bbbbbb| 成人一区二区视频在线观看| 国产免费一级a男人的天堂| 国产伦精品一区二区三区视频9| 午夜免费鲁丝| 99热6这里只有精品| 小蜜桃在线观看免费完整版高清| 天堂中文最新版在线下载| 七月丁香在线播放| 男女边吃奶边做爰视频| 亚洲人与动物交配视频| 如何舔出高潮| 亚洲国产精品国产精品| 欧美高清成人免费视频www| 欧美激情极品国产一区二区三区 | 黄色日韩在线| 国产女主播在线喷水免费视频网站| 少妇人妻 视频| 啦啦啦在线观看免费高清www| 十分钟在线观看高清视频www | 在线观看免费视频网站a站| 亚洲精品一二三| av在线蜜桃| 99久久精品热视频| 亚洲色图综合在线观看| 国产免费一区二区三区四区乱码| 午夜免费鲁丝| 爱豆传媒免费全集在线观看| 国产一区二区三区av在线| 国产免费福利视频在线观看| 国产精品.久久久| 中文在线观看免费www的网站| 国产美女午夜福利| 成人特级av手机在线观看| 日本-黄色视频高清免费观看| 久久久精品免费免费高清| 狠狠精品人妻久久久久久综合| 水蜜桃什么品种好| av又黄又爽大尺度在线免费看| 中文欧美无线码| 欧美老熟妇乱子伦牲交| tube8黄色片| 日韩 亚洲 欧美在线| 亚洲精品久久午夜乱码| 一区在线观看完整版| 午夜福利在线观看免费完整高清在| 一级爰片在线观看| 久久久久网色| 啦啦啦视频在线资源免费观看| 综合色丁香网| 精华霜和精华液先用哪个| 亚洲第一区二区三区不卡| 国产欧美日韩精品一区二区| 亚洲精品中文字幕在线视频 | 国产精品不卡视频一区二区| 乱系列少妇在线播放| 日韩av免费高清视频| 亚洲av日韩在线播放| 九草在线视频观看| 男人狂女人下面高潮的视频| 丰满少妇做爰视频| 嫩草影院新地址| 欧美少妇被猛烈插入视频| 精品国产三级普通话版| 最近中文字幕2019免费版| 精品一品国产午夜福利视频| 久久午夜福利片| 在线观看三级黄色| 深夜a级毛片| 成人毛片a级毛片在线播放| 看免费成人av毛片| 亚洲精品中文字幕在线视频 | 在线免费十八禁| 激情 狠狠 欧美| 日日撸夜夜添| 亚洲熟女精品中文字幕| 免费大片黄手机在线观看| 亚洲国产成人一精品久久久| 性色avwww在线观看| 色婷婷av一区二区三区视频| 亚洲av二区三区四区| 五月开心婷婷网| 日产精品乱码卡一卡2卡三| 久久ye,这里只有精品| 国产精品99久久99久久久不卡 | av卡一久久| 黄片wwwwww| 男女边摸边吃奶| 高清午夜精品一区二区三区| 亚洲欧美成人精品一区二区| 中文字幕制服av| 久久av网站| 卡戴珊不雅视频在线播放| 草草在线视频免费看| 老司机影院毛片| 国产成人精品福利久久| 国产伦精品一区二区三区四那| a级毛色黄片| 国产淫片久久久久久久久| 2021少妇久久久久久久久久久| 婷婷色综合www| 国产精品99久久99久久久不卡 | 久久精品国产亚洲av涩爱| 日本免费在线观看一区| 啦啦啦在线观看免费高清www| 青春草亚洲视频在线观看| 校园人妻丝袜中文字幕| 欧美xxxx黑人xx丫x性爽| 免费看av在线观看网站| 亚洲欧美日韩东京热| 少妇人妻 视频| 亚洲高清免费不卡视频| 午夜精品国产一区二区电影| 麻豆成人av视频| 汤姆久久久久久久影院中文字幕| 亚洲精品乱久久久久久| 中国国产av一级| 国产人妻一区二区三区在| 精品一品国产午夜福利视频| 午夜福利高清视频| 国产精品精品国产色婷婷| 日韩亚洲欧美综合| 国产一区有黄有色的免费视频| 中文资源天堂在线| 高清午夜精品一区二区三区| 九九久久精品国产亚洲av麻豆| 深爱激情五月婷婷| 少妇熟女欧美另类| 男人添女人高潮全过程视频| 身体一侧抽搐| 麻豆国产97在线/欧美| 狂野欧美白嫩少妇大欣赏| 国产精品人妻久久久久久| 自拍偷自拍亚洲精品老妇| 久久人人爽av亚洲精品天堂 | 在线看a的网站| 黑人猛操日本美女一级片| 亚洲成人一二三区av| 国产成人freesex在线| 日韩精品有码人妻一区| 好男人视频免费观看在线| 国产亚洲av片在线观看秒播厂| 免费观看av网站的网址| 国产国拍精品亚洲av在线观看| 久久精品国产亚洲网站| 伊人久久精品亚洲午夜| 美女国产视频在线观看| 日日啪夜夜撸| 精品国产三级普通话版| 国内精品宾馆在线| 国产精品99久久久久久久久| 亚洲精品aⅴ在线观看| 最近中文字幕高清免费大全6| 欧美精品亚洲一区二区| 日日摸夜夜添夜夜添av毛片| 少妇被粗大猛烈的视频| 丰满少妇做爰视频| 日本欧美视频一区| 亚洲性久久影院| 日韩人妻高清精品专区| 黄色怎么调成土黄色| 18禁动态无遮挡网站| 欧美激情国产日韩精品一区| 久久99热这里只有精品18| 精品久久久精品久久久| 中文字幕免费在线视频6| 免费观看av网站的网址| 欧美少妇被猛烈插入视频| 亚洲精品日本国产第一区| 亚洲三级黄色毛片| av在线观看视频网站免费| 一区二区三区四区激情视频| 欧美成人午夜免费资源| 国产成人精品久久久久久| 激情五月婷婷亚洲| 免费看日本二区| 亚洲激情五月婷婷啪啪| 99久国产av精品国产电影| 女人久久www免费人成看片| 丰满迷人的少妇在线观看| 黄色欧美视频在线观看| av免费在线看不卡| 国产大屁股一区二区在线视频| 哪个播放器可以免费观看大片| 免费黄色在线免费观看| 日韩欧美 国产精品| av国产久精品久网站免费入址| 国产在线视频一区二区| 狂野欧美激情性xxxx在线观看| 精品久久久噜噜| 日本wwww免费看| 如何舔出高潮| av在线观看视频网站免费| 乱系列少妇在线播放| 免费观看av网站的网址| 夜夜骑夜夜射夜夜干| 性高湖久久久久久久久免费观看| 欧美zozozo另类| 尾随美女入室| 国产精品一区二区在线不卡| 亚洲欧美成人精品一区二区| 国产精品一区二区在线不卡| 亚洲欧美成人精品一区二区| 国产精品精品国产色婷婷| 亚洲av中文字字幕乱码综合| 一级毛片电影观看| 99久久精品一区二区三区| 国产欧美另类精品又又久久亚洲欧美| 亚洲国产毛片av蜜桃av| 国产精品人妻久久久影院| 97在线人人人人妻| www.色视频.com| 亚洲国产成人一精品久久久| 久久国产乱子免费精品| 超碰97精品在线观看| 亚洲激情五月婷婷啪啪| 亚洲欧美日韩另类电影网站 | xxx大片免费视频| 国产大屁股一区二区在线视频| 亚洲美女视频黄频| 高清毛片免费看| 精品久久久久久久久亚洲| 免费看av在线观看网站| 哪个播放器可以免费观看大片| 日本黄色片子视频| 国产 一区精品| 韩国高清视频一区二区三区| 街头女战士在线观看网站| 国产成人aa在线观看| 国产高潮美女av| 亚洲欧美清纯卡通| 色网站视频免费| 亚洲婷婷狠狠爱综合网| 一个人看的www免费观看视频| 国精品久久久久久国模美| 伊人久久精品亚洲午夜| 一区二区三区四区激情视频| 性高湖久久久久久久久免费观看| 99久久综合免费| 黄色配什么色好看| 久久 成人 亚洲| 欧美激情极品国产一区二区三区 | av.在线天堂| 最后的刺客免费高清国语| 国产精品一区二区在线观看99| 午夜福利视频精品| 夫妻性生交免费视频一级片| 一本久久精品| 亚洲欧美一区二区三区黑人 | 久热这里只有精品99| 自拍欧美九色日韩亚洲蝌蚪91 | 国产伦精品一区二区三区四那| 精品人妻视频免费看| 最近中文字幕高清免费大全6| 女人十人毛片免费观看3o分钟| av线在线观看网站| 精华霜和精华液先用哪个| 国产老妇伦熟女老妇高清| 五月开心婷婷网| 18禁在线播放成人免费| tube8黄色片| 日韩伦理黄色片| 国产无遮挡羞羞视频在线观看| 欧美三级亚洲精品| 丰满迷人的少妇在线观看| 国产色婷婷99| 久久精品久久久久久噜噜老黄| 色视频www国产| 亚洲精品乱码久久久久久按摩| 免费观看的影片在线观看| 亚洲成色77777| 日韩成人av中文字幕在线观看| 青春草视频在线免费观看| 亚洲国产av新网站| 久久精品久久久久久久性| 黑丝袜美女国产一区| 在线观看国产h片| 成人综合一区亚洲| 99久久精品热视频| 一级黄片播放器| 成年美女黄网站色视频大全免费 | 五月开心婷婷网| 18禁在线播放成人免费| 联通29元200g的流量卡| av国产免费在线观看| 人妻少妇偷人精品九色| 精品熟女少妇av免费看| 在线观看免费高清a一片| 国产探花极品一区二区| 精品午夜福利在线看| 亚洲国产精品国产精品| 99热国产这里只有精品6| 国产精品久久久久久精品古装| 亚洲精品自拍成人| 国产 精品1| 国产高潮美女av| 成年人午夜在线观看视频| 亚洲av二区三区四区| 国产精品爽爽va在线观看网站| 男人添女人高潮全过程视频| 成人国产av品久久久| 晚上一个人看的免费电影| 蜜臀久久99精品久久宅男| 51国产日韩欧美| 成人毛片60女人毛片免费| 国产亚洲精品久久久com| 伦精品一区二区三区| 大码成人一级视频| 女性被躁到高潮视频| 夫妻性生交免费视频一级片| 夫妻午夜视频| 国产亚洲精品第一综合不卡| 女警被强在线播放| 久久久久久免费高清国产稀缺| 男女之事视频高清在线观看 | 国产高清视频在线播放一区 | 嫁个100分男人电影在线观看 | 五月开心婷婷网| www.熟女人妻精品国产| 国产野战对白在线观看| 欧美成狂野欧美在线观看| 久热爱精品视频在线9| 少妇裸体淫交视频免费看高清 | 国产又色又爽无遮挡免| 成人国产av品久久久| 亚洲人成电影观看| 国产精品亚洲av一区麻豆| 黄网站色视频无遮挡免费观看| 国产熟女欧美一区二区| 在线观看免费高清a一片| 亚洲人成网站在线观看播放| 黄色片一级片一级黄色片| 99热全是精品| 亚洲一码二码三码区别大吗| 我的亚洲天堂| 欧美老熟妇乱子伦牲交| 精品久久蜜臀av无| 成年av动漫网址| 久久九九热精品免费| 国产亚洲av高清不卡| 久久精品成人免费网站| 欧美 亚洲 国产 日韩一| 欧美日韩综合久久久久久| 国产免费视频播放在线视频| 欧美黑人欧美精品刺激| 国产成人啪精品午夜网站| 午夜免费成人在线视频| 香蕉国产在线看| 性色av一级| 亚洲熟女精品中文字幕| 建设人人有责人人尽责人人享有的| 亚洲一码二码三码区别大吗| 婷婷色av中文字幕| 高清不卡的av网站| 美女脱内裤让男人舔精品视频| 五月天丁香电影| 日韩av在线免费看完整版不卡| 婷婷色av中文字幕| e午夜精品久久久久久久| 精品一区二区三卡| 久久久久久久久免费视频了| 久久国产精品大桥未久av| 五月天丁香电影| 十八禁人妻一区二区| 亚洲精品国产av蜜桃| av片东京热男人的天堂| 精品一品国产午夜福利视频| 久久青草综合色| 天天添夜夜摸| 一级a爱视频在线免费观看| 亚洲天堂av无毛| videos熟女内射| 欧美国产精品一级二级三级| 九草在线视频观看| 一本综合久久免费| 婷婷色麻豆天堂久久| 在线亚洲精品国产二区图片欧美| 各种免费的搞黄视频| 欧美变态另类bdsm刘玥| 日日爽夜夜爽网站| 国产成人精品久久二区二区91| 丰满饥渴人妻一区二区三| 亚洲欧洲国产日韩| 在线观看免费视频网站a站| 午夜免费鲁丝| 精品人妻熟女毛片av久久网站| 国产成人精品久久二区二区免费| 欧美精品人与动牲交sv欧美| 久久久久久久大尺度免费视频| 亚洲精品乱久久久久久| 亚洲第一青青草原| 9191精品国产免费久久| 久久久亚洲精品成人影院| 久久久久久久大尺度免费视频| 各种免费的搞黄视频| a级片在线免费高清观看视频| 看免费av毛片| 狠狠婷婷综合久久久久久88av| 国产日韩欧美视频二区| 亚洲av成人精品一二三区| 精品视频人人做人人爽| av在线播放精品| 中文字幕人妻丝袜制服| 性色av一级| 国产成人免费观看mmmm| 国产一卡二卡三卡精品| 亚洲国产最新在线播放| 十八禁人妻一区二区| 国产在线一区二区三区精| 国产精品欧美亚洲77777| 大香蕉久久网| 黄色视频在线播放观看不卡| 久久精品久久久久久噜噜老黄| 日韩 欧美 亚洲 中文字幕| 高潮久久久久久久久久久不卡| 少妇裸体淫交视频免费看高清 | 99久久人妻综合| 夫妻午夜视频| 久久女婷五月综合色啪小说| 亚洲精品一卡2卡三卡4卡5卡 | 2018国产大陆天天弄谢| 91精品三级在线观看| 亚洲欧美清纯卡通| 中文欧美无线码| 无遮挡黄片免费观看| 久久国产精品人妻蜜桃| 香蕉丝袜av| 啦啦啦啦在线视频资源| 爱豆传媒免费全集在线观看| 男女床上黄色一级片免费看| 久久精品亚洲熟妇少妇任你| 国产熟女午夜一区二区三区| 男女无遮挡免费网站观看| 色视频在线一区二区三区| 天堂中文最新版在线下载| 亚洲精品中文字幕在线视频| 自拍欧美九色日韩亚洲蝌蚪91| 精品人妻熟女毛片av久久网站| 在现免费观看毛片| 最黄视频免费看| 精品视频人人做人人爽| 少妇猛男粗大的猛烈进出视频| 中文字幕人妻丝袜制服| 欧美日韩精品网址| 这个男人来自地球电影免费观看| 日韩中文字幕欧美一区二区 | 老司机靠b影院| 久久久精品国产亚洲av高清涩受| 丝袜喷水一区| 成人国产av品久久久| 777米奇影视久久| 一个人免费看片子| 极品人妻少妇av视频| 精品亚洲成国产av| 色综合欧美亚洲国产小说| 在线观看免费视频网站a站| 一级片'在线观看视频| 久久狼人影院| 免费久久久久久久精品成人欧美视频| 一区二区日韩欧美中文字幕| 男女国产视频网站| 深夜精品福利| 亚洲成人手机| 亚洲成色77777| 99九九在线精品视频| 色网站视频免费| 欧美国产精品va在线观看不卡| 人成视频在线观看免费观看| 秋霞在线观看毛片| 男人操女人黄网站| 亚洲少妇的诱惑av| xxx大片免费视频| av欧美777| 制服人妻中文乱码| www.av在线官网国产| av在线播放精品| 久久国产精品大桥未久av| 老司机深夜福利视频在线观看 | 免费人妻精品一区二区三区视频| 老汉色∧v一级毛片| 女警被强在线播放| 国产亚洲av片在线观看秒播厂| 一本—道久久a久久精品蜜桃钙片| 精品一品国产午夜福利视频| 人人澡人人妻人| 成人免费观看视频高清| 女人爽到高潮嗷嗷叫在线视频| 大型av网站在线播放| 亚洲熟女毛片儿| 最黄视频免费看| 日本91视频免费播放| 丰满人妻熟妇乱又伦精品不卡| 日本黄色日本黄色录像| av国产久精品久网站免费入址| 国产女主播在线喷水免费视频网站| 欧美成人午夜精品| 成年人黄色毛片网站| 日本五十路高清| 又粗又硬又长又爽又黄的视频| 亚洲国产毛片av蜜桃av| 国产97色在线日韩免费| 久热这里只有精品99| 一级毛片女人18水好多 | 日本vs欧美在线观看视频| 精品亚洲成国产av| 男女国产视频网站| 日韩熟女老妇一区二区性免费视频| 亚洲欧美清纯卡通| 色精品久久人妻99蜜桃| 亚洲精品国产区一区二| 18禁裸乳无遮挡动漫免费视频| 免费看不卡的av|