• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Learning Discriminatory Information for Object Detection on Urine Sediment Image

    2024-02-19 12:01:28SixianChanBinghuiWuGuodaoZhangYuanYaoandHongqiangWang

    Sixian Chan,Binghui Wu,Guodao Zhang,Yuan Yao and Hongqiang Wang

    1College of Computer Science and Technology,Zhejiang University of Technology,Hangzhou,310023,China

    2The Hefei Institutes of Physical Science,Chinese Academy of Sciences,Hefei,230071,China

    3Department of Digital Media Technology,Hangzhou Dianzi University,Hangzhou,310018,China

    4School of Computer Science,University of Nottingham Ningbo China,Ningbo,315100,China

    ABSTRACT

    In clinical practice,the microscopic examination of urine sediment is considered an important in vitro examination with many broad applications.Measuring the amount of each type of urine sediment allows for screening,diagnosis and evaluation of kidney and urinary tract disease,providing insight into the specific type and severity.However,manual urine sediment examination is labor-intensive,time-consuming,and subjective.Traditional machine learning based object detection methods require hand-crafted features for localization and classification,which have poor generalization capabilities and are difficult to quickly and accurately detect the number of urine sediments.Deep learning based object detection methods have the potential to address the challenges mentioned above,but these methods require access to large urine sediment image datasets.Unfortunately,only a limited number of publicly available urine sediment datasets are currently available.To alleviate the lack of urine sediment datasets in medical image analysis,we propose a new dataset named UriSed2K,which contains 2465 high-quality images annotated with expert guidance.Two main challenges are associated with our dataset:a large number of small objects and the occlusion between these small objects.Our manuscript focuses on applying deep learning object detection methods to the urine sediment dataset and addressing the challenges presented by this dataset.Specifically,our goal is to improve the accuracy and efficiency of the detection algorithm and,in doing so,provide medical professionals with an automatic detector that saves time and effort.We propose an improved lightweight one-stage object detection algorithm called Discriminatory-YOLO.The proposed algorithm comprises a local context attention module and a global background suppression module,which aid the detector in distinguishing urine sediment features in the image.The local context attention module captures context information beyond the object region,while the global background suppression module emphasizes objects in uninformative backgrounds.We comprehensively evaluate our method on the UriSed2K dataset,which includes seven categories of urine sediments,such as erythrocytes(red blood cells),leukocytes(white blood cells),epithelial cells,crystals,mycetes,broken erythrocytes,and broken leukocytes,achieving the best average precision(AP)of 95.3% while taking only 10 ms per image.The source code and dataset are available at https://github.com/binghuiwu98/discriminatoryyolov5.

    KEYWORDS

    Object detection;attention mechanism;medical image;urine sediment

    1 Introduction

    In clinical practice,there are typically two primary methods used to test urine specimens.These include the urine routine test and the urine sediment examination.Urine routine test indicators generally include urine red blood cells (RBCs),white blood cells (WBCs),and urine protein.Based on the indicators of urine routine tests,it is possible to determine whether a patient has the renal tubular disease,blood disease,glomerular function,liver and gallbladder disease,etc.[1].For example,a large number of RBCs,calcium oxalate crystals(CAOXs),and hyaline casts(HYALs)indicate the presence of urinary tract calculi[2].However,the accuracy of urine routine tests can be easily affected by the patient’s metabolic problems and drug use.Consequently,medical professionals are increasingly turning to urine sediment examination for certain important diagnoses.Urine sediment examination is one of the most important methods,providing a scientific basis for clinical diagnosis and improving the accuracy of diagnoses[3].Manual testing is time-consuming,labor-intensive,and subjective,which requires highly trained technicians.By detecting the amount of various urine sediment in sampling images,the amount of urine sediment in a sample can be calculated.Markedly elevated levels of red blood cells in urine can be indicative of the presence of kidney disease,lower urinary tract disease,and extrarenal disease.Similarly,a substantial increase in white blood cells may suggest the possibility of inflammation within the urinary system,while increased epithelial cells may indicate the presence of inflammation.The presence of an increased amount of crystals in urine may signal the presence of urinary stones,and high levels of mycetes are typically observed in patients with diabetes.Therefore,analyzing the urinary sediment content from a urine sample is a valuable and informative method for evaluating the condition of various organs within the urinary system.

    Object detection has made great progress in the field of natural images,remote sensing images and medical images[4,5].In the urine sediment examination,object detection automatically and efficiently locates and classifies particle instances from microscope images,which is precisely the task of object detection.To detect the images taken by the urine sediment analyzer,combining online and offline integration with the Internet is a trend that has emerged in recent years[6].We deploy the model on the server and then can get the amount of various urine sediment by uploading the urine sediment images to the server for detection,which certainly helps medical personnel save time and effort.This application requires fast speed to meet real-time detection needs.Although the machine learning method [7] is effective in stabilizing features and handling high-dimensional and low-dimensional data,its detection speed can sometimes be slower than desired.Like Support Vector Machine(SVM)uses Histogram of Oriented Gradient[8]to extract features,which takes about 1 s,and its detection performance is limited by intra-class differences.

    The deep learning method is significant because of its high speed and accuracy in urine sediment detection.CNN-based(Convolutional Neural Network)algorithms combine with traditional feature processing methods are not efficient.Ji et al.[9]added branches to identify and classify small objects separately,such as erythrocytes and leukocytes.The branches classify particles based on edge features in the segmented region.However,it is not easily generalizable to other categories,as each branch needs to be trained independently for a specific category.The CNN-based algorithms combined with the graphical neural network[10]can establish the relationship between features to improve classification accuracy.At present,convolution neural networks are getting deeper and deeper,and GCN is difficult to adapt.One-stage target detection algorithms like YOLOs [11,12] predict directly on the feature maps and take less time during inference.Introducing the attention mechanism into the deep learning model has proven to be an effective approach for enhancing the detection performance of small urine sediments [13].This method only utilizes the spatial information of the current layer’s feature map,without interacting with the feature maps of other layers.

    Datasets are the foundation of the computer vision on which the proposed methods are evaluated.Many natural image datasets have been proposed in recent years,facilitating the development of deep learning algorithms.In medical images,a few urine sediment datasets are publicly available,and only a few contain comprehensive annotation and classification.This manuscript annotates a urine sediment dataset to facilitate the application of deep learning in medical images.The challenge for the urine sediment dataset is that most particles are small-scale in the dataset.What’s more,there are locally dense cases in urine images which means particles overlap each other.Dense object detection is also a big challenge for detectors.Generic object detection algorithms contain [14] a large number of parameters,which tend to overfit the algorithm if trained on a dataset with a small number of samples,and these algorithms have poor detection performance for small objects.

    Based on the above observations,we propose the lightweight detector discriminatory-YOLOv5s based on[12],which takes advantage of high speed and outstanding detection performance.YOLOv5 is convenient for adjusting the model size by changing hyperparameters.What’s more,it is easy to be deployed to the chip due to the reproducibility of the structure.This detector achieves better performance than normal object detectors with the proposed local context attention (LCA)module and global background suppression(GBS)module.These two modules improve the models’discriminatory ability.The LCA module makes the model pay more attention to context information around the particles to distinguish dense particles.The GBS module is designed to reduce the impact of background by fusing different level features.The feature fusion highlights the low-level features for better classification and localization.Our contributions are as follows:

    · We present a new urine sediment dataset called UriSed2K containing 2465 images with 7 categories.This dataset contains 29450 annotated labels.

    · We design a novel intelligent automatic system,which contains a detector called discriminatory-YOLOv5s with LCA module and GBS module for object detection on urine sediment images.The LCA module is designed to distinguish dense particles.This module helps the model find correlated features around the particles by capturing context information.The GBS module is proposed to suppress background information by learning the importance weight of feature maps.It makes the model focus on non-background information and improves the detection performance of small particles.

    · Our proposed method achieves better performance than the mainstream object detection methods in two urine sediment datasets,while our model includes fewer parameters.

    The following content of this manuscript is summarized as follows.We briefly review the research related to our work in Section 2.We describe in detail the specifics of our annotated dataset in Section 3.Then,the specifics of LCA module and GBS module are given in Section 4.The experimental results are presented in Section 5.Finally,Section 6 summarizes our results and provides some ideas for future work.

    2 Related Work

    2.1 Urine Sediment Image Dataset

    In recent years,the size of image datasets has shown an increasing trend,from VOC [15] to ImageNet [16],MS-COCO [17].The Pascal VOC dataset contains 21,500 images in 20 classes.The MS-COCO dataset contains over 300,000 images in 80 classes,200,000 annotated,and more than 2 million instances in the entire dataset.The ImageNet dataset has more than 14 million images across more than 20,000 categories.More than 1 million of these images have clear category annotations and bounding boxes for the location of objects in the image.

    Compared to the vast natural image datasets,medical image datasets are relatively small in scale[18].We found a few existing works on urine sediment datasets for recognition or detection:the dataset in[19]contains 1671 images,which is used for recognition by using deep learning.The urine sediment dataset USE [20–22] contains 5376 useful urinary microscopical images after filtering,resulting in 43000 meaningful instances.Undoubtedly,it takes a lot of time to remove images that only include noise.The research [9] collects 650 images for 10 categories and gets 300000 segments from these images.UMID(Urine Microscopic Image Dataset)[23]comprises 367 annotated microscopic images of urine sediment of dimensions(1280,720).The UMID contains 3 types of cells,RBCs,WBCs and epithelial cells.What’s more,annotating labels requires relevant medical expertise.Since urine sediment images are difficult to obtain and labeling them is labor intensive,urine sediment datasets are valuable,and most are not publicly available.

    2.2 Object Detection of Urine Sediment Images

    There are many proven network models in the field of object detection.Object detection models are similar in extracting features.These features are extracted bottom-up from the image by convolution.Two-stage models such as Faster R-CNN[14]select regions that may contain objects,so they need to generate a large number of proposals which takes more time to process.Two-stage models take advantage of high precision.One-stage models such as YOLOs[11,12],FCOS[24],make predictions directly for each region of the image without the process of generating proposals.Their inference speeds are fast at the expense of precision,but the expense of precision is tolerable in practical application scenarios.

    These network models are originally designed for natural image datasets such as VOC,COCO,which have complex backgrounds and different categories.These models contain many parameters,and the models require sufficient data for adequate training.Urine sediment datasets are usually small in size,and these models may be overfitted when trained directly on these datasets.In practice,lightweight models containing fewer parameters should be used,which are also easy to implement.The distinctive characteristic of the urine sediment dataset is the small objects.Most of the particles take up less than 1% area of the image.In the field of medical images,feature fusion can use its own feature information to improve detection performance significantly [25].The Feature Pyramid Network(FPN)[26]fuses feature maps of various scales from top to bottom,transmitting semantic information from deep feature maps to shallow ones,which improves the detection performance of small objects.The FPN combined with DenseNet[22],alleviates the problem of category confusion in the urine sediment images and performs well in an image with dozens of small particles.B-PesNet enriches semantic information in shallow layers so that the deep convolution layers can leverage more information when concatenating feature maps [27].Reference [21] exchanged context information between multi-scale features in the FPN and then fuse these features by element-wise addition.The Path Aggregation Network(PAN)[28]followed behind FPN and transmits low-level information from the bottom up to the high level.

    CNNs are able to capture local features due to the nature of convolution.However,CNNs are limited by the receptive field and are less capable of modeling regions outside the receptive field.In order to interact between feature map pixels,it is becoming increasingly common to use self-attention architecture for CNN models to obtain global modeling capabilities[29].

    3 UriSed2K Dataset

    We collect 2465 unstained images that are captured by an automatic urine sediment detector.The automatic urine sediment detector can automatically detect the position of the sample,aspirate it,mix it,and allow it to settle.It then captures an image of the sample through a microscope in the catheter.The resolution of the urine sediment microscopy images is 800 by 600 pixels.Fig.1 displays the seven major categories present in our urine sediment dataset,which include leukocyte,erythrocyte,broken leukocyte,broken erythrocyte,mycete,crystal,and epithelial cell.

    Figure 1 :The samples of seven categories of particles in the UriSed2K

    We annotate the images using rectangular boxes with labels in the VOC format.Each box includes the category value and the coordinates of the upper left and lower right corners,which arec,x1,y1,x2,y2.To enable a wider range of algorithms to be trained on this dataset,we have developed scripts to convert the dataset from the VOC format to either YOLO or COCO format.The YOLO format annotations include the center coordinates of the bounding box and its width and height relative to the original image.The COCO format annotations are similar to VOC format and are saved in a single JSON file for all images.According to the labels we annotated,we draw two figures:Fig.2 shows the number of seven particles in our dataset,Fig.3 shows the distribution of rectangular boxes’ratio of width and height.There are three main challenges in the urine sediment dataset:

    Small-scale object:Fig.3 shows the distribution of rectangular anchor boxes’size.In our dataset,the area of the largest object is no more than 14% of the whole image.Over 80% of the particles take up less than 1% area of the image.About 50% of the particles are smaller than 0.5% of the image area.This strongly suggests that the urine sediment dataset can be regarded as a small-scale object detection dataset.

    Intra-class variation:Fig.1 shows the various shapes of the same particles.The shapes of mycete are determined by its reproduction.The shapes of crystals depend on the formation conditions.Additionally,the appearance of particles can be affected by various shooting conditions,including the intensity of the light source and the magnification level.

    Locally dense distribution: The distribution of particles in the urine images is globally sparse.Notably,even after dilution,some particles continue to cluster together.It is challenging to detect these aggregated particles one by one.

    Figure 2 :The number of particles.The numbers of various particles from left to right are 4322,18335,2723,1148,1083,894,and 945,respectively.The letter‘b’stands for‘broken’

    Figure 3 :The distribution of rectangular anchor boxes’size.The value on the coordinate axis indicates the ratio of the width and height of the anchor box to the width and height of the image.The darker the color of the pixel point,the more anchor boxes are on that scale

    4 Methods

    In view of the characteristics of the urine sediment dataset,we propose two modules to improve the detection performance in detecting small objects by stronger discrimination.Fig.4 shows our network model with two modules.

    Figure 4 : The network of the discriminatory-YOLOv5s.The backbone extracts feature maps at different levels.The local context attention(LCA)module captures long-range information and concatenates feature maps.The global background suppression(GBS)module suppresses the background through feature fusion.The prediction head is responsible for predicting three scale feature maps.The variable C represents the number of classes in the dataset.The value of 5+C contains information about the predicted bounding box,including the center coordinates,height,width,objectness score,and probabilities of C categories

    4.1 Local Context Attention Module

    Determined by the properties of convolutional neural networks,the low-level feature maps contain more detailed features,which are effective for the model in detecting and predicting small particles.However,in the urine sediment dataset,particles tend to aggregate and block each other,making it difficult for models relying solely on low-level features to locate and classify individual particles.To distinguish these aggregated particles,we introduce self-attention mechanism to make the model focus on regions outside the receptive field.We propose the Local Context Attention(LCA)module based on FPN [26] incorporating the self-attentive mechanism.The module transfers high-level semantic information after attention to low-level features.Inspired by[29]and[30],we design several multi-head self-attention(MHSA)layers with relative position encoding to the FPN,overcoming the limitation of the receptive field by the convolutional kernel size.The improved model can increase the interaction between information from distant locations in the feature map.

    To capture the correlation between two positions in the feature map,the MHSA layer attends to the features at those positions and computes a weighted sum of them.We then concatenate the resulting high-level feature maps with low-level feature maps to increase the model’s attention on dense areas in the low-level feature maps.As shown in Fig.4,we add the MHSA layers to different scale feature maps.To fit these feature maps,we design the length of the position encoding to adapt to the input feature map’s scale.Adjusting the length of the position encoding to the input feature map’s scale is necessary to ensure that the relative positions of features are accurately captured across all feature map scales.The position encoding is used to represent the spatial relationships between features in the input feature map,and its length needs to be adjusted according to the scale of the feature map to account for variations in the relative distances between features at different scales.

    In the MHSA layer,the input feature map is divided into multiple heads equally to allow for better capturing of the spatial relationships between pixels.Specifically,we divide the feature map into four heads based on the channel dimension in our module.Fig.5 shows the architecture of the MHSA layer with position encoding.For each head,we compute query,key,and value projections by 1×1 convolutions.These projections are used to compute attention scores between each pair of positions in the feature map.To account for the relative positions of features,we use two learnable vectors for each head.These position encodings are added to the query and key projections before the attention scores are computed.The context informationcontent_contextis computed byq,k.

    whereTrepresents transpose,and ? is the matrix multiplication.Thecontent_contextcontains correlation information between two pixels,which represents the importance of each location in the input feature map relative toq.In order to consider location information in space,we introduce relative position encoding to the feature maps.

    whereris the relative position distance encodings,RhandRware the learnable vectors to represent the relative position distances along the height and width dimensions,respectively,and ⊕is the elementwise sum.The position of each pixel point can be represented according to the offset in the height and width dimensions.

    wherecontent_positionis the position information,it is obtained by matrix multiplication between the transposed position encodingrTand the query projection.

    whereAttentionis the attention map which is the summary ofcontent_contextandcontent_positionaftersoftmax.The value in the attention map is ∈(0,1),which represents the relative importance of different positions in the input feature map.

    wherevis the value projection.SAis the reweighted feature map that reflects the relation between two pixels of the feature map.

    wherei∈[1,n],nis the preset number of head.

    whereOutis concatenated by all heads,the parameters in each head are independent.After the above operations,we utilize MHSA to enhance the position of feature maps with high similarity in each feature map before fusion.The LCA module overcomes the limitation of the receptive field and interacts with the detailed information of each position,such as texture,color,shape,etc.,improving the representative ability of the model.

    4.2 Global Background Suppression Module

    From Section 3,we learn that the distribution of particles in urine images is globally sparse,which means that the background takes up most of the image and provides little information.When we humans look at images,we actively ignore the background to focus on the objects we are interested in.To enable our model to do the same,we have designed the GBS module,which assigns low weights to the background and allows the model to ignore it.The GBS module is inspired by[31]and fuses different levels of features by learning the importance weights of spatial information adaptively.This means that low-level features such as edges,corners,colors,and shapes are given more importance for localization.After adding the GBS module,the model pays much less attention to the background,just as our human visual system does.

    Figure 5 : The architecture of one of the heads in a multi-head self-attention layer.C,H and W demonstrate the channel number,height and width of the given feature map.WQ,WK,and WV are three different 1×1 convolution layers,Rh and Rw are learnable vectors,and their lengths are initialized based on the input feature map,⊕is the element-wise sum,? is the matrix multiplication.The head focuses on the important positions of the given feature map,which is determined by the generated attention map

    Fig.6 shows the feature fusion of level 2 in the GBS module.For each level,the process of fusing the futures at the corresponding levellis as follows:

    whereFlrefers to the target level feature,andl,nare both integers in the set{1,2,3}.fn→lrepresents the feature map that has been resized from levelnto match the size of levell.αl,βl,γlrepresent the spatial importance weights of levell,which are shared across all channels.iandjdenote the position of the feature map.The weights of each point are obtained by the softmax function,and∈(0,1).For example,getting the weight ofαlas follows:

    whereopis the resizing operation,△and ▽represent the up-sampling and down-sampling,respectively.In the LCA module,the size of the feature map becomes larger and the number of channels decreases.Therefore,the resizing operation can be determined by the number of channels.For upsampling,we use a 1×1 convolution layer to compress dimension to the same levell.Then the feature map is enlarged by nearest interpolation,where the multiplier for spatial size is equal to 2×|l-n|.For down-sampling,if|l-n|=1,we use a 3×3 convolution layer with stride 2,similar to the downsampling in the YOLOv3 backbone.It is different when|l-n|=2,a 3×3 kernel size max pool layer with stride 2 added before the convolution layer.

    Figure 6 :The details of level 2 in the GBS module.△,▽,⊙,⊕represent up-sampling,down-sampling,element-wise multiplication and element-wise sum,respectively.α,β,γ are the spatial importance weights.f 1,f 2,f 3 are the feature maps.f 1→2,f 2→2,f 3→3 are the feature maps after resizing.The CA is a channel attention mechanism

    αl,βl,γlare learned from the spatial information and are shared across each channel.From the perspective of channel importance,we add channel attention to enhance cross-channel interaction.Inspired by [32],we add simple channel attention toFl.This channel attention first compresses the spatial information of each feature map using global average pooling and then learns the importance of each channel using one-dimensional convolution across channels,which can consider both local channel information and global channel information with low computational complexity.The last output is defined as follows:

    whereCONVrepresents the 1 dimensional convolution with kernel size 5,⊙represents the elementwise multiplication.GAPrepresents the global average pooling,andflis the feature of levellbefore operation.The fusion parameters in this module are obtained through convolution and are able to participate in standard back-propagation for learning.Through training,these parameters can be effectively adjusted to reweight the feature maps and reduce the weight of background regions,leading to an improvement in the model’s detection performance.

    4.3 Loss Function

    Same as normal object detection algorithms,the loss consists of three components:

    whereLcls,LboxandLobjindicate the classification loss,the box regression loss and the objectness score loss.λcls,λbox,λobjare the coefficients,they are preset as 0.5,0.05,1,respectively.The classification lossLclsis defined as follows:

    whereS2,Bare feature map size,and number of candidate boxes,respectively,BCEis binary cross entropy function,?cis the ground truth,andcis the predicted class.TheLboxis defined as follows:

    The intersection over union(IoU)is DIoU[33],better reflecting the overlap between the predicted box and ground truth.The objectness score loss is defined as follows:

    5 Experiments

    5.1 Implementation

    Dataset partition.Before training,we partitioned the urine sediment dataset.The dataset was randomly divided into training set,validation set,test set.The ratio of images in the three parts is 3:1:1.In order to verify the robustness of the proposed method,we performed three divisions to obtain three different divided datasets.Then we trained on the three datasets separately and got three models.

    Training.Our network model is trained end-to-end on a computer with 2 2080Ti GPUs.We use small YOLOv5 [12] as baseline whose network depth is 1/3 and the number of channels is 1/2 of YOLOv5.This network model contains fewer parameters.The input image size is 640×640,which is close to the real image.The main data augmentation trick is Mosaic,which randomly selects 4 images from a batch,then cuts and combines them.What’s more,images are randomly flipped vertically and horizontally according to a preset probability.The network is trained with stochastic gradient descent(SGD)by 50 epochs,in which the batch size,the weight decay and momentum are set to 16,5×10-4,0.937.In the first 3 epochs,the warm learning rate increases from 0.01 to 0.1 and then decreases the learning rate by cosine annealing to 0.002.

    Inference.Our anchor boxes have 3 aspect ratios for each level[0.5,1,2].For every pixel at each level,every anchor box matches the most similar shape object and then conducts classification and box regression.The number of objects in the images is much less than the generated anchor boxes.It is indispensable to remove abundant prediction boxes of the same class.Interaction over union(IoU)is the overlap rate of two boxes.If the IoU of two prediction boxes exceeds the threshold,this means that they are most likely the same object.Non-maximum suppression(NMS)[34]with the threshold at 0.6 is applied to each class separately.

    5.2 Comparison with Other Object Detection Algorithms

    Our method is compared to classical object detection algorithms such as Faster R-CNN [14],YOLOv3[11],FCOS[24],PAN[28],BCPNet[21].Faster R-CNN is the representative algorithm of the two-stage algorithms,and PAN adds a PANet after FPN on the basis of the Faster R-CNN.The PANet YOLOv3 and FCOS are one-stage algorithms,where YOLOv3 is the anchor-based algorithm and FCOS is the anchor-free algorithm.The backbone of YOLOv3 is DarkNet53,the others’backbones are the same,all are ResNet50.Despite their different names,both backbone networks are used to extract different scale features of images and both have residual structures.These networks are trained on the MMdetection framework[35]with 2 NVIDIA 2080Ti GPUs.Since our dataset is small compared to the natural image dataset,these networks are trained 1 schedule(12 epochs).

    Table 1 compares the detection performance of other mainstream detectors with ours in terms of AP(average precision).It demonstrates that our model outperforms other detectors in three different splits,while also having far fewer parameters.AP is the classical evaluation metric used to measure the performance of a detector.Typically,AP is defined as the average of the ratios of true positives to all positives for all recall values.0.5 is the IoU threshold of true positive.mAP@0.5:0.95 represents the mean value of 10 APs calculated with different IoU thresholds,evenly distributed between 0.5 and 0.95.

    Table 1 : The comparison of detection performance on the UriSed2k dataset

    Table 2 compares the detection performance on the UMID dataset [23].The UMID dataset labels three types of urine sediment,RBCs,WBCs,and epithelial cells.Besides,there are many urine sediments in the images that are not labeled.Despite these situations,we can see from the table that our method performs better on this dataset compared to other methods.

    Table 2 : The comparison of detection performance on the UMID

    Fig.7 shows some samples to demonstrate the effectiveness of our approach.We visualize the ground truth and the prediction boxes of each method overlaid onto the original images.When urine sediments are sparsely distributed and not occluded,our method performs almost as well as commonly used object detection methods.In addition,in the case of dense local occlusion,our method is able to more effectively distinguish these particles from the whole part.

    5.3 Ablation Study

    To validate the effectiveness of each component in our proposed method,we conducted ablation experiments on the UriSed2K dataset.The baseline model used was a small YOLOv5 object detector without the local context attention (LCA) module and the global background suppression (GBS)module.Table 3 shows two overall metrics and the precision of the seven classes of urine sediment particles at the IoU of 0.5.Our results demonstrate that the addition of the LCA and GBS modules significantly improves the precision of the seven types of particles.Overall,our results show that our proposed method effectively works on the urine sediment dataset.

    Table 3 : Ablation experimental performance on the UriSed2K dataset

    Figure 7 :The ground truth and prediction of each method

    Fig.8 shows the heat maps detected by four network models,reflecting where the network focuses.Each heat map is generated based on the convolutional output of the last layer of the network.Specifically,the calculation method is to normalize the predicted objectness scores for each position to between 0 and 1,then map them onto a heat map of the same size as the input image.This allows us to judge the degree of network attention to different areas based on the color of each position.These heat maps have a gradient of colors from blue to red,with darker red areas indicating more attention from the network.

    The heat maps in the third column demonstrate the effectiveness of our LCA module in capturing context information.This enables the model to increase its perception of information surrounding particles,thus effectively separating locally dense particles.However,we also notice that the LCA module can overly focus on the background,which is redundant.This is where the GBS module comes in,as it suppresses the model’s attention to the background.The heat maps in the fourth column show how the GBS module helps the model focus less on the background.Overall,the LCA and GBS modules work together to help the model deal with local density and global sparsity in urine sediment detection.In Fig.9,we can clearly see which categories the objects were predicted to belong to.Our improved method has significantly improved the detection recall rate compared to the baseline,and the accuracy of most categories has also improved.

    Figure 8 : We show three examples of local densities highlighted by red boxes.The heat maps are visualized to origin images by four network modules.The baseline is small YOLOv5 [12].The local context attention(LCA)module and global background suppression(GBS)module are our proposed methods.From the detection results,we see that the objects in the red boxes are well separately detected

    Figure 9 :The confusion matrix.The values in each cell of the confusion matrix are normalized to a range between 0 and 1.The values on the diagonal represent the recall rate of the predicted category.The accuracy of the prediction can be calculated by dividing the value on the diagonal by the sum of the corresponding row

    6 Conclusion

    This manuscript presented a urine sediment dataset called UriSed2K containing 2465 images with a total of 29450 instances.We proposed the improved method of discriminatory-YOLOv5s.It was a lightweight network model with two effective modules,the local context attention(LCA)module and the ground background suppression(GBS)module,for object detection on urine sediment images.The LCA module enhanced the model’s perception ability by capturing context information around the targets.The GBS module suppressed the model’s attention to the background,therefore enabling the model to concentrate on the locations with features.These two modules,proposed for the main challenges of the urine sediment dataset,obviously improved the detector’s performance in small objects and got the best average precision of 95.3% at IoU=0.5.Our dataset will contribute to object detection and recognition in urine sediment images.The UriSed2K dataset enables the development of automated urinalysis techniques,which reduces the time and labor costs associated with manual analysis and improves the accuracy and efficiency of analyses.Analyzing urine sediment datasets can provide insights into changes in urine sediments in different diseases or pathological states,and can offer valuable data support for related medical research.Our method can be combined with automatic urine sediment analyzers to obtain real-time counts of various urine sediment particles in images.This allows medical professionals to save time and analyze samples more efficiently,ultimately aiding in the diagnosis of medical conditions.Our method still requires many images to obtain detection performance,and there are many different forms of urine sediment in the images,which are difficult to label.There is great potential to apply this dataset in few-shot detection to reduce annotation efforts.

    Acknowledgement:The authors would like to thank the anonymous reviewers and the editors of the journal.Your constructive comments have improved the quality of this paper.

    Funding Statement:This work was partially supported by the National Natural Science Foundation of China (Grant Nos.61906168,U20A20171),Zhejiang Provincial Natural Science Foundation of China(Grant Nos.LY23F020023,LY21F020027),Construction of Hubei Provincial Key Laboratory for Intelligent Visual Monitoring of Hydropower Projects(Grant Nos.2022SDSJ01).

    Author Contributions: The authors confirm contribution to the paper as follows: data collection:Hongqiang Wang; study conception and design: Sixian Chan,Binghui Wu; coding,experimental evaluation:Binghui Wu;analysis and interpretation of results:Binghui Wu,Guodao Zhang;funding support:Sixian Chan,Hongqiang Wang,Yuan Yao;draft manuscript preparation:Guodao Zhang,Yuan Yao;draft review:Binghui Wu.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The dataset UriSed2K that supports the findings of this study is publicly available from https://github.com/binghuiwu98/discriminatory-yolov5.The dataset UMID is publicly available from[23].

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产97色在线日韩免费| 欧美三级亚洲精品| 国产真人三级小视频在线观看| 久久久久久大精品| 午夜精品在线福利| 九色国产91popny在线| 国产乱人伦免费视频| 人人妻人人澡欧美一区二区| 亚洲av日韩精品久久久久久密| 久久午夜亚洲精品久久| 后天国语完整版免费观看| 特大巨黑吊av在线直播| 在线观看舔阴道视频| 国产午夜福利久久久久久| 在线永久观看黄色视频| 亚洲黑人精品在线| 国产精品永久免费网站| 久久香蕉精品热| 国内揄拍国产精品人妻在线| 狠狠狠狠99中文字幕| 亚洲成a人片在线一区二区| 国产激情欧美一区二区| 欧美+亚洲+日韩+国产| 免费高清视频大片| 国产欧美日韩一区二区精品| 精品一区二区三区av网在线观看| 90打野战视频偷拍视频| 狂野欧美白嫩少妇大欣赏| 午夜精品一区二区三区免费看| 国产亚洲av高清不卡| 一级a爱片免费观看的视频| 1024香蕉在线观看| 国内毛片毛片毛片毛片毛片| 18禁裸乳无遮挡免费网站照片| 一本一本综合久久| 精品欧美国产一区二区三| 国产美女午夜福利| 日本黄色片子视频| 99久久成人亚洲精品观看| 亚洲在线观看片| 国产熟女xx| 色老头精品视频在线观看| 动漫黄色视频在线观看| 国产高潮美女av| 午夜影院日韩av| 一区福利在线观看| 十八禁人妻一区二区| 欧美国产日韩亚洲一区| 动漫黄色视频在线观看| 国产高清视频在线观看网站| 后天国语完整版免费观看| 国产在线精品亚洲第一网站| 男人舔女人的私密视频| 欧美性猛交黑人性爽| 美女 人体艺术 gogo| 免费一级毛片在线播放高清视频| 男女午夜视频在线观看| 日本免费a在线| 国产97色在线日韩免费| 亚洲一区二区三区不卡视频| 亚洲中文字幕一区二区三区有码在线看 | 免费在线观看视频国产中文字幕亚洲| 日韩欧美在线乱码| 午夜日韩欧美国产| 一个人看视频在线观看www免费 | 亚洲人成网站在线播放欧美日韩| 全区人妻精品视频| 日韩欧美 国产精品| 亚洲乱码一区二区免费版| 国产又色又爽无遮挡免费看| 国产精品一区二区三区四区久久| 极品教师在线免费播放| 免费观看的影片在线观看| av黄色大香蕉| 身体一侧抽搐| 精品国产美女av久久久久小说| 国产精品香港三级国产av潘金莲| 91在线观看av| 好看av亚洲va欧美ⅴa在| 亚洲专区国产一区二区| 成人国产一区最新在线观看| 精华霜和精华液先用哪个| 国产精品一及| 听说在线观看完整版免费高清| 一个人观看的视频www高清免费观看 | 欧美性猛交黑人性爽| 久久天躁狠狠躁夜夜2o2o| 精品一区二区三区四区五区乱码| 国产久久久一区二区三区| 99热精品在线国产| 1000部很黄的大片| 听说在线观看完整版免费高清| 最近在线观看免费完整版| www国产在线视频色| 欧美日本视频| 免费在线观看影片大全网站| 视频区欧美日本亚洲| 国产麻豆成人av免费视频| 国产亚洲精品综合一区在线观看| 欧美日本亚洲视频在线播放| 中亚洲国语对白在线视频| 天堂影院成人在线观看| 成人国产一区最新在线观看| 免费搜索国产男女视频| 久久国产精品人妻蜜桃| 亚洲精品在线观看二区| 欧美黑人巨大hd| 最近最新中文字幕大全电影3| 精品一区二区三区视频在线观看免费| 亚洲性夜色夜夜综合| 又爽又黄无遮挡网站| 久久久久国产精品人妻aⅴ院| 最新中文字幕久久久久 | 熟妇人妻久久中文字幕3abv| 热99在线观看视频| 在线观看日韩欧美| 两性午夜刺激爽爽歪歪视频在线观看| 丰满的人妻完整版| 在线观看午夜福利视频| 久久草成人影院| 成年女人永久免费观看视频| 嫩草影院精品99| 国产高清有码在线观看视频| 久久天躁狠狠躁夜夜2o2o| 天天躁日日操中文字幕| 一级作爱视频免费观看| 又爽又黄无遮挡网站| 亚洲在线自拍视频| 岛国在线观看网站| 999久久久国产精品视频| 国产主播在线观看一区二区| 国产精品爽爽va在线观看网站| 午夜福利在线观看吧| 欧美色欧美亚洲另类二区| 脱女人内裤的视频| 亚洲精品在线观看二区| 又大又爽又粗| 亚洲国产日韩欧美精品在线观看 | 岛国视频午夜一区免费看| 亚洲成人久久爱视频| 久久久久免费精品人妻一区二区| 免费av毛片视频| 午夜精品久久久久久毛片777| 中文字幕av在线有码专区| 亚洲成人精品中文字幕电影| 美女cb高潮喷水在线观看 | а√天堂www在线а√下载| 99久久成人亚洲精品观看| 每晚都被弄得嗷嗷叫到高潮| 听说在线观看完整版免费高清| 美女扒开内裤让男人捅视频| 两个人的视频大全免费| 成人午夜高清在线视频| 国产精品电影一区二区三区| 欧美在线黄色| 欧美成狂野欧美在线观看| 人人妻人人看人人澡| 一级a爱片免费观看的视频| 性色avwww在线观看| 天堂av国产一区二区熟女人妻| 丁香欧美五月| 一级作爱视频免费观看| 国产黄片美女视频| 久久久精品欧美日韩精品| 特大巨黑吊av在线直播| 999久久久国产精品视频| 日本a在线网址| 国产综合懂色| 无遮挡黄片免费观看| 午夜福利在线观看吧| 国产av麻豆久久久久久久| 少妇人妻一区二区三区视频| 后天国语完整版免费观看| 国产精品野战在线观看| 大型黄色视频在线免费观看| 国产精品九九99| 在线看三级毛片| 一个人看视频在线观看www免费 | 91av网一区二区| 亚洲国产欧美一区二区综合| 欧洲精品卡2卡3卡4卡5卡区| 小说图片视频综合网站| 天堂动漫精品| 美女 人体艺术 gogo| 久久精品国产综合久久久| 亚洲乱码一区二区免费版| 制服丝袜大香蕉在线| 中出人妻视频一区二区| 成人永久免费在线观看视频| 亚洲乱码一区二区免费版| 久久中文字幕人妻熟女| www国产在线视频色| 日韩欧美在线乱码| 999精品在线视频| 国产亚洲欧美在线一区二区| 一进一出抽搐动态| 久久久久久久久中文| 日本免费一区二区三区高清不卡| 久久久久久久午夜电影| 国产麻豆成人av免费视频| 国产成人精品久久二区二区免费| 波多野结衣高清无吗| 亚洲无线在线观看| 欧美丝袜亚洲另类 | 人人妻人人澡欧美一区二区| 亚洲九九香蕉| av片东京热男人的天堂| 国产精品电影一区二区三区| 国内少妇人妻偷人精品xxx网站 | 国产 一区 欧美 日韩| 精品不卡国产一区二区三区| 少妇人妻一区二区三区视频| 噜噜噜噜噜久久久久久91| 麻豆成人午夜福利视频| 国产精品一区二区免费欧美| 国产精品99久久久久久久久| 国产午夜福利久久久久久| 免费电影在线观看免费观看| 老鸭窝网址在线观看| 人妻久久中文字幕网| 国产精品乱码一区二三区的特点| 国产男靠女视频免费网站| av在线天堂中文字幕| 成人鲁丝片一二三区免费| 亚洲欧美精品综合一区二区三区| 国产一区二区激情短视频| 成熟少妇高潮喷水视频| 日韩欧美国产一区二区入口| 母亲3免费完整高清在线观看| 国产精品野战在线观看| 亚洲精品色激情综合| 欧美乱妇无乱码| 国产精品永久免费网站| 久久国产精品影院| 国产一区二区激情短视频| svipshipincom国产片| 日韩欧美国产一区二区入口| 禁无遮挡网站| 精品福利观看| 午夜亚洲福利在线播放| 亚洲国产中文字幕在线视频| 免费看a级黄色片| 日本a在线网址| 日韩欧美国产在线观看| 男女下面进入的视频免费午夜| 欧美日韩一级在线毛片| 日韩中文字幕欧美一区二区| 嫩草影院入口| 免费看光身美女| 日韩三级视频一区二区三区| 丰满人妻熟妇乱又伦精品不卡| 免费看a级黄色片| 精品欧美国产一区二区三| 老熟妇乱子伦视频在线观看| 免费看美女性在线毛片视频| 免费一级毛片在线播放高清视频| 国产成年人精品一区二区| 色av中文字幕| 亚洲欧美日韩无卡精品| aaaaa片日本免费| 日韩中文字幕欧美一区二区| 午夜福利在线观看吧| 欧美乱色亚洲激情| 亚洲男人的天堂狠狠| 日韩人妻高清精品专区| 两性夫妻黄色片| 欧美精品啪啪一区二区三区| 国产精品精品国产色婷婷| 欧美一级a爱片免费观看看| 国产真人三级小视频在线观看| 黄色片一级片一级黄色片| 夜夜看夜夜爽夜夜摸| 成人午夜高清在线视频| 免费在线观看视频国产中文字幕亚洲| 听说在线观看完整版免费高清| 高潮久久久久久久久久久不卡| 一区福利在线观看| 精品日产1卡2卡| www.自偷自拍.com| 香蕉国产在线看| 国语自产精品视频在线第100页| 午夜福利免费观看在线| 99热精品在线国产| 波多野结衣高清作品| 蜜桃久久精品国产亚洲av| 女人高潮潮喷娇喘18禁视频| 非洲黑人性xxxx精品又粗又长| 国产亚洲精品久久久久久毛片| 成人午夜高清在线视频| 天天一区二区日本电影三级| 男女之事视频高清在线观看| 国产午夜精品论理片| 亚洲熟妇中文字幕五十中出| 精品久久久久久久久久免费视频| 免费av毛片视频| 成熟少妇高潮喷水视频| www日本黄色视频网| 精品无人区乱码1区二区| 国内精品美女久久久久久| 天天躁日日操中文字幕| 欧美日韩精品网址| 欧美在线一区亚洲| 日韩精品中文字幕看吧| 欧美激情在线99| 一本综合久久免费| 最近视频中文字幕2019在线8| 啦啦啦韩国在线观看视频| 他把我摸到了高潮在线观看| 国产精品永久免费网站| 九九久久精品国产亚洲av麻豆 | 夜夜看夜夜爽夜夜摸| 久久亚洲真实| 国产黄片美女视频| 亚洲国产欧美一区二区综合| h日本视频在线播放| 亚洲欧美一区二区三区黑人| 两性午夜刺激爽爽歪歪视频在线观看| 一个人免费在线观看电影 | 男女视频在线观看网站免费| 久久久久国产精品人妻aⅴ院| 日韩欧美免费精品| 亚洲无线观看免费| 国产精品av久久久久免费| 国产激情欧美一区二区| 国产成人福利小说| 亚洲 欧美一区二区三区| 欧美黑人欧美精品刺激| 久久久久久久久免费视频了| 五月玫瑰六月丁香| 91老司机精品| 免费av毛片视频| 日本三级黄在线观看| 丁香欧美五月| 欧美极品一区二区三区四区| 国产一区二区在线观看日韩 | 精品久久久久久久久久久久久| 啦啦啦免费观看视频1| 狂野欧美激情性xxxx| 午夜免费成人在线视频| 亚洲成a人片在线一区二区| 又粗又爽又猛毛片免费看| 麻豆成人午夜福利视频| 精品电影一区二区在线| 精品无人区乱码1区二区| 大型黄色视频在线免费观看| 精品一区二区三区四区五区乱码| 日韩人妻高清精品专区| 在线观看免费午夜福利视频| 嫩草影视91久久| 熟女人妻精品中文字幕| 嫩草影视91久久| 在线观看免费午夜福利视频| 97碰自拍视频| 精品一区二区三区四区五区乱码| 亚洲真实伦在线观看| 午夜激情欧美在线| 国产av在哪里看| 国产成人av教育| 男人舔女人的私密视频| 免费在线观看成人毛片| 亚洲真实伦在线观看| aaaaa片日本免费| av天堂在线播放| 国产美女午夜福利| 国产av在哪里看| 亚洲成a人片在线一区二区| 国产av在哪里看| av片东京热男人的天堂| 97人妻精品一区二区三区麻豆| 又爽又黄无遮挡网站| 日韩欧美精品v在线| 狂野欧美激情性xxxx| 国产精品久久久av美女十八| 婷婷丁香在线五月| 国产主播在线观看一区二区| av黄色大香蕉| 亚洲18禁久久av| 日韩大尺度精品在线看网址| 色综合站精品国产| 亚洲无线观看免费| 观看美女的网站| 老司机深夜福利视频在线观看| 黄频高清免费视频| 国产黄片美女视频| 男插女下体视频免费在线播放| 日本黄色视频三级网站网址| 精品一区二区三区视频在线 | 国产午夜精品久久久久久| 日日摸夜夜添夜夜添小说| 国产三级黄色录像| 国产亚洲精品综合一区在线观看| 午夜亚洲福利在线播放| 国产黄片美女视频| 亚洲五月婷婷丁香| 一区二区三区高清视频在线| 久久久久性生活片| 欧美在线黄色| av黄色大香蕉| 长腿黑丝高跟| 看黄色毛片网站| 听说在线观看完整版免费高清| 国产成+人综合+亚洲专区| 51午夜福利影视在线观看| 成人高潮视频无遮挡免费网站| 99国产精品一区二区蜜桃av| 又紧又爽又黄一区二区| 久久性视频一级片| 日韩免费av在线播放| 精品一区二区三区四区五区乱码| 亚洲国产精品sss在线观看| 色av中文字幕| 宅男免费午夜| 成人一区二区视频在线观看| 亚洲性夜色夜夜综合| 国产精品99久久久久久久久| 日本 欧美在线| 夜夜看夜夜爽夜夜摸| 麻豆国产av国片精品| 999久久久国产精品视频| 免费在线观看亚洲国产| 五月伊人婷婷丁香| 丰满人妻熟妇乱又伦精品不卡| 丰满人妻一区二区三区视频av | 香蕉久久夜色| 免费搜索国产男女视频| 亚洲精品在线美女| 非洲黑人性xxxx精品又粗又长| 村上凉子中文字幕在线| 亚洲精品乱码久久久v下载方式 | 午夜精品在线福利| 日韩精品青青久久久久久| 最新在线观看一区二区三区| 国产高清激情床上av| 国产又黄又爽又无遮挡在线| 韩国av一区二区三区四区| 岛国视频午夜一区免费看| 国产欧美日韩精品一区二区| 大型黄色视频在线免费观看| 美女高潮的动态| 五月伊人婷婷丁香| 最新美女视频免费是黄的| 九色成人免费人妻av| 成人精品一区二区免费| 亚洲欧美一区二区三区黑人| 国产乱人伦免费视频| 日本一二三区视频观看| 国产三级中文精品| 亚洲第一欧美日韩一区二区三区| 久久久成人免费电影| 午夜精品一区二区三区免费看| 久久精品综合一区二区三区| 特级一级黄色大片| 一边摸一边抽搐一进一小说| 国内少妇人妻偷人精品xxx网站 | 欧美午夜高清在线| 99热精品在线国产| 观看美女的网站| 国产精品一区二区精品视频观看| 国产欧美日韩一区二区精品| 日本与韩国留学比较| 校园春色视频在线观看| 日韩欧美三级三区| 国产亚洲精品综合一区在线观看| 91久久精品国产一区二区成人 | 一区二区三区高清视频在线| 男人和女人高潮做爰伦理| 国产亚洲精品久久久久久毛片| 久久久久国内视频| 国语自产精品视频在线第100页| 夜夜爽天天搞| 欧美精品啪啪一区二区三区| 麻豆av在线久日| 悠悠久久av| 色视频www国产| 色噜噜av男人的天堂激情| 久久热在线av| 国产精品精品国产色婷婷| 老熟妇乱子伦视频在线观看| 麻豆国产av国片精品| 人人妻人人澡欧美一区二区| 高清在线国产一区| 亚洲av日韩精品久久久久久密| 亚洲va日本ⅴa欧美va伊人久久| 免费看十八禁软件| 久久久久久九九精品二区国产| 国产成人欧美在线观看| 久久中文看片网| 小蜜桃在线观看免费完整版高清| cao死你这个sao货| 999精品在线视频| 国产97色在线日韩免费| 1024香蕉在线观看| 日韩欧美在线乱码| 中亚洲国语对白在线视频| 成人av一区二区三区在线看| 欧美一级毛片孕妇| 国产成年人精品一区二区| 91九色精品人成在线观看| 母亲3免费完整高清在线观看| 1000部很黄的大片| 99久久综合精品五月天人人| 中文字幕人妻丝袜一区二区| 一本综合久久免费| 男插女下体视频免费在线播放| 久久久久亚洲av毛片大全| 精品99又大又爽又粗少妇毛片 | 亚洲国产色片| 久久久久久九九精品二区国产| 在线看三级毛片| 婷婷六月久久综合丁香| 精品久久久久久久人妻蜜臀av| 欧美激情在线99| 黄色 视频免费看| 欧美中文日本在线观看视频| 88av欧美| 国产成+人综合+亚洲专区| 国产成人影院久久av| 又粗又爽又猛毛片免费看| 精品久久久久久成人av| 19禁男女啪啪无遮挡网站| 两性午夜刺激爽爽歪歪视频在线观看| 成年人黄色毛片网站| 日韩成人在线观看一区二区三区| 日本一本二区三区精品| 日本免费a在线| 精品国产超薄肉色丝袜足j| 久久久久性生活片| 1024香蕉在线观看| 亚洲一区二区三区不卡视频| 国产亚洲av嫩草精品影院| 精品久久久久久,| 91老司机精品| 91在线精品国自产拍蜜月 | 床上黄色一级片| 岛国在线观看网站| 久99久视频精品免费| 国产又色又爽无遮挡免费看| 日日摸夜夜添夜夜添小说| 国产精品久久久久久人妻精品电影| 三级男女做爰猛烈吃奶摸视频| 亚洲五月婷婷丁香| 国内少妇人妻偷人精品xxx网站 | 亚洲 国产 在线| 在线看三级毛片| 国产成人系列免费观看| 校园春色视频在线观看| 国产伦精品一区二区三区视频9 | 一级作爱视频免费观看| 久久久久久人人人人人| 国产精品美女特级片免费视频播放器 | 亚洲av美国av| 精品国内亚洲2022精品成人| 久久精品综合一区二区三区| 9191精品国产免费久久| 黄频高清免费视频| 亚洲色图av天堂| 99riav亚洲国产免费| 每晚都被弄得嗷嗷叫到高潮| 三级毛片av免费| 村上凉子中文字幕在线| 亚洲av电影不卡..在线观看| 国产蜜桃级精品一区二区三区| 真实男女啪啪啪动态图| 国产一区二区在线av高清观看| 91在线观看av| 欧美成人性av电影在线观看| 国产乱人伦免费视频| 桃红色精品国产亚洲av| 男人舔女人的私密视频| 午夜福利成人在线免费观看| 国产精品久久久人人做人人爽| 观看免费一级毛片| 天堂√8在线中文| 亚洲中文日韩欧美视频| 免费观看人在逋| 国产黄片美女视频| 毛片女人毛片| 男人舔奶头视频| 精品久久蜜臀av无| 99久久精品热视频| 亚洲av成人精品一区久久| 亚洲中文av在线| 搞女人的毛片| 欧美xxxx黑人xx丫x性爽| 麻豆久久精品国产亚洲av| 99久久国产精品久久久| 精品久久久久久,| 日本在线视频免费播放| 十八禁网站免费在线| 亚洲成av人片在线播放无| 最近最新中文字幕大全免费视频| 九九在线视频观看精品| 最近在线观看免费完整版| 亚洲精品一区av在线观看| 亚洲国产中文字幕在线视频| 国产成年人精品一区二区| 日韩欧美在线乱码| 亚洲av熟女| 亚洲电影在线观看av| 无人区码免费观看不卡| 国内少妇人妻偷人精品xxx网站 | 最新在线观看一区二区三区| 欧美在线黄色| 午夜久久久久精精品| 变态另类丝袜制服| 国产伦在线观看视频一区| 亚洲av熟女| 一级黄色大片毛片| 久久久国产成人免费| 亚洲欧美日韩无卡精品| 日本在线视频免费播放| 国产高清视频在线观看网站| 亚洲av成人一区二区三| 99国产精品一区二区蜜桃av| 一区二区三区激情视频|