• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Interactive medical image segmentation with self-adaptive confidence calibration?#

    2023-09-21 06:31:08ChuyunSHENWenhaoLIQisenXUBinHUBoJINHaibinCAIFengpingZHUYuxinLIXiangfengWANG

    Chuyun SHEN ,Wenhao LI ,Qisen XU ,Bin HU ,Bo JIN,Haibin CAI,Fengping ZHU,Yuxin LI,Xiangfeng WANG?

    1School of Computer Science and Technology, East China Normal University, Shanghai 200062, China

    2Huashan Hospital, Fudan University, Shanghai 200040, China

    3Software Engineering Institute, East China Normal University, Shanghai 200062, China

    Abstract: Interactive medical image segmentation based on human-in-the-loop machine learning is a novel paradigm that draws on human expert knowledge to assist medical image segmentation.However,existing methods often fall into what we call interactive misunderstanding,the essence of which is the dilemma in trading offshort-and long-term interaction information.To better use the interaction information at various timescales,we propose an interactive segmentation framework,called interactive MEdical image segmentation with self-adaptive Confidence CAlibration (MECCA),which combines action-based confidence learning and multi-agent reinforcement learning.A novel confidence network is learned by predicting the alignment level of the action with short-term interaction information.A confidence-based reward-shaping mechanism is then proposed to explicitly incorporate confidence in the policy gradient calculation,thus directly correcting the model’s interactive misunderstanding.MECCA also enables user-friendly interactions by reducing the interaction intensity and difficulty via label generation and interaction guidance,respectively.Numerical experiments on different segmentation tasks show that MECCA can significantly improve short-and long-term interaction information utilization efficiency with remarkably fewer labeled samples.The demo video is available at https://bit.ly/mecca-demo-video.

    Key words: Medical image segmentation;Interactive segmentation;Multi-agent reinforcement learning;Confidence learning;Semi-supervised learning

    1 Introduction

    Medical image segmentation is a crucial task in computer-assisted medical diagnosis.However,traditional convolutional neural networks (CNNs)(Wang et al.,2018,2019;Liao et al.,2020) used in segmentation algorithms often struggle to meet clinical standards due to the variability of pathological conditions,the presence of dark lesion areas,and the inconsistent quality of training data.Factors such as discrepancies in imaging scanners,operators,and annotators make the task of medical image segmentation more challenging.To improve accuracy,interactive image segmentation algorithms (Lin D et al.,2016;Xu et al.,2016;Rajchl et al.,2017;Bredell et al.,2018;Wang et al.,2018,2019;Liao et al.,2020;Ma et al.,2021) have been developed using interactive correction information,such as clicks,scribbles,or bounding boxes,to refine the results.

    Fig.1 illustrates the general interactive segmentation process,which contains two modules,i.e.,an interactive module and a utilization module.In the interactive module,users or experts provide interactive correction information,such as clicks and scribbles.The utilization module then uses this interactive correction information to efficiently refine the previous result.Interactive segmentation algorithms have the potential to outperform traditional algorithms by incorporating additional interactive correction information.The interaction process can be viewed as a continuous collaboration between the model and a human expert.Therefore,an effective interactive segmentation model should be able to incorporate and respond to expert feedback to improve its performance through collaboration.

    Fig.1 Interactive segmentation process.In the interactive module,the expert observes the current (or initial) segmentation and provides correction information (the red hint);in the utilization module,new segmentation is refined based on the correction information.References to color refer to the online version of this figure

    Methods such as BIFSeg (Wang et al.,2018) and DeepIGeoS (Wang et al.,2019) model user interactions as hard constraints through conditional random fields.However,these models focus more on one-step interaction,and refine segmentation after the first step cannot be efficiently used (Liao et al.,2020;Ma et al.,2021).Thus,these models cannot use long-term interaction information efficiently.The follow-up InterCNN (Bredell et al.,2018) and other methods (Liao et al.,2020;Ma et al.,2021) improve DeepIGeoS and BIFSeg,and model the problem as an iterative interaction problem,which is more focused on multi-step interaction.Therefore,these methods can cover the data distribution corresponding to different interaction levels in the training phase to effectively use long-term interactions,while ignoring the stochasticity or uncertainty of the model,which makes it difficult for them to use short-term interaction information effectively.

    Existing interactive methods cannot effectively use short-and long-term interaction information simultaneously,which leads to interactive misunderstanding.Fig.2 presents this phenomenon when implementing the popular interactive segmentation algorithm InterCNN (Bredell et al.,2018) on the BraTS2015 dataset.The algorithm ignores the expert’s correction information (Fig.2a),and is even adversely affected (Fig.2b).These inconsistencies between hint information and the refinement results indicate that existing algorithms still face the critical challenge of inefficient utilization of interactive correction information.The reason for the segmentation refining failure is that,at the end of the interaction process,the total loss of the model will guarantee the main area’s priority and ignore some small but challenging areas,such as edges.The regions that are insensitive to interactive correction information could be considered hard-to-segment regions (Shrivastava et al.,2016;Nie et al.,2019),leading to suboptimal refinement of the segmentation process.If these regions are not given additional attention during the training phase,the segmentation model may prioritize easy-to-segment areas at the expense of these more challenging regions.This problem is more severe for medical images,where the hard-tosegment regions usually are tumor boundaries and are very important for clinical diagnosis and surgery.Therefore,it has become urgent to improve the utilization of correction information,especially for the hard-to-segment regions.

    Fig.2 Segmentation refining failure after long-term interactions on BraTS2015 with InterCNN: (a) the model cannot fully understand or ignore the hint information;(b) the model could misunderstand the correction,which results in a worse result

    Another significant challenge in the development of interactive segmentation algorithms is the need for a large number of labeled images.Accurate annotation of medical images can be time-consuming and costly,making it difficult to obtain a sufficient quantity of high-quality data.On the other hand,it is much more cost-effective to obtain a larger number of unlabeled images,but existing interactive segmentation methods do not use these low-cost resources.As a result,it is important to explore how to use the benefits of unlabeled data to reduce the dependence on expert-annotated images.

    In this paper,we propose a novel interactive segmentation algorithm for three-dimensional (3D) medical images,called interactive MEdical image segmentation with self-adaptive Confidence CAlibration (MECCA).MECCA combines an action-based confidence learning network with the multi-agent reinforcement learning (MARL) framework (Zhang KQ et al.,2021).The action-based confidence network can evaluate the corrective action quality by directly calculating the confidence.Unlike learning the confidence of the overall segmentation result,the confidence of actions could be used to better evaluate whether the segmentation model has correctly used the experts’ interaction information.We formulate the iterative interaction process as a Markov decision process (MDP) to model the dynamic process and further introduce the MARL technique (Lee and Song,2018).Further,instead of setting each image or each patient (a series of images) as the agent,we consider each voxel (the smallest unit in the 3D space) as an agent,while each agent aims to learn the segmentation policy and makes its decision (Furuta et al.,2020).Specifically,after receiving the interactive correction information,each agent will modify its label by changing (increasing or decreasing) the category probability.The novel action-based confidence network will directly evaluate each agent’s corrective action obtained from the MARL corrective policy.Using this action-based confidence network,two additional techniques are further proposed to improve the utilization efficiency of the interactive correction information: (1) compared with former manual rewards,a self-adaptive reward function of each action is constructed which could provide more meticulous feedback under a flexible framework;(2) a simulated label generation mechanism is established using the interactive correction information as the weakly supervisory signal.Combining the confidence network,the simulated label generation mechanism can approximately generate the labels for unsupervised images and reduce over-reliance on labeled data.

    The main contributions of this work are summarized as follows:

    1.A novel framework is proposed for interactive medical image segmentation by combining the action-based confidence learning network with MARL.

    2.A self-adaptive feedback mechanism (selfadaptive reward) is constructed with the actionbased confidence network to alleviate interactive misunderstanding during the interaction process.

    3.Simulated supervisory signals can be generated based on the confidence learning network and actions;hence,much fewer labeled data (ground truth data) are needed to achieve the same performance.

    2 Related works

    2.1 Deep interactive image segmentation

    Xu et al.(2016) segmented images based on CNN interactively.DeepCut (Rajchl et al.,2017) and ScribbleSup (Lin D et al.,2016) both employ weak supervision to establish interactive image segmentation methods.DeepIGeoS (Wang et al.,2019) employs the geodesic distance metric to construct a hint map.The interactive segmentation process can be considered a sequential process which is natural to model by reinforcement learning (RL).Polygon-RNN (Castrejón et al.,2017) fundamentally segments each target as a polygon and iteratively chooses the polygon vertexes through a recurrent neural network (RNN).Polygon-RNN++(Acuna et al.,2018) employs almost the same idea as Polygon-RNN but learns to choose vertexes by RL.SeedNet (Lee and Song,2018) trains an expert interaction generation RL model that obtains new simulated interaction information at each interaction step.IteR-MRL (Liao et al.,2020) and BS-IRIS (Ma et al.,2021) both model the dynamic interaction process as an MDP and employ MARL models to segment images.Some researchers aimed to reduce the annotation cost of interactive image segmentation.IFSL (Feng et al.,2021) introduces interactive learning into the few-shot learning strategy and addresses the annotation burden of medical image segmentation models.IOG (Zhang SY et al.,2020) uses a practical inside-outside guidance approach to minimize the labeling cost.These interactive methods can hardly use experts’short-and long-term interaction information simultaneously,thus creating error correction operations.

    2.2 Uncertainty image segmentation

    Uncertainty estimation is helpful in the context of deployed machine learning systems because it can detect when a neural network is likely to make an incorrect prediction or when the input may be out of distribution.Recent methods generally adopt the neural network to learn what the uncertainty should be for any given input,i.e.,learningbased uncertainty estimation or confidence learning,as demonstrated in Kendall and Gal (2017),Moeskops et al.(2017),DeVries and Taylor (2018b),Hung et al.(2018),Robinson et al.(2018),Jungo and Reyes (2019),and Nie et al.(2019).These methods commonly consist of a segmentation network and a confidence network,and are more computationally efficient than other techniques.Thus,they are better suited when computational resources are limited or when real-time inference is required,such as the interactive segmentation scenario considered in this paper.Specifically,Kendall and Gal (2017) introduced a confidence network to predict the aleatoric and epistemic uncertainties by imitating classic Bayesian tools.The segmentation network of DeVries and Taylor (2018a,2018b) produces two separate outputs: prediction probabilities and a confidence estimate.Confidence estimate is motivated by interpolation between the predicted probability distribution and the target distribution during training,where the degree of interpolation is proportional to the confidence estimate.A series of works (Moeskops et al.,2017;Hung et al.,2018;Nie et al.,2019) focus on incorporating uncertainty estimation in the adversarial learning process,where the segmentation network corresponds to the generator,and the confidence network is the discriminator accordingly.Moeskops et al.(2017) first employed generative adversarial networks (GANs) to improve the CNN-based brain magnetic resonance imaging (MRI) segmentation method.The semi-supervised learning technique was used in Hung et al.(2018) to predict trustworthy regions in unlabeled images.Nie et al.(2019) proposed a difficulty-aware attention mechanism to handle those difficult samples or challenging regions.Different from the previous work on learning uncertainty through imitation,joint training,or adversarial learning,a simple but powerful alternative is to introduce an auxiliary task,such as predicting the overlap between a proposed segmentation and its ground truth (Robinson et al.,2018),or to predict the voxel-wise false positive and false negative (Jungo and Reyes,2019).

    2.3 Remark

    In our proposed algorithm,the confidence network should evaluate the confidence of calibrating actions instead of the segmentation results,which is the most significant difference between MECCA and previous uncertainty estimation methods.Therefore,we design a novel action-oriented auxiliary task to predict whether the direction of voxel-wise action is consistent with the ground truth.

    3 Proposed algorithm

    This section introduces the proposed interactive segmentation algorithm,MECCA,which can iteratively evaluate the refinement actions and feedback to the segmentation model.The algorithm framework follows the MARL structure (Sections 3.1 and 3.2),and an action-based confidence learning module (Section 3.3) is introduced to evaluate the confidence of corrective actions.This action-based confidence learning module can be used to establish the self-adaptive reward scheme and simulated label generation mechanism to use the interactive correction information efficiently.The architecture overview of MECCA is depicted in Fig.3a.The model’s state information includes the original 3D image,the previous segmentation probability,and the hint map generated from interaction and confidence maps.Based on the current state information,the segmentation module gives suggested actions to refine previous segmentation results by adjusting the segmentation probability of each voxel (agent).Further,the state information will be used to evaluate the confidence of the obtained actions through the confidence network,with a confidence map as the output.The self-adaptive reward is designed through a self-adaptive weighting scheme based on the action confidence evaluation (Section 3.4).The self-adaptive reward map can be considered a value map with the same size as the original input image (each agent has its self-adaptive reward).It can reflect the performance of the corresponding agent’s action.In addition,MECCA will suggest some lowconfidence regions for the user to segment during the interaction (Section 3.5).The confidence map can also be used to generate the simulated label by comparison with actions,which will be described in Section 3.6.The newly obtained hint map,the adjusted segmentation probability result,and the original 3D image form a new state.The process described above is repeated until the segmentation result meets the requirements.To emphasize,during the testing stage in Fig.3b,there is no need to calculate the selfadaptive rewards,and at the same time,we need only the obtained actions and suggested interaction areas.

    Fig.3 The architecture (a) and testing stage (b) of MECCA.In (a),the segmentation module outputs actions to change the segmentation probability of each voxel (agent) at each interaction step.Meanwhile,the confidence network will estimate the confidence of the actions,which will generate the self-adaptive reward and simulated label.The confidence map can provide the advice regions of the next interaction step to experts.In (b),all voxels in the segmentation probability map are initialized at a fixed value (set to 0.5 in the experiment).Users or experts randomly mark hint points according to the initialized segmentation probability map

    3.1 MARL-driven interactive segmentation

    We employ the MARL structure to formulate the interactive segmentation process and continuously give error-corrective actions at each interaction step.Letx=(x1,x2,···,xN) denote the input image,wherexi(i=1,2,···,N) denotes theithvoxel of the image.In MARL setting,every voxelxiis treated as an agent with its own refinement policyAt time stept,agentxiobtains actionfrom the segmentation network according to its current state.After taking the action,the agent will receive a rewardaccording to the segmentation result.

    3.2 Segmentation network

    The segmentation network adopts P-Net (Wang et al.,2019) as the backbone.It has two heads: a policy head and a value head.Two heads share the first three 3D convolutional blocks to extract low-level features.Each of the blocks has two convolution layers,and the size of the convolution kernel is fixed as 3×3 in all these convolution layers.Dilated convolution is employed in all the convolution kernels,which can reduce the loss of resolution.Both heads have another two 3D convolutional blocks to extract specific high-level features.Fig.4 shows the detailed architecture.Specifically,the policy head outputs policy,which is the distribution of action probabilities under the current state.By taking actions on different scales,the probabilities will be dynamically adjusted.The value head estimates the value of the current state,reflecting how good the current state is and estimating the expected return:

    Fig.4 Architecture of the segmentation network.The confidence network is the same as the value branch of the segmentation network,but the parameters are not shared

    whereTis the terminal time step in the interaction process,andγis the discount factor.θvdenotes the parameters of the value head,and the gradient with respect toθvis

    The policy head’s goal is to maximize the expected return by selecting proper actions in states(t).We useθpto denote the parameters of the policy head,and the gradient forθpis denoted as

    Usually,the policy head is updated more slowly.

    3.3 Action confidence learning

    As we mentioned in Section 1,there will be some situations where the segmentation model misunderstands or ignores the hint information.To some extent,these samples (or regions) with the phenomenon of interactive misunderstanding are hard samples (or regions).Although these samples may account for a small portion of the dataset,they are critical for improving generalization and robustness.The most important thing is finding a professional easy-or-hard representer (Nie et al.,2019) to identify them during the interaction.Focal loss (Lin TY et al.,2017) evaluates the easy-or-hard samples through a predicted probability.Nie et al.(2019) applied adversarial learning to train the easy-or-hard representer.Each representer has its advantages,but all evaluate easy-or-hard samples based on the final segmentation result.As such,these methods cannot be directly applied to interactive segmentation.For example,if the model predicts the category probability of a voxel to be 0.8,and then takes an actiona(t)i=-0.1,the next prediction will be 0.7 after interactionIfyi=1,the results are both correctly predicted becauseare both larger than 0.5,whereas for interactive segmentation,the probability is changing in the wrong direction.This change is what we call the interactive misunderstanding phenomenon,and the formal definition is shown in the following:

    Definition 1(Interactive misunderstanding) For a binary classification problem,the sign of the foreground labely=1 is denoted as positive,and the sign of the background labely=0 is denoted as negative.In an interactive medical image segmentation task (i.e.,a voxel-wise binary classification problem),for any voxeli,if the sign of the change of the segmentation probability outputted by the algorithm for two consecutive interaction steps,sign (Δ(p(i))),is not equal to sign (yi),then this phenomenon is defined as interactive misunderstanding.

    The confidence network learns the confidence of the given actions to avoid misunderstanding of hint information and to take accurate actions.We argue that confidential information can be used to regularize action choices and suggest more efficient interaction.The confidence network structure also uses P-Net as the backbone.The confidence network contains six 3D convolutional blocks.Each of the blocks has two convolution layers,and the size of the convolution kernel is fixed as 3×3 in all these convolution layers.The detailed architecture is also shown in Fig.4.

    The confidence network is trained using the previous state and action as input and a confidence map as output.The confidence network is optimized by minimizing the summation of binary cross-entropy loss over actions (shown in Eq.(6)) at each time stept.Here we useCto denote the confidence network,wCthe confidence network parameters,andLBCEthe binary cross-entropy loss:

    andg(t)indicates whether the direction of the action is consistent with the label.Fora ⊕b,the statement is true only if eithera>0 orb>0.

    One potential issue when training the confidence network is the imbalance of samples.Early in RL training,the error rate of actions outputted by the segmentation network is high,whereas most actions are correct when the network gradually converges.Inspired by discriminator training in GANs,this study introduces symmetric samples into Eq.(6) to speed up training.An obvious advantage is that it improves sample utilization efficiency because the confidence network can know what a“bad sample” is and obtain a corresponding“good sample.”

    3.4 Self-adaptive reward

    Essentially,no matter whether focal loss or adversarial learning is applied to train the segmentation model,the objective is to weigh these hard samples to prevent the model from being dominated by easy samples.However,the interactive segmentation task differs from those fully automatic segmentation tasks because the interactive segmentation model needs to cooperate with the user and understand the user’s hint information.The action to refine the segmentation result shows how the segmentation model understands hint information.Therefore,it is necessary to ensure that the hint information is correctly understood and that the correct action is taken.

    Specifically,the previously described actionconfidence learning can provide the segmentation model with a confidence map to alleviate interactive misunderstanding.Using this confidence map,easy-or-hard samples can be better recognized because the confidence values for these “hard regions”are lower than those for other regions.We formulate this action-aware learning as the self-adaptive reward function,r(t),which is shown in Eq.(8),to adapt this mechanism to MARL training:

    3.5 Interaction guide

    Another challenge for interactive image segmentation is that users usually need to decide where to interact in many slices,which is very timeconsuming,especially for 3D images.Our framework provides users with an interaction guide mechanism to save user interaction time.After refinement,our framework will suggest some possible areas for users to segment during the interaction.Specifically,our framework will filter out those areas with low action confidence and provide them to users (Fig.5).First,the original 3D image will be segmented with super voxels;each super voxel can be regarded as a group of voxels that share common characteristics.We use a simple linear iterative clustering (SLIC) (Achanta et al.,2012) technique with spacing=[2,2,2],compactness=0.1 to generate super voxels,and the number of super voxels equals 100 at the beginning and gradually declines during the refinement iterations for training and testing.Second,the proposed algorithm will compute the mean action confidence in each super voxel and rank them in descending order.Finally,the top five super voxels will be marked and recommended to users.Users need to select the best interaction positions from these super voxels.

    Fig.5 Interaction guide mechanism: (a) probability map;(b) action map;(c) confidence map;(d) advice region;(e) interaction point.The areas surrounded by the green lines are advice regions provided to users,and the red point in (e) is the real hint information selected from advice regions.The brighter the color (closer to yellow),the larger the positive value;conversely,the darker the color (closer to black),the smaller the negative value.References to color refer to the online version of this figure

    3.6 Simulated label generation

    In medical imaging,there are much more unlabeled data than labeled data due to difficulties in labeling medical images.To address this lack of annotations,the proposed algorithm leverages the action confidence not only to improve the utilization efficiency of hint information but also to generate a simulated label (Fig.6) for unlabeled data.We define the simulated label as follows:

    Fig.6 Illustration of simulated label generation.The simulated label generation mechanism uses the confidence map and the action map to generate the simulated label.The confidence map is used to calibrate the action,and the direction of the calibrated action is the simulated label of each voxel

    whereM=I(max (c(t),1 -c(t)) >δ),andis the advantage (defined in Mnih et al.(2016)) at time steptof takinga(t)in the condition of states(t),indicating the actual accumulated reward without being affected by the state and reducing the gradient variance.A mask,M,is used to constrain the training of unlabeled data.The backward gradients of unlabeled data occur only when the action confidence exceeds thresholdδ(which is gradually increased during the training process).Unlike traditional pseudo-label training,the supervised signal does not come from the segmentation network but from the confidence network.These filtered data with hint information are more valuable and provide more accurate supervised signals.Generally,the training processes with labeled data and simulated labeled data are carried out simultaneously.When using the simulated label generation mechanism,the pseudocode is as shown in Algorithm 1.

    4 Experiments and results

    4.1 Datasets and implementation details

    To comprehensively evaluate our proposed method,we apply our algorithm to four 3D medical image datasets.All datasets are divided into two parts:Dtrain/Dtest.The details of these datasets are as follows: (1) BraTS2015: Brain Tumor Segmentation Challenge 2015(Menze et al.,2015) contains 274(234/40) multi-parametric MRI images (Flair,T1,T1C,T2) from brain tumor patients.In our task,we use only the Flair image and segment the whole brain tumor.(2) BraTS2020: Brain Tumor Segmentation Challenge 2020 (Menze et al.,2015) contains 285(235/50) multi-parametric MRI images (Flair,T1,T1C,T2) from brain tumor patients.In our task,we use only the Flair image and segment the whole brain tumor.(3) MM-WHS: Multi-Modality Whole Heart Segmentation (Zhuang and Shen,2016) contains 24(20/4) multi-modality whole heart images covering the whole heart substructures.In our task,we choose to segment the left atrium blood cavity.(4) Medical Segmentation Decathlon: This is a generalizable 3D semantic segmentation dataset containing different organ segmentation tasks (Simpson et al.,2019).We use the spleen and liver datasets,which provide 61 (41/20) and 106 (96/10) CT images,respectively.

    We implement our method with PyTorch (Paszke et al.,2019).The demo video of MECCA is available at https://bit.ly/mecca-demo-video.The segmentation and confidence networks are both initialized using the Xavier (Glorot and Bengio,2010) method,and the learning rate is initialized to 1e-4.Other parameters are set as follows:T=5,γ=0.95,α=0.8,β=1.The maskMranges from 0.85 to 0.99 and increases by 0.000 25 at every epoch.Adam (Kingma and Ba,2015) is adopted as the optimizer.The original image is cropped by the bounding box based on the ground truth with a random extension in the range of 1-11 voxels.Each image is then resized and normalized to 55×55×30.The data are augmented by random flips and rotations.The proposed algorithm’s training time with one Nvidia 2080ti GPU varies from 5 to 13 h for different datasets.

    4.2 Experimental settings

    During the process of interaction,we adopt the edge points as the hint information for two reasons.On one hand,click operation saves more time than scribbles or other interactive methods.On the other hand,edge points can provide more information about object edges because the bottleneck of medical image segmentation is usually the inaccurate segmentation of object edges.Specifically,we provide each method with 45 points during the whole interaction process.At every step,users will click some edges that are not correctly predicted.We generate a 3D Gaussian (with a kernel size of eight voxels) centered on each of the edge points as the hint map to let networks receive hint information.Then the hint map will be inputted to the segmentation network as part of the state.We use the Dice score and the average symmetric surface distance (ASSD) to evaluate the segmentation results.According to these evaluation metrics,doctors can judge the patient’s condition:

    whereSpandSgdenote the prediction of an algorithm and the ground truth,respectively.

    whereSaandSbrepresent the sets of surface points of the segmentation results predicted by the algorithm and the ground truth,respectively.d(i,Sb) is the shortest Euclidean distance betweeniandSb.The Dice and ASSD in all tables are the average values of five algorithm test results.

    4.3 Comparisons with baselines

    We compare MECCA with four state-of-the-art interactive segmentation methods: DeepIGeoS (Wang et al.,2019),InterCNN (Bredell et al.,2018),IteR-MRL (Liao et al.,2020),and BS-IRIS (Ma et al.,2021).InterCNN is the multi-step version of DeepIGeoS.We also introduce a state-of-the-art medical image segmentation method,U-Net (Ronneberger et al.,2015),as a baseline.Table 1 shows the quantitative comparison of the seven segmentation methods on different datasets.For a fair comparison,except U-Net,all CNN-based methods adopt the same network structure (P-Net),which was proposed in Wang et al.(2019).In particular,U-Net and P-Net are methods without hint information.We can see that MECCA performs better than or is on par with the state-of-the-art methods on three datasets.We visualize the results in Fig.7,which shows that our method achieves more accurate results in edge segmentation.

    Table 1 Quantitative comparison of different methods

    Fig.7 Visualization of segmentation results by different methods.The green lines represent the boundaries of the ground truth,and the yellow lines represent the predicted boundaries.References to color refer to the online version of this figure

    To demonstrate that MECCA can take advantage of hint information more efficiently,we compare the relative improvement of different methods at each interaction step.Due to the different interaction processes of different methods,we set up a unified interaction process: all these methods have five interaction steps and receive 25 identical points at the first step to generate the initial segmentation result.After that,users will click five points at each step in the following four interaction steps (a total of 25+5×4=45 points),and these methods will iteratively refine previous segmentation results.The DeepIGeoS method does not model the interaction sequence,and simply combines current and previous hint information to refine the previous segmentation.The results are shown in Table 2 and Fig.8.

    Table 2 Dice scores at each interaction step by different methods

    Fig.8 The performance improvement of different interactive medical segmentation methods at different interaction steps.All these testing results were obtained on the BraTS2015 dataset

    The results show that MECCA performs better under the same amount of hint information and improves the Dice score more after the five steps.Compared with CNN-based methods,the main advantage of RL-based methods is that they can always maintain notable improvement.There are two reasons for this result.The first one is that RLbased methods model the whole interaction process to avoid interaction conflict.The second reason is the relative entropy-based reward which encourages the model to keep refining results.However,we should realize that the RL-based methods still cannot guarantee the high confidence of the corrective actions.As we can see in Fig.8,the performance of IteRMRL is not as good as those of other methods at the beginning,which is caused by numerous incorrect actions.However,after learning the confidence of actions and applying the self-adaptive reward to update the segmentation model,our proposed MECCA can perform well at each step and continue significantly refining the result.

    4.4 Comparison of different rewards

    One main contribution of our work is that the self-adaptive reward mechanism is further proposed based on the confidence map.This mechanism makes different actions have different feedback levels,so the segmentation network can identify wrong actions as much as possible.To measure the impact of different means of reward weighting on the segmentation result,we compare MECCA with the original IteR-MRL without a weighting reward,Focal-Reward IteR-MRL (Focal-IteR-MRL),and the one with a weighted reward in segmentation error regions (Err-IteR-MRL).Focal-IteR-MRL adopts the idea of focal loss (Lin TY et al.,2017):

    The reward function of Err-IteR-MRL is a linear scaling of the IteR-MRL reward:

    whereλiis a positive real number hyperparameter.The performance of different weighting rewards is shown in Fig.9.It can be seen that the actionbased reward weighting method is more suitable for interactive segmentation.Also,this study visualizes different masks that weigh the primary reward in Fig.10.As we can see,the focal mask pays more attention to regions with low prediction probabilities,and the error region mask contains all regions segmented incorrectly in the previous result.It is not enough for the focal mask to obtain more structured information by considering only the probability of prediction.As such,the performance of Focal-IteRMRL is erratic,even the poorest on the left atrium dataset.The performance of Err-IteR-MRL is more stable,but the improvement is not notable.This is because the weighting process of Err-IteR-MRL is based on the segmentation result while the refinement process of interactive image segmentation is based on actions.

    Fig.9 Average Dice scores of methods with different weighting rewards.The reward function of IteR-MRL is not weighted;the reward function of Focal-IteRMRL is weighted via Eq.(12),the reward function of Err-IteR-MRL is weighted in segmentation error regions,and the reward function of MECCA is weighted through action confidence

    Fig.10 Visualization of the original image (a),ground truth (b),action map (c),self-adaptive mask (d),error region mask (e),and focal mask (f).All masks are obtained at the first interaction step.The brighter the color (closer to yellow),the larger the positive value;conversely,the darker the color (closer to black),the smaller the negative value.References to color refer to the online version of this figure

    4.5 Weakly supervised segmentation

    As mentioned in Section 3.6,MECCA can reduce the dependence on voxel-level annotations of images using the simulated label generated from the action confidence map.This study validates MECCA on the BraTS2015 dataset by randomly selecting different proportions of samples as fully labeled data and using the rest of the training images as unlabeled data,which provide only hint information during the interaction.This study compares MECCA against P-Net (without hint),DeepIGeoS,and IteR-MRL (Liao et al.,2020).However,only MECCA uses unlabeled data,and the three baselines use only a fixed proportion of labeled data.This does not fully demonstrate the performance of MECCA on semi-supervised problems.For this reason,we introduce a state-of-the-art semi-supervised method UA-MT (Yu et al.,2019).Table 3 shows the results of different methods.MECCA achieves a Dice score of 87.14% with only 12.5% labeled data,and it performs better than the other methods with 25%labeled data.

    Table 3 Quantitative comparison between MECCA and other methods on the BraTS2015 dataset with different portions of labeled data

    Because UA-MT is not an interactive segmentation algorithm,its absolute performance is worse than those of interactive segmentation baselines but better than that of the non-interactive method PNet.In addition,it can be seen from the last column of Table 3 that UA-MT has the most negligible performance loss with different proportions of labeled data.After MECCA introduces the simulated label generation mechanism,the performance loss is similar to that of UA-MT.Interestingly,DeepIGeoS can maintain low performance loss without using a semi-supervised learning mechanism.Considering the performance loss of IteR-MRL,we can conclude that in the semi-supervised interactive segmentation task,the interactive misunderstanding phenomenon will exacerbate the performance loss caused by missing data.DeepIGeoS and MECCA alleviate the interactive misunderstanding phenomenon through hard constraints and self-adaptive confidence calibration,respectively,so both of them can achieve better results in semi-supervised settings.However,note that DeepIGeoS does not consider multi-step interactions or the relationship between consecutive interactions.It cannot fully use long-term interaction information,so there is a large gap with MECCA in absolute performance.

    We also evaluate MECCA with different numbers of interactions and percentages of labeled data.Based on the experimental results (Table 4),we conclude the following: (1) if there are enough labeled data,MECCA can achieve good results after a few interactions;(2) the number of interactions required by MECCA to achieve the same performance is roughly inversely proportional to the portion of labeled data;(3) although MECCA requires more interactions (about 2-3 times) when only part of the labeled data is available,it can approach the performance of the algorithm trained with all data being labeled in the end.

    Table 4 Dice scores with different numbers of interactions and percentages of labeled data

    4.6 Ablation studies

    Table 5 shows the impact of different mechanisms on MECCA.It is clear that the simulated label generation mechanism and self-adaptive reward mechanism significantly improve the algorithm’s performance.The simulated label generation mechanism can achieve a 1.70% and 2.26% gain over the initial algorithm without any mechanism for the BraTS2015 dataset and the Liver dataset in Medical Segmentation Decathlon,respectively.The main reason for this significant improvement is the utilization of these unlabeled data.

    Table 5 Ablation study of our proposed algorithm on BraTS2015 and Liver datasets with 25% labeled data

    Table 6 shows the comparison of interactive segmentation methods in computational and interaction time.The results show that MECCA takes only about half the time of other methods when performing an interaction.The main contribution to the reduction of interaction time is the interaction guide mechanism mentioned.Furthermore,to explore the efficiency of MECCA in authentic tasks,we design an interactive software platform.We ask oncologists to use it for natural interactive segmentation,and the results are shown in Table 7.It can be seen from the results that the MECCA algorithm also takes only about half the time of the baselines due to interaction guidance.In addition,as the segmentation results become more and more accurate,the interaction time required gradually decreases.

    Table 6 Comparison of computational and interaction time among different interactive segmentation methods

    Table 7 Quantitative comparison of interaction time among different interactive segmentation methods with different numbers of interactions on a realistic segmentation platform operated by experts

    5 Conclusions

    We present a novel action-based confidence learning method for interactive 3D image segmentation.Specifically,we propose a method for learning the confidence of actions that continuously refine the segmentation result during the interaction process,so that hint information can be used more effectively.Based on this process,a self-adaptive reward is proposed for the segmentation module,which can prevent the misunderstanding phenomenon during the interaction and help reduce the time cost of interaction by providing users with the advice regions with which they should interact next.In addition,the confidence map can replace the ground truth to generate feedback for the unlabeled samples for the segmentation module.These samples,without voxel-level annotations,can also be used to model training.By integrating these components,MECCA is demonstrated to improve medical segmentation efficiently using fewer annotated samples.

    Although our method can greatly enhance the utilization of short-and long-term interaction information with significantly fewer labeled samples,there are also certain limitations that we need to address.One of the primary challenges in medical image segmentation is the high dimension and resolution of medical images,leading to a requirement for numerous interactive clicks.Specifically,Wang et al.(2018) provided the actual use time collected from real physicians.Thus,there is a need for more efficient interactive methods,and Aljabri et al.(2022) introduced the currently available annotation tools for medical imaging.In future work,we may incorporate more efficient annotation methods to minimize the cost of physician interactions.

    Further,RL from human feedback (RLHF),which is used in ChatGPT (OpenAI,2022),is increasingly popular.This approach collects comparison data and trains a reward model,where the data are human rankings of the model’s outputs from best to worst.RLHF is another application that benefits from human feedback,and in future work,we may consider another interactive method in which radiographers rank the different outputs as reward signals.Another limitation in the field of medical image segmentation is the presence of multiple centers,where data are collected from different devices,resulting in varying distributions.With such non-independent and identically distributed (non-IID) data,the generalization ability of deep learning based models is limited (Li et al.,2021;Ye et al.,2022).To overcome this,our proposed MECCA method may require additional design;e.g.,Aljabri et al.(2022) introduced a novel robust weakly supervised learning paradigm.

    Finally,as technology continues to advance,changes in patient distribution and the introduction of new medical equipment may necessitate the development of online learning methods to ensure that the algorithm remains robust.Popular methods such as iCaRL (Rebuffiet al.,2017),GDumb (Prabhu et al.,2020),and iCaRL (Rebuffiet al.,2017) maintain a buffer to store the most important samples for the model.Specifically,in the RL area,we can borrow the ideas and theoretical justifications from lifelong RL,in which setting agents must solve a series of related tasks drawn from a task distribution rather than a single,isolated task (Abel et al.,2018).Abel et al.(2018) explored which knowledge should be transferred in lifelong RL through two simple families of results.Xie et al.(2020) leveraged latent variable models to learn a representation of the environment from experience and performed off-policy RL with this representation.

    Contributors

    Chuyun SHEN,Wenhao LI,and Qishen XU designed the research and conducted the experiments.Bin HU,Fengping ZHU,and Yuxin LI ensured the validity of the experiments.Bo JIN,Haibin CAI,and Xiangfeng WANG offered support across various experimental aspects.Chuyun SHEN drafted the paper.All the authors revised and finalized the paper.

    Compliance with ethics guidelines

    Chuyun SHEN,Wenhao LI,Qisen XU,Bin HU,Bo JIN,Haibin CAI,Fengping ZHU,Yuxin LI,and Xiangfeng WANG declare that they have no conflict of interest.

    Data availability

    The demo video is available at https://bit.ly/meccademo-video.The other data that support the findings of this study are available from the corresponding author upon reasonable request.

    List of supplementary materials

    1 More related works

    2 More visualizations

    3 Robustness of MECCA

    4 Comparison of baseline responses to the same user interaction

    Fig.S1 MECCA segmentation process

    Fig.S2 Qualitative segmentation results of MECCA for the BraTS2015 validation set

    Figs.S3-S5 Results of different methods’ responses to the same user interactions according to the same initial segmentation on different testing instances and different channels for the Liver dataset in Medical Segmentation Decathlon

    Table S1 Dice of our method which varies with the number of interactions under different cases

    Table S2 MECCA’s tolerance to inaccurate interaction points

    欧美中文综合在线视频| xxx96com| 变态另类成人亚洲欧美熟女 | 日韩三级视频一区二区三区| 午夜免费激情av| 老司机午夜福利在线观看视频| 色播在线永久视频| 人人妻,人人澡人人爽秒播| 免费在线观看完整版高清| 亚洲aⅴ乱码一区二区在线播放 | 免费高清视频大片| 日韩大码丰满熟妇| 91老司机精品| 高清欧美精品videossex| 中文字幕精品免费在线观看视频| 超碰成人久久| 丰满饥渴人妻一区二区三| 搡老岳熟女国产| 久久久水蜜桃国产精品网| 免费女性裸体啪啪无遮挡网站| 51午夜福利影视在线观看| 久久精品国产清高在天天线| 一级a爱片免费观看的视频| 啦啦啦免费观看视频1| 天堂√8在线中文| 久久99一区二区三区| 久久午夜亚洲精品久久| 老司机午夜福利在线观看视频| 久久热在线av| 国产精品野战在线观看 | 亚洲成人免费av在线播放| 亚洲成人国产一区在线观看| 后天国语完整版免费观看| 制服人妻中文乱码| 侵犯人妻中文字幕一二三四区| av在线播放免费不卡| 51午夜福利影视在线观看| 中文字幕色久视频| 欧美精品啪啪一区二区三区| 国产精品久久久人人做人人爽| 国产成人精品无人区| 三级毛片av免费| 真人做人爱边吃奶动态| 每晚都被弄得嗷嗷叫到高潮| 亚洲人成77777在线视频| 精品熟女少妇八av免费久了| 丝袜美足系列| 亚洲国产精品999在线| 国产亚洲精品久久久久5区| 日本wwww免费看| xxx96com| 国产深夜福利视频在线观看| 久久人妻熟女aⅴ| 国产单亲对白刺激| 亚洲精品一卡2卡三卡4卡5卡| 午夜a级毛片| 69av精品久久久久久| 欧美另类亚洲清纯唯美| 精品少妇一区二区三区视频日本电影| 欧美黄色片欧美黄色片| 麻豆国产av国片精品| 我的亚洲天堂| 女人被躁到高潮嗷嗷叫费观| 国产一区二区三区在线臀色熟女 | 在线视频色国产色| 免费搜索国产男女视频| 黄片播放在线免费| 精品国产亚洲在线| 久久精品人人爽人人爽视色| 国产成人欧美在线观看| 变态另类成人亚洲欧美熟女 | 女生性感内裤真人,穿戴方法视频| 黑人欧美特级aaaaaa片| 欧美大码av| netflix在线观看网站| 欧美激情久久久久久爽电影 | 一级片'在线观看视频| 69精品国产乱码久久久| 性少妇av在线| 欧美成人免费av一区二区三区| 亚洲男人的天堂狠狠| 成人亚洲精品av一区二区 | 亚洲av成人不卡在线观看播放网| 色婷婷久久久亚洲欧美| 又黄又爽又免费观看的视频| 久久国产精品影院| 女同久久另类99精品国产91| 18禁观看日本| 嫩草影视91久久| 日韩精品免费视频一区二区三区| 久久中文看片网| 99热只有精品国产| 精品久久久久久久毛片微露脸| 亚洲av第一区精品v没综合| 男人的好看免费观看在线视频 | 免费观看人在逋| 国产蜜桃级精品一区二区三区| 久久国产乱子伦精品免费另类| av视频免费观看在线观看| 久久精品人人爽人人爽视色| 99精国产麻豆久久婷婷| 狠狠狠狠99中文字幕| 老鸭窝网址在线观看| 精品熟女少妇八av免费久了| 午夜成年电影在线免费观看| 美国免费a级毛片| 亚洲国产毛片av蜜桃av| 性色av乱码一区二区三区2| 亚洲一卡2卡3卡4卡5卡精品中文| 中文字幕人妻熟女乱码| 欧美日韩福利视频一区二区| 久久天躁狠狠躁夜夜2o2o| 日韩高清综合在线| 久久久精品欧美日韩精品| 日本精品一区二区三区蜜桃| 日韩视频一区二区在线观看| 久久精品国产亚洲av高清一级| 后天国语完整版免费观看| 日韩欧美在线二视频| 亚洲欧美日韩高清在线视频| 9191精品国产免费久久| 精品卡一卡二卡四卡免费| 成人亚洲精品av一区二区 | 国产精品秋霞免费鲁丝片| 成人国语在线视频| 久久国产精品人妻蜜桃| 色婷婷久久久亚洲欧美| 老熟妇乱子伦视频在线观看| 欧美成人午夜精品| 在线观看免费视频日本深夜| 日本黄色视频三级网站网址| 精品人妻1区二区| 波多野结衣av一区二区av| 美女高潮喷水抽搐中文字幕| 久久久精品欧美日韩精品| 另类亚洲欧美激情| 欧美激情极品国产一区二区三区| 亚洲aⅴ乱码一区二区在线播放 | 精品久久久久久电影网| e午夜精品久久久久久久| 精品久久久久久电影网| 久久久久国产一级毛片高清牌| www国产在线视频色| 久久人妻av系列| 色播在线永久视频| 亚洲一区二区三区色噜噜 | 熟女少妇亚洲综合色aaa.| 在线观看66精品国产| 亚洲熟妇熟女久久| 免费在线观看亚洲国产| 黄片播放在线免费| 水蜜桃什么品种好| 国产99白浆流出| 成年版毛片免费区| 精品一区二区三区视频在线观看免费 | a级毛片在线看网站| 大香蕉久久成人网| 日本撒尿小便嘘嘘汇集6| 91大片在线观看| 一本综合久久免费| 99久久人妻综合| 19禁男女啪啪无遮挡网站| 亚洲七黄色美女视频| 这个男人来自地球电影免费观看| 黄频高清免费视频| 男女下面插进去视频免费观看| 亚洲色图 男人天堂 中文字幕| 精品日产1卡2卡| 黄色毛片三级朝国网站| 欧美日韩黄片免| 精品国产美女av久久久久小说| 制服诱惑二区| 国产av一区在线观看免费| 国产精品一区二区精品视频观看| 国产真人三级小视频在线观看| 精品国产一区二区三区四区第35| 色在线成人网| 啦啦啦免费观看视频1| 久久久国产精品麻豆| a级毛片在线看网站| 久久这里只有精品19| 成人18禁高潮啪啪吃奶动态图| 国产乱人伦免费视频| 一进一出好大好爽视频| 欧美在线一区亚洲| 欧美乱妇无乱码| 成年女人毛片免费观看观看9| 国产亚洲欧美98| 性少妇av在线| 精品午夜福利视频在线观看一区| 黄色丝袜av网址大全| 久久精品国产综合久久久| 嫩草影视91久久| av网站免费在线观看视频| 国产精品电影一区二区三区| 久久精品aⅴ一区二区三区四区| 69精品国产乱码久久久| 麻豆成人av在线观看| 国产一区二区激情短视频| 校园春色视频在线观看| 老熟妇仑乱视频hdxx| 69av精品久久久久久| 精品国内亚洲2022精品成人| 伊人久久大香线蕉亚洲五| 亚洲中文av在线| 真人一进一出gif抽搐免费| 首页视频小说图片口味搜索| 日韩有码中文字幕| www国产在线视频色| 黄色毛片三级朝国网站| 免费在线观看影片大全网站| 国产欧美日韩一区二区三| 免费女性裸体啪啪无遮挡网站| 别揉我奶头~嗯~啊~动态视频| 久久青草综合色| 国产在线观看jvid| 日韩大码丰满熟妇| 国产av一区在线观看免费| 9热在线视频观看99| 久久久久久免费高清国产稀缺| 亚洲精品中文字幕在线视频| 国产主播在线观看一区二区| 国产三级黄色录像| 在线观看日韩欧美| 国产激情欧美一区二区| 夜夜看夜夜爽夜夜摸 | 女性生殖器流出的白浆| 搡老乐熟女国产| 日本免费一区二区三区高清不卡 | 精品国产美女av久久久久小说| 亚洲av日韩精品久久久久久密| 少妇裸体淫交视频免费看高清 | 色哟哟哟哟哟哟| 色在线成人网| 久久久国产精品麻豆| 日本wwww免费看| 精品高清国产在线一区| 不卡一级毛片| 男女做爰动态图高潮gif福利片 | 中文字幕色久视频| 少妇粗大呻吟视频| 成人手机av| 欧美 亚洲 国产 日韩一| 老汉色av国产亚洲站长工具| 制服人妻中文乱码| 80岁老熟妇乱子伦牲交| 久久久久国产一级毛片高清牌| 成人av一区二区三区在线看| 免费人成视频x8x8入口观看| 亚洲自偷自拍图片 自拍| 久久精品影院6| 欧洲精品卡2卡3卡4卡5卡区| 精品国产超薄肉色丝袜足j| 亚洲熟妇中文字幕五十中出 | 亚洲一区二区三区不卡视频| 日韩免费高清中文字幕av| 日本五十路高清| 亚洲精品美女久久av网站| 91国产中文字幕| 国产精品久久电影中文字幕| 欧美日韩国产mv在线观看视频| 巨乳人妻的诱惑在线观看| 一级作爱视频免费观看| 97人妻天天添夜夜摸| 88av欧美| 女警被强在线播放| 91大片在线观看| 99国产精品一区二区三区| 热re99久久精品国产66热6| 嫩草影视91久久| 十八禁网站免费在线| 免费看十八禁软件| 婷婷丁香在线五月| 色老头精品视频在线观看| 日韩欧美一区二区三区在线观看| 大码成人一级视频| 男女午夜视频在线观看| 亚洲一卡2卡3卡4卡5卡精品中文| 国产精品一区二区精品视频观看| 亚洲aⅴ乱码一区二区在线播放 | 久久影院123| 欧美日韩瑟瑟在线播放| 久久久久久免费高清国产稀缺| 成人三级黄色视频| 亚洲人成77777在线视频| 久久久国产成人免费| 黄色a级毛片大全视频| 99久久久亚洲精品蜜臀av| 久久天躁狠狠躁夜夜2o2o| 欧美 亚洲 国产 日韩一| 久久精品人人爽人人爽视色| 国产亚洲欧美精品永久| 成人特级黄色片久久久久久久| 日本黄色日本黄色录像| 日本一区二区免费在线视频| 国产亚洲欧美精品永久| 色老头精品视频在线观看| 国产亚洲精品第一综合不卡| 天堂中文最新版在线下载| 成人国语在线视频| 亚洲av片天天在线观看| 亚洲精品久久成人aⅴ小说| 在线av久久热| 999久久久国产精品视频| 精品免费久久久久久久清纯| 精品久久蜜臀av无| 丝袜美足系列| 国产精品99久久99久久久不卡| 国产欧美日韩一区二区三区在线| 成人三级黄色视频| 国产av精品麻豆| 亚洲欧美精品综合久久99| 精品高清国产在线一区| 97碰自拍视频| 亚洲国产欧美一区二区综合| 精品福利观看| 日韩人妻精品一区2区三区| 老司机亚洲免费影院| 国产免费现黄频在线看| 麻豆一二三区av精品| 天堂动漫精品| 91大片在线观看| 男女做爰动态图高潮gif福利片 | www.熟女人妻精品国产| 精品一区二区三卡| 99精品久久久久人妻精品| 国产精品久久久久成人av| 欧美+亚洲+日韩+国产| 午夜福利,免费看| 日日干狠狠操夜夜爽| 国产成人一区二区三区免费视频网站| 午夜91福利影院| 国产真人三级小视频在线观看| 亚洲七黄色美女视频| 亚洲成国产人片在线观看| 精品国产亚洲在线| 如日韩欧美国产精品一区二区三区| 国产亚洲精品久久久久久毛片| 亚洲aⅴ乱码一区二区在线播放 | av天堂在线播放| 精品久久久久久久毛片微露脸| 99香蕉大伊视频| 久久热在线av| 亚洲人成网站在线播放欧美日韩| 欧美激情 高清一区二区三区| 免费在线观看黄色视频的| 热99国产精品久久久久久7| 丁香欧美五月| 亚洲aⅴ乱码一区二区在线播放 | 国产亚洲精品久久久久5区| 黑人操中国人逼视频| 91成年电影在线观看| 可以在线观看毛片的网站| 男女床上黄色一级片免费看| 亚洲成国产人片在线观看| 亚洲一区二区三区色噜噜 | 亚洲aⅴ乱码一区二区在线播放 | 精品国产国语对白av| 精品电影一区二区在线| 久久精品影院6| 大香蕉久久成人网| 日日夜夜操网爽| 好男人电影高清在线观看| 80岁老熟妇乱子伦牲交| 久久性视频一级片| 悠悠久久av| 欧美成狂野欧美在线观看| 精品国产超薄肉色丝袜足j| 1024视频免费在线观看| 亚洲一区二区三区不卡视频| 少妇被粗大的猛进出69影院| 日本免费a在线| 精品无人区乱码1区二区| 黑人操中国人逼视频| 国产男靠女视频免费网站| 国产伦人伦偷精品视频| 正在播放国产对白刺激| 黄网站色视频无遮挡免费观看| 免费在线观看完整版高清| 亚洲欧美激情综合另类| 最好的美女福利视频网| 久久久久国产一级毛片高清牌| 丁香欧美五月| 国产成人av教育| 精品福利观看| 免费在线观看完整版高清| 国产精品成人在线| xxx96com| 老汉色∧v一级毛片| 久久国产精品男人的天堂亚洲| 亚洲一码二码三码区别大吗| 国产精品亚洲av一区麻豆| 国产精品98久久久久久宅男小说| 一区福利在线观看| 久久九九热精品免费| 免费观看精品视频网站| 99国产精品免费福利视频| 精品国内亚洲2022精品成人| 色综合站精品国产| 99久久99久久久精品蜜桃| 亚洲人成77777在线视频| 国产亚洲精品一区二区www| 99国产综合亚洲精品| 中文字幕精品免费在线观看视频| 国产成人精品无人区| 亚洲欧美精品综合一区二区三区| a级毛片在线看网站| 一边摸一边做爽爽视频免费| 国产色视频综合| 成年版毛片免费区| 亚洲精品美女久久av网站| 亚洲avbb在线观看| 日本黄色视频三级网站网址| 久久天堂一区二区三区四区| 香蕉国产在线看| 国产1区2区3区精品| 国产深夜福利视频在线观看| 久久久久久久久久久久大奶| 脱女人内裤的视频| 天天躁夜夜躁狠狠躁躁| 狠狠狠狠99中文字幕| 国产亚洲精品综合一区在线观看 | 亚洲色图综合在线观看| 两个人免费观看高清视频| 女人精品久久久久毛片| 欧美人与性动交α欧美软件| 日韩欧美一区二区三区在线观看| 麻豆av在线久日| 国产日韩一区二区三区精品不卡| 女同久久另类99精品国产91| 女生性感内裤真人,穿戴方法视频| 免费在线观看黄色视频的| 亚洲av片天天在线观看| 天堂影院成人在线观看| 国产亚洲精品综合一区在线观看 | 精品国产乱子伦一区二区三区| 丰满的人妻完整版| 精品无人区乱码1区二区| 两性午夜刺激爽爽歪歪视频在线观看 | 窝窝影院91人妻| 好男人电影高清在线观看| 狠狠狠狠99中文字幕| √禁漫天堂资源中文www| 身体一侧抽搐| 国产成人精品无人区| 9热在线视频观看99| 18美女黄网站色大片免费观看| 在线观看www视频免费| 一级作爱视频免费观看| 久久国产乱子伦精品免费另类| 免费久久久久久久精品成人欧美视频| 欧美老熟妇乱子伦牲交| 国产亚洲欧美精品永久| 国产av在哪里看| 国产精品久久视频播放| 高清av免费在线| av免费在线观看网站| 久热这里只有精品99| 又大又爽又粗| 成人18禁高潮啪啪吃奶动态图| 丰满饥渴人妻一区二区三| 精品免费久久久久久久清纯| 熟女少妇亚洲综合色aaa.| 亚洲三区欧美一区| 俄罗斯特黄特色一大片| 岛国在线观看网站| 一边摸一边抽搐一进一小说| 亚洲专区中文字幕在线| 欧美黄色淫秽网站| 久久欧美精品欧美久久欧美| 高清毛片免费观看视频网站 | 国产日韩一区二区三区精品不卡| 日韩中文字幕欧美一区二区| 五月开心婷婷网| 亚洲国产中文字幕在线视频| 身体一侧抽搐| 色综合站精品国产| 中出人妻视频一区二区| 成人国产一区最新在线观看| 在线av久久热| 免费看a级黄色片| 国产精品亚洲av一区麻豆| 亚洲精品久久成人aⅴ小说| 淫妇啪啪啪对白视频| 91成年电影在线观看| 一边摸一边抽搐一进一出视频| 亚洲人成电影观看| 国产熟女xx| 国产精品日韩av在线免费观看 | 久久久久久亚洲精品国产蜜桃av| 国产精品亚洲一级av第二区| av网站免费在线观看视频| 午夜日韩欧美国产| 五月开心婷婷网| 日日夜夜操网爽| 黄片小视频在线播放| 国产国语露脸激情在线看| 动漫黄色视频在线观看| 露出奶头的视频| 亚洲成国产人片在线观看| 国产精品99久久99久久久不卡| 岛国视频午夜一区免费看| 精品午夜福利视频在线观看一区| 三上悠亚av全集在线观看| 久久性视频一级片| 757午夜福利合集在线观看| 日韩三级视频一区二区三区| 色尼玛亚洲综合影院| 欧美av亚洲av综合av国产av| 欧美丝袜亚洲另类 | 啦啦啦在线免费观看视频4| 久久国产乱子伦精品免费另类| 国产一区二区三区在线臀色熟女 | 亚洲国产精品999在线| 日韩高清综合在线| 国产熟女xx| 国产精品九九99| 国产av一区在线观看免费| av中文乱码字幕在线| 国产av又大| 不卡一级毛片| 成年女人毛片免费观看观看9| 日韩精品免费视频一区二区三区| 淫妇啪啪啪对白视频| 成人亚洲精品一区在线观看| 咕卡用的链子| 亚洲视频免费观看视频| 国产精品亚洲av一区麻豆| 少妇 在线观看| 天天躁狠狠躁夜夜躁狠狠躁| 狠狠狠狠99中文字幕| 日韩大尺度精品在线看网址 | 丝袜在线中文字幕| 日日干狠狠操夜夜爽| 女人爽到高潮嗷嗷叫在线视频| 99国产精品免费福利视频| 91麻豆av在线| 91成人精品电影| 久久热在线av| 黄色视频,在线免费观看| 国产亚洲av高清不卡| 精品一区二区三区四区五区乱码| 免费av毛片视频| 色综合婷婷激情| 一级,二级,三级黄色视频| 国产精品影院久久| 级片在线观看| 成年人免费黄色播放视频| 成人精品一区二区免费| 露出奶头的视频| 一级毛片高清免费大全| 国产精品二区激情视频| 国产激情久久老熟女| 一进一出抽搐gif免费好疼 | 一区二区三区国产精品乱码| 美女高潮喷水抽搐中文字幕| 国产亚洲欧美精品永久| av中文乱码字幕在线| 十八禁人妻一区二区| 一二三四在线观看免费中文在| 亚洲一卡2卡3卡4卡5卡精品中文| 日韩 欧美 亚洲 中文字幕| xxxhd国产人妻xxx| 母亲3免费完整高清在线观看| 三上悠亚av全集在线观看| 久久草成人影院| 欧美黑人精品巨大| 国产精品一区二区免费欧美| 日本精品一区二区三区蜜桃| 50天的宝宝边吃奶边哭怎么回事| 深夜精品福利| 黑人欧美特级aaaaaa片| 1024香蕉在线观看| 成人18禁在线播放| 波多野结衣av一区二区av| 夜夜夜夜夜久久久久| 亚洲伊人色综图| 最新在线观看一区二区三区| 国产野战对白在线观看| www国产在线视频色| 国产又爽黄色视频| 精品国内亚洲2022精品成人| 亚洲aⅴ乱码一区二区在线播放 | 亚洲国产看品久久| 成人国语在线视频| 亚洲少妇的诱惑av| 久久久久久亚洲精品国产蜜桃av| 亚洲 国产 在线| av福利片在线| 国产深夜福利视频在线观看| 中文字幕人妻熟女乱码| 999久久久精品免费观看国产| 桃红色精品国产亚洲av| 欧美成人午夜精品| 午夜精品在线福利| 波多野结衣高清无吗| 国产人伦9x9x在线观看| 免费观看人在逋| 99久久久亚洲精品蜜臀av| www.熟女人妻精品国产| 久久久国产精品麻豆| 高清欧美精品videossex| 亚洲国产欧美一区二区综合| 夜夜看夜夜爽夜夜摸 | 在线av久久热| 搡老岳熟女国产| 欧美成人免费av一区二区三区| 国产在线观看jvid| 夫妻午夜视频| 巨乳人妻的诱惑在线观看| 国产片内射在线| 欧美黑人欧美精品刺激| 亚洲精品av麻豆狂野| 国产一区二区三区在线臀色熟女 | 久久精品人人爽人人爽视色|