• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Multi-Layer Deep Sparse Representation for Biological Slice Image Inpainting

    2023-10-26 13:15:30HaitaoHuHongmeiMaandShuliMei
    Computers Materials&Continua 2023年9期

    Haitao Hu ,Hongmei Ma and Shuli Mei,?

    1College of Information and Electrical Engineering,China Agricultural University,Beijing,100000,China

    2Yantai Research Institute,China Agricultural University,Yantai,264670,China

    ABSTRACT Biological slices are an effective tool for studying the physiological structure and evolution mechanism of biological systems.However,due to the complexity of preparation technology and the presence of many uncontrollable factors during the preparation processing,leads to problems such as difficulty in preparing slice images and breakage of slice images.Therefore,we proposed a biological slice image small-scale corruption inpainting algorithm with interpretability based on multi-layer deep sparse representation,achieving the high-fidelity reconstruction of slice images.We further discussed the relationship between deep convolutional neural networks and sparse representation,ensuring the high-fidelity characteristic of the algorithm first.A novel deep wavelet dictionary is proposed that can better obtain image prior and possess learnable feature.And multi-layer deep sparse representation is used to implement dictionary learning,acquiring better signal expression.Compared with methods such as NLABH,Shearlet,Partial Differential Equation(PDE),K-Singular Value Decomposition(K-SVD),Convolutional Sparse Coding,and Deep Image Prior,the proposed algorithm has better subjective reconstruction and objective evaluation with small-scale image data,which realized high-fidelity inpainting,under the condition of small-scale image data.And the O(n2)-level time complexity makes the proposed algorithm practical.The proposed algorithm can be effectively extended to other cross-sectional image inpainting problems,such as magnetic resonance images,and computed tomography images.

    KEYWORDS Deep sparse representation;image inpainting;convolutional sparse modelling;deep neural network

    1 Introduction

    Biological slice is a technique that uses frozen or paraffin slicing to obtain thin slices of biological tissue,which is an important approach to studying the interaction mechanism of biological tissue and the system.For example,mouse brain slices are an important model for studying the development of mouse neural networks,synapses,and brain area function[1].In addition,lust slice images can provide strong support for the reconstruction of lust 3D models[2].Biological slice images,as cross-sectional images,contain rich texture and contour information,which are important features for identifying objects in the images[3].However,during the slicing process,image degradation such as corruption may occur,which affects further research on biological mechanisms.Therefore,we need to perform inpainting on small-scale corruption in images.Besides,The uniqueness and individual differences of biological organisms make high-fidelity inpainting of biological slice images necessary to ensure the authenticity of the inpainting.

    The example-based method[4]uses the structure and redundancy of an image to search for the best-matched block in the image and fill in the missing areas.However,this method is only suitable for images with self-similarity and repetitive textures.In recent years,the use of deep learning to achieve image inpainting has gained some popularity.Pathak et al.[5]proposed the use of Deep Convolutional Neural Network(DCNN)to obtain high-dimensional information to guide image inpainting,showing some promising results.Generative deep learning methods (represented by Generative Adversarial Networks and denoising diffusion inpainting models[6])have demonstrated powerful abilities in the inpainting of large-scale corrupted images.However,deep learning relies on statistical information and only focuses on visual plausibility,and lacks interpretability.The reliability of the image reconstruction results cannot be guaranteed,casting doubts on the application of deep learning in biological slice image inpainting[7].

    Sparse representation(SR)[8,9]is a research direction in solving ill-posed problems.It has been extensively used in solving inverse problems like image denoising,restoration,and deblurring.Studying the interpretability problem of deep learning through sparse representation has become a feasible method[10–12].Usually,sparse representation solves the following constrained optimization problem:

    where D represents the dictionary,z represents the sparse coding,and λ represents the regularization parameter.

    The sparse coding reconstruction of an image is achieved by the linear combination of atoms in the dictionary [13].The dictionary serves as a set of basis vectors for signal representation and can be divided into fixed and learned dictionaries.Fixed dictionaries are no longer considered due to their inflexibility and rigidity.Learned dictionaries are initialized appropriately and obtained through training.The form of the initial dictionary affects the optimal results of dictionary learning.The initialization of the dictionary can be obtained by calculating Hilbert space basis functions,such as using Discrete Cosine Transform (DCT),Wavelet Transform,etc.[14].Some scholars use the convolution kernel function as the initial dictionary or construct convolutional sparse dictionaries by using convolutional methods[15].The lack of prior knowledge of the image,and the presence of noise and random initialization of the dictionary,make the learning of sparse coding difficult.

    The K-SVD algorithm[16]can obtain a set of overcomplete basis vectors,which can effectively represent signals.However,the time complexity of the optimization process is daunting.Although denoising algorithms such as BM3D[17]and DnCNN[18]have surpassed the denoising performance of the K-SVD algorithm,the K-SVD algorithm still has a wide range of applications.Can the K-SVD algorithm be revitalized in the era of deep learning? Scetbon,Elad,and other scholars proposed the Deep K-SVD algorithm[19],which re-interpreted the dictionary update strategy in the K-SVD algorithm as a differentiable form and constructed an interpretable end-to-end deep neural network model with a lightweight architecture that achieves a denoising performance approaching that of the DCNN.

    In high-reliability requirements for the task of restoring biological slice images,deep learning uses generative means,but the authenticity of the reconstruction results cannot be guaranteed.Using deep neural networks such as Autoencoder and restricted Boltzmann machines as deep sparse representation models [20],the interpretability is not clear.Deep sparse representation is a deep learning model based on sparse representation,which possesses a certain level of interpretability and learning capability.It enables better high-fidelity restoration and meets the authenticity requirements of biological slice images.Deep sparse representation still requires a large amount of data,and selecting hyperparameters for training poses certain difficulties.

    We summarize the contributions of this paper as follows: (1) We propose an end-to-end deep neural network model based on deep sparse representation to address the task of small-scale damage restoration in biological slice images.(2)We investigated the learnability of wavelet dictionaries and proposed a deep wavelet dictionary along with its updating algorithm.(3)We conducted an in-depth analysis of the relationship between sparse representation and deep neural networks,providing a comprehensive discussion on the interpretability of deep sparse representation.

    The paper is organized as follows.Section 2 introduces related work,Section 3 discusses the proposed algorithm,Section 4 reports experimental results and analysis,and Section 5 concludes the paper.

    2 Related Work

    2.1 Wavelet Transform

    Wavelet transform can capture local features of an image from multiple perspectives and achieve energy concentration.But,wavelet basis functions are manually designed,and fixed basis functions may not adapt well to signal families.The wavelet function can be obtained through the discretization of the parameters a and b of the continuous wavelet function.

    The response of the wavelet function to signalS(t)∈L2(R)is shown in Eq.(3).

    whereψ(t)is the wavelet function.

    In the L2(R)space,the one-dimensional discrete wavelet transform convolves and downsamples the signalf(t)using a high-pass filterψ(x)and a low-pass filterφ(x)to obtain approximation componentsdjand detail componentsaj[21].For an imageY∈L2(R2),the two-dimensional discrete wavelet filter is obtained by taking the tensor product of the one-dimensional wavelet filter,defined as follows[22]:

    In the equation,Φ(x,y)represents the scaling function,Ψ(x,y)represents the wavelet function.Fast Mallat algorithm is commonly used to efficiently compute the discrete wavelet transform,which uses the alternating convolution and down-sampling operations of the low-pass filterΦ(x,y)and high-pass filterΨ(x,y)to obtain the approximation and detail components of the image.

    Deep learning has shown great charm with its powerful fitting ability[23,24].This leads people to think about whether it is feasible to re-examine traditional methods,including wavelet transform,from the perspective of deep learning.Fortunately,following the train of thought of traditional methods and reconstructing algorithms from the perspective of deep learning has become a trend nowadays[25].Efforts have been made to improve the interpretability of DCNN,and it has been found that its learned results tend to approach wavelet transform or sparse representations[12].We consider building a trainable wavelet dictionary using learnable wavelet functions to obtain a better dictionary.

    2.2 Convolutional Sparse Modelling

    Convolutional Sparse Coding(CSC)is one approach to sparse representation in image processing,which is supported by strong theoretical foundations and has good biological plausibility.However,in recent years,the performance of CSC has been surpassed by deep learning.Building a“deep”CSC model has potential application values in various fields such as image restoration[15,26],image classification[27],and image registration[28].

    Sparsity has been integrated into the development of deep neural networks[29].CSC can also be used to construct deep sparse models by applying deep architecture,which transforms the optimization problem in Eq.(1)into the following form:

    since solving the?1norm is an NP-hard problem,it is usually related to the?1norm,which can be transformed into a convex optimization problem.

    Autoencoders,Restricted Boltzmann Machines,and other deep neural networks lack clear interpretability when solving Eq.(1).Sparse coding usually uses greedy algorithms or iterative thresholding algorithms,with the latter being able to approximate basis pursuit and implement sparsification in the form of network unfolding.Daubechies proposed Iterative shrinkage-thresholding algorithm(ISTA)[30]to approximate layered basis pursuit,updating the sparse coefficients through j rounds of iteration,with small values set to zero in each round while the rest remains almost unchanged.It is defined as follows:

    in the ISTA algorithm,Lis the step size with a value of the maximum eigenvalue ofDTD.λis the regularization parameter,and the soft thresholding operator is denoted asSθand is defined as:

    when the dictionaryDis enforced to be shared,the thresholding scheme can be approximately viewed as a“recurrent neural network”.Deep sparse representation is a deep learning model based on sparse representation.It has advantages such as stronger representation power and lower time complexity.However,it poses challenges in terms of selecting model hyperparameters and constructing datasets.

    3 Sparse Representation and Deep Neural Network

    3.1 Connection Between Sparse Representation and Deep Neural Networks

    Deep neural networks are developed based on the study of biological neural systems.The representation of signals in deep neural networks is non-linear,and the feature extraction is complex with a multi-scale network hierarchy.Deep convolutional neural networks(DCNN),as the representative of deep neural networks,use operations such as convolutional layers,linear layers,and pooling layers.The convolutional layer contains operations like convolution,activation,and bias.The linear layer performs linear transformations on the convolutional coefficients.The pooling layer implements the multi-resolution analysis of the neural network.

    The convolutional layer simulates the functions and structures of biological neurons.For input signalsy∈RM?N,the convolution operation is a group of convolutions applied toy,i.e.,x=φy,whereφrepresents a convolution matrix.The result is strengthened by the biasband then passed through an unsaturated non-linear activation function,such as the Rectified Linear Unit(ReLU),to achieve sparse activation.Therefore,neurons in DCNN can be defined as follows:

    in the equation,ReLU(α)=max(0,α).The multi-layer forward propagation mechanism in DCNN allows for multi-scale analysis.The parameter updates using gradient descent and backpropagation allow DCNN to provide feedback to the model based on the loss function.The transition from linear activation functions to saturated non-linear activation functions and to unsaturated non-linear activation functions enhances the non-linear characteristics of DCNN while increasing the sparsity.This indicates that neural networks are evolving towards sparser responses that resemble those of biological neural systems[30].

    The starting point for the construction of sparse representation theory is the sparse response of neurons in the visual cortex of the brain to visual signals.The update of sparse coefficients is generally achieved by using an iterative thresholding function to realize sparsification,such as the ISTA algorithm shown in Eqs.(6)and(7).It can be observed that the sparse coding algorithm in deep sparse representation theory and the structure of neurons(convolution operation)in neural networks have equivalence.Both apply a convolution followed by a non-linear transformation with a threshold to output an activation or inhibition state.

    As shown in Fig.1,the ReLU function and the ISTA algorithm both achieve sparseness and nonlinearity as a variant of the threshold function.The convolution operation applied to the convolutional coefficients in DCNN is equivalent to the convolution between dictionaryφand sparse coefficientsαin sparse representation theory.The biasbfunctionally equivalent to the thresholdθ,and the contraction operation both result in the sparsity of coefficients.Therefore,the functionality of neurons in the forward propagation of DCNN is equivalent to the sparse coding of sparse representation theory.

    Figure 1:Sparse regularization strategy

    The above discussion confines the problem to a single-layer model.The multiscale nature of DCNN is evident,formally defined as follows:

    By extending the basic deep sparse representation to multiple layers,we can obtain the Multi-layer Deep Sparse Representation(ML-DSR)model,formally defined below:

    this indicates that deep sparse representation also has a multi-scale analysis mechanism.

    DCNN and sparse representation share similarities in optimization and feedback mechanisms.In DCNN,backpropagation is applied in each iteration to update the network parameters with respect to the loss function,using the gradient descent method.Such optimization and feedback mechanisms ensure that the final solution moves towards minimizing the error.In sparse representation,the dictionary is updated based on the feedback from the reconstruction error,and the update direction is towards minimizing the loss.This optimization mechanism is evident in convolutional sparse models,improving the learnability of dictionaries and network parameters.

    3.2 High-Fidelity Reconstruction

    Starting from the basic form of deep sparse representationX=Dα,assumingX={x1,x2,...,xn-1,xn}Tandα={α1,α2,...,αn-1,αn}T,to achieve the reconstruction of the true valueX,we need to pursueargmin,where i=1···n,to obtain the optimal sparse coefficientα.Reference[10] provided the conditions for stable recovery of the original signal in the worst-case scenario.As follows:

    μ(D)represents the coherence of the dictionaryD,which is defined by the maximum inner product of different atoms in the dictionary.The solution obtained by applying the pursuit algorithm to this is stable and reliable,satisfying the condition.

    the symbolεrepresents bounded noise disturbance to the signal,which indicates that perfect signal reconstruction can be achieved when,thereby achieving high-fidelity reconstruction.However,the precise pursuit of the sparse coefficientαis very difficult,so we relax the constraint on,and achieve a high-fidelity representation of the image under the condition thatδ=<σ.

    Based on the Deep K-SVD algorithm,we constructed an artificial image dataset to validate its ability to effectively represent images through deep sparse representation,as shown in Fig.2.The root mean square error(RMSE)was used to measure the errorδof the reconstructed artificial images.Fig.3 shows the results of several artificial images reconstructed using deep sparse representation,which met our theoretical expectations in terms of visual effects and metrics analysis.The RMSE errors of the four test images reconstructed were 0.0185,0.0667,0.0326,and 0.0108,respectively,and the errors were within an acceptable range.This indicates that deep sparse representation can achieve the precise reconstruction of biological slice images,which establishes the basis for high-fidelity image inpainting.

    Figure 2:Artificial image datasets

    Figure 3:The reconstructed results of the test images.a1,b1,c1,and d1 are the original artificial images,while a2,b2,c2,and d2 are the reconstructed results

    4 Deep Sparse Representation Inpainting

    Our end-to-end biological slice image inpainting model is shown in Fig.4.First,we construct a deep wavelet dictionary (DWDict) for the imagey.Next,we decomposeyinto a collection of overlapping image blocks and generate corresponding sparse code.We then iteratively update the dictionary and sparse code in k layers.Finally,we obtain the reconstruction result by simple averaging.The error between the reconstruction result and the ground truth is fed back to the model to update the network’s parameters.Next,we will explain the details of dictionary construction and sparse coding.

    Figure 4:The algorithm architecture.The upper part represents the process of reconstructing damaged images,while the lower part shows the details of the encoding stage and dictionary update stage

    4.1 Deep Wavelet Dictionary

    Multiscale geometric analysis tools can achieve sparse representation of target images[31].Using wavelet transform to construct a wavelet dictionary can further exploit the sparsity performance of the algorithm,and the prior knowledge obtained from the images can enhance the performance of deep learning algorithms[32].However,wavelet transform basis functions rely on laborious designs,obtaining the optimal sparse representation of signal sets difficult.Therefore,we designed learnable wavelet basis functions and used the image prior knowledge to construct a Deep Wavelet Dictionary(DWDict),as shown in Fig.5.

    To meet the multiscale characteristics of wavelet decomposition,we first construct 2D discrete wavelet filters based on Eq.(5)and then construct a cascaded deep wavelet network as the convolutional kernel function to achieve convolutional operation and parameter learnability.Meanwhile,the down-sampling operation with a step size of 2 is performed to meet the requirements of each layer decomposition in the Fast Mallat algorithm.The form of the deep wavelet decomposition network is shown in Fig.6.LetFi(n)be the non-linear function of the convolutional kernel function,and for signalS(t),the response of the deep wavelet decomposition network is shown in Eq.(13):

    where ↓2 represents the down-sampling operator in the equation.

    The imageYis decomposed into a coefficient setα={αi|i=1,...,n} by the deep wavelet decomposition network,and we decompose the coefficient setαinto a normalized coefficient block collectionaccording to the step sizepand specifications.The coefficient block retains local information such as the edges of the image.

    The dictionary should cover the entire space of x in spaceRNas much as possible.Does the K-SVD algorithm tend to learn what kind of atoms and does it meet the above ideas?We constructed some artificial images and used the K-SVD algorithm to generate their dictionaries,as shown in Fig.7.

    Figure 5:Algorithm flow for the construction of the deep wavelet dictionary

    Figure 6:Deep wavelet decomposition network

    The K-SVD algorithm tends to favor learning the parts of the image with obvious features and selects representative parts from many feature areas as atoms.Based on this characteristic,we decided to prioritize image blocks with clear edge texture features in the construction of the deep wavelet dictionary.Therefore,we use the mean gradient as a measuring factor,which can characterize the degree of grayscale change,to select potential image blocks.The formula is defined as follows:

    in the formula,Prepresents the decomposed small blocks.To avoid the disturbance caused by the damaged areas on the dictionary construction,the image blocks in these areas will be excluded in advance.

    Figure 7:The results of K-SVD dictionary learning:(a)ring,(b)dictionary;(c)square,(d)dictionary

    4.2 Sparse Coding

    Our sparsification strategy is to use a learnable ISTA algorithm to achieve sparse coding.The parameters of the ISTA algorithm rely entirely on manual design,which makes it difficult to achieve optimal sparse representation,especially for ill-posed inverse problems.To address this issue,Gregor et al.[33] proposed the Learned-ISTA algorithm to learn the model’s parameters,but its adaptability is not strong.

    Due to the powerful representation capability of DCNN,the ISTA-Net algorithm[34]showed that nonlinear transformation functions can achieve sparse representation of images.TheF(·)operator constructed using nonlinear operators and ReLU linear units can achieve a richer expression of images.We use the learnability and nonlinearity of theF(·)operator to improve the sparsification performance of the sparse coding algorithm.Sparse coefficientαgets byα=F(α),and then as input to Eq.(15).

    in the equation,ρ=λ/Lis a learnable parameter and its value is obtained by the MLP module to satisfy an appropriate stopping criterion for each small block.

    ML-DSR shows that D orαcan be a component of a layer,or can be used as the target signal.Therefore,we use the dictionary D as the target signal and deploy a shallow deep neural network model based on ML-DSR,which is updated in each iteration as defined below:

    the dictionary also uses a deep differentiable method to optimize it during model feedback.

    4.3 General Structure

    The issue of small-scale corruption can be described asy=A⊙x,where ⊙represents the elementwise product andArepresents the limited operator.When addressing the damage caused by the limited operator,a new optimization objective function is obtained as shown in Eq.(19).

    In the Deep K-SVD algorithm,the?2norm is used to measure the error between the true value and the reconstructed result,i.e.,loss=‖x-f(y)‖2.However,the?2norm alone does not necessarily produce good sparse solutions and may cause artifacts,ignoring the visual perception demands of the human eye.Therefore,according to the suggestion in reference[33],a mixed form ofms_ssimand?1loss is used as the loss function.

    As it is a global reconstruction for images with small-scale damage,information loss in the nondamaged areas of the image is inevitable.Therefore,we adopt the complement operation on the reconstructed result X,which allows us to focus on the reconstruction of the damaged areas and obtain better reconstruction results.The complement operation is defined as follows:

    the symbol“~”represents a negation operation on the mask M.Finally,we can get the following algorithm flow.

    5 Experiment

    5.1 Data Sets and Preparatory Work

    We used the frozen slice method to prepare the biological slices and stained them with uranyl acetate and lead citrate.Then we obtained 90 to 100 micrographs of lust microscopic slices by photographing them with a microscope.Processing the slice image by image enhancement methods such as shading correction and cropping.The process of slice image acquisition is shown in Fig.8,The final obtained lust slice images contain rich texture information,complex edge contour structures,and segmental smoothness.In some tissue slices,there are also self-similar fractal structures[35],which reflect the complexity of biological structures.Sliced image inpainting has the following challenges:first,sliced images have complex texture contour information and thus need to be restored realistically at the time of inpainting,second,the low amount of sliced image data,and third,the need to avoid application difficulties due to high time complexity.

    Figure 8:Biological sliced image preparation process

    Due to the complexity of the slicing process and the small size of the data,we extract a specified number of images from the training set to effectively address the problem of insufficient data.Specifically,am?nsized image is cropped into(m-k+1)?(n-k+1)sub-images of sizek?k.A manually crafted set of around 200 small-scale masks with different shapes are used to initialize the position of masks on the sliced images during data loading,to improve the generalization performance of the model.

    Our model is built using the PyTorch framework and trained using the ADMM optimizer with a learning rate of 1e-4.The patch size is 8.The computer used for model training is equipped with an Intel(R) Xeon(R) Platinum 8255C CPU @ 2.50 GHz and NVIDIA GeForce RTX 3080.For model testing,both MATLAB R2021a and Python 3.7.4 environments were used on a computer with an Intel(R)Core(TM)i7-9750H CPU@2.60 GHz and NVIDIA GeForce GTX 1650.

    5.2 Experiment Result

    We design experiments from several aspects such as restoration effect,model complexity,and practicality,and use both simulated breakage and real breakage to show the restoration effect in the evaluation of inpainting effect.Fig.9 shows the training loss,which was completed in just over 3 h.During testing and analysis,we used grayscale images of size 512 ?512 pixels for evaluation,including subjective and objective evaluations,time complexity comparison,model size comparison,and other assessment experiments that will be discussed in the following.

    Figure 9:Training loss

    As shown in Fig.10,we constructed a DCT dictionary with 256 atoms and compared it to a deep wavelet dictionary with the same number of atoms.We found that the deep wavelet dictionary captures local feature information such as curved stripes,which are edge-like features,whereas the DCT dictionary mainly consists of striped and grid-like information.This indicates that the variety of atoms in the deep wavelet dictionary is more diverse,and the prior knowledge obtained from images can lead to better recovery performance of the algorithm.

    Figure 10:Dictionary construction algorithm,(a)is the DCT dictionary,(b)is the DWDict

    To evaluate the effectiveness of our proposed algorithm,we compared it with classical image inpainting algorithms and deep learning-based image inpainting algorithms,including NLABH[36],Shearlet [37],Mumford-Shah [38],K-SVD [16],Deep Image Prior (DIP) [32],and Local Block Coordinate Descent Algorithm(LoBCoD)[39].To assess the reconstruction quality of the algorithms,we used both subjective visual evaluation and objective evaluation metrics such as peak signal to noise ratio (PSNR),structural similarity index (SSIM),and root mean square error (RMSE),which can reflect the quality of image inpainting.

    Fig.11 shows the reconstruction results of different models after inpainting,and their evaluation metrics can be obtained from Table 1,which used simulated breakage masks.It can be observed that PDE-based inpainting methods such as Mumford-Shah algorithm exhibit strong competitiveness in small-scale damage inpainting by reasoning from the damaged edges towards the inside.LoBCoD algorithm,as a convolutional sparse coding model,uses the local block coordinate descent method to find the optimal basis vector and achieves good reconstruction performance.Compared with K-SVD algorithm,NLABH algorithm,our proposed algorithm achieves excellent subjective visual effects and optimal PSNR and RMSE scores,with SSIM score also approaching the optimal,reflecting the superior performance of our algorithm in reconstruction quality.

    Figure 11:Slice image inpainting experiment:An original image,(b)masked image,(c)Mumford-Shah,(d)Shearlet,(e)NLABH,(f)K-SVDm,(g)DIP,(h)LoBCoD,and(i)Proposed algorithm

    Table 1:Evaluation of image contrast inpainting effect

    The time complexity of the algorithm isO(n2),which the bottleneck lies in the construction of the deep wavelet dictionary.Table 2 provides the running time of each model.We aim to evaluate the application value of the algorithms through a rough analysis of their running time.On the other hand,our proposed algorithm only takes less than 13 s to achieve comparable inpainting results to the KSVD algorithm.Although the K-SVD algorithm also has good performance,its resource consumption is huge,reaching an intolerable time scale,making it difficult to apply in practical work.Table 2 and the reconstruction results discussed above demonstrate that our proposed algorithm not only achieves competitive inpainting results but also has practical application value.This is the goal we are pursuing.

    Table 2:Run time comparison of algorithms

    We conducted experimental comparisons on different forms of biological slice images to demonstrate the effectiveness of our proposed algorithm,as shown in Fig.12 and Table 3.The highlighted red boxes demonstrate the rationality of our proposed algorithm in the inpainting of details.It can be seen that the PSNR indicator of the image after reconstruction using the multi-scale geometric analysis tool as a“sparse”expression tool is not ideal.This is because the geometric analysis tool adopts an approximation method,which brings information loss in the decomposition and reconstruction process.Compared to algorithms such as K-SVD and DIP,the algorithm proposed in this paper achieves a more natural and continuous transition at the boundaries of damaged regions.The SSIM value of the proposed algorithm reaches 0.993,which is an optimal result compared with other algorithms.

    To further highlight the advantages of our proposed model in terms of small scale and low complexity,we used two indicators:the total number of parameters and the number of floating-point operations per second(FLOPS).The total number of parameters reflects the scale of the model,while FLOPS reflects the complexity of the model.As shown in Table 4,compared with the lightweight model of DIP,it can be seen that our proposed algorithm has a total number of parameters of only 1.789M,which is lower than the 2.15M of DIP.The GFLOPS indicator of our proposed model is 19.288G,much lower than the 1.930T of DIP.The significant difference in GFLOPS is that our proposed algorithm does not drastically increase in complexity with the scale of the problem.

    Figure 12:Slice image inpainting experiment:An original image,(b)masked image,(c)Mumford-Shah,(d)Shearlet,(e)NLABH,(f)K-SVDm,(g)DIP,(h)LoBCoD,and(i)Proposed algorithm

    Another issue worth discussing is whether it is possible to train the model with fewer data to cope with the difficulty of acquiring images.As shown in Table 5,we trained the model with 16,32,and 44 images,respectively,and evaluated the results.It can be seen that even with a small amount of data,the reconstructed results are acceptable,and the error level is kept below 0.004.This indicates that our proposed model has value in situations where data is difficult to obtain,and can adapt well to scenarios where it is hard to acquire biological image data.

    Table 3:Evaluation of image contrast inpainting effect

    Lastly,to demonstrate the practical application capability of the proposed algorithm in this paper,this chapter focuses on the application of the high-fidelity restoration algorithm to the problem of small-scale damages in biological slice images.As shown in Fig.13,the location marked by the red box in the left image indicates the damaged area,which is labeled with a mask.It can be observed that the algorithm presented in this chapter achieves excellent restoration results for this real small-scale damage.

    Figure 13:Realistic corruptions inpainting in slice images

    6 Conclusion

    This article focuses on the issue of small-scale corruption in biological slice image preparation,particularly in the case of limited data.We analyzed the relationship between sparse representation and deep neural networks and established a deep network model based on deep sparse representation.Our proposed model can effectively inpainting small-scale corruption in biological slice images while preserving the edge texture and contour structure of the slice images.We conducted tests using simulated damages,compared with other methods such as PDE-based methods,Shearlet algorithm,DIP algorithm,K-SVD algorithm,and sparse coding algorithm,our proposed model achieved good results in terms of effectiveness,time,and model scale.And then we demonstrated the application capability of our proposed method in addressing true corruptions in biological slice images that highfidelity inpainting has been achieved.

    Unlike algorithms such as K-SVD and LoBCOD,the algorithm proposed in this article not only achieves high-fidelity inpainting of biological slice images but also benefits from well-time complexity,making it valuable for practical applications.The proposed algorithm can also be effectively applied to other cross-sectional image restoration tasks,such as MRI and CT scan images.In future work,we will focus on explainable deep learning research based on deep sparse representation.Furthermore,the efficiency of the proposed model still needs further improvement,we will focus on dictionary learning with an emphasis on the learnable wavelet dictionary.

    Acknowledgement:The authors extend their appreciation to the anonymous reviewers for their constructive comments and suggestions.

    Funding Statement:This work was supported by the National Natural Science Foundation of China (Grant No.61871380),the Shandong Provincial Natural Science Foundation (Grant No.ZR2020MF019),and Beijing Natural Science Foundation(Grant No.4172034).

    Author Contributions:Study conception and design: Haitao Hu,Shuli Mei;data collection: Haitao Hu,Hongmei Ma;analysis and interpretation of results:Haitao Hu,Shuli Mei,Hongmei Ma;draft manuscript preparation: Haitao Hu,Shuli Mei,Hongmei Ma.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data that support the findings of this study are available from the corresponding author upon reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    欧美日本亚洲视频在线播放| 黑人欧美特级aaaaaa片| 亚洲色图 男人天堂 中文字幕| 日本免费a在线| 欧美黑人精品巨大| 亚洲欧美精品综合一区二区三区| 国产精品一区二区免费欧美| 午夜福利成人在线免费观看| 麻豆av在线久日| 国产黄色小视频在线观看| 国产一卡二卡三卡精品| 亚洲第一av免费看| 黄色 视频免费看| 亚洲自偷自拍图片 自拍| 亚洲欧美日韩高清在线视频| 成人三级做爰电影| 99久久国产精品久久久| 视频在线观看一区二区三区| 欧美性猛交黑人性爽| 999久久久国产精品视频| 国产一区二区三区视频了| 侵犯人妻中文字幕一二三四区| 亚洲精品av麻豆狂野| 日本三级黄在线观看| 可以在线观看的亚洲视频| 日本免费a在线| 午夜免费观看网址| 欧美日本亚洲视频在线播放| 悠悠久久av| 国产精品一区二区精品视频观看| 无限看片的www在线观看| 可以在线观看毛片的网站| 国产精品永久免费网站| 国内精品久久久久久久电影| 久久久久九九精品影院| 99国产综合亚洲精品| 国产精品九九99| 久久天堂一区二区三区四区| 国产三级在线视频| 日韩精品青青久久久久久| 久热爱精品视频在线9| 欧美乱色亚洲激情| 中文资源天堂在线| 免费观看精品视频网站| 老司机福利观看| 性欧美人与动物交配| 国产97色在线日韩免费| 久久久精品国产亚洲av高清涩受| 香蕉国产在线看| 日韩大码丰满熟妇| 国产精品 国内视频| 亚洲国产精品成人综合色| 女性生殖器流出的白浆| 国产欧美日韩精品亚洲av| 美女免费视频网站| 欧美又色又爽又黄视频| 日韩精品青青久久久久久| 国产精品综合久久久久久久免费| 在线观看66精品国产| 中文资源天堂在线| 无限看片的www在线观看| 精品国产国语对白av| 国产1区2区3区精品| 男男h啪啪无遮挡| 欧美另类亚洲清纯唯美| 亚洲一区高清亚洲精品| 夜夜夜夜夜久久久久| 国产区一区二久久| 亚洲av日韩精品久久久久久密| 可以在线观看的亚洲视频| 日本 欧美在线| 成人免费观看视频高清| 国产欧美日韩一区二区三| 国内揄拍国产精品人妻在线 | 久久国产精品影院| 精品少妇一区二区三区视频日本电影| 日本成人三级电影网站| 天天一区二区日本电影三级| 亚洲专区字幕在线| 午夜日韩欧美国产| 欧美乱码精品一区二区三区| 一级毛片高清免费大全| 午夜免费鲁丝| 特大巨黑吊av在线直播 | 好男人电影高清在线观看| 欧美性长视频在线观看| 看黄色毛片网站| 99精品在免费线老司机午夜| 国内久久婷婷六月综合欲色啪| 热99re8久久精品国产| 丝袜人妻中文字幕| www.www免费av| 久久午夜亚洲精品久久| 亚洲精品久久国产高清桃花| 免费女性裸体啪啪无遮挡网站| 久久久久久久午夜电影| 久久中文看片网| 99re在线观看精品视频| av欧美777| 啦啦啦免费观看视频1| 男女午夜视频在线观看| 亚洲av成人不卡在线观看播放网| 熟女电影av网| 免费无遮挡裸体视频| 午夜两性在线视频| 欧美成人午夜精品| 在线观看www视频免费| 亚洲精品一卡2卡三卡4卡5卡| 国产亚洲精品久久久久久毛片| 亚洲精品在线美女| 在线观看www视频免费| 午夜视频精品福利| 日韩精品青青久久久久久| 久久久久亚洲av毛片大全| 韩国精品一区二区三区| 亚洲精品一区av在线观看| 亚洲国产精品999在线| 国产野战对白在线观看| 亚洲成人精品中文字幕电影| 久久精品aⅴ一区二区三区四区| 老汉色av国产亚洲站长工具| 婷婷六月久久综合丁香| 1024香蕉在线观看| 日本 av在线| 国产一区二区在线av高清观看| 欧美激情极品国产一区二区三区| 久久精品影院6| 深夜精品福利| 久久香蕉国产精品| 99国产精品一区二区蜜桃av| 最近最新免费中文字幕在线| 一夜夜www| 亚洲精品久久国产高清桃花| 精品熟女少妇八av免费久了| 很黄的视频免费| 久久久久久亚洲精品国产蜜桃av| 在线看三级毛片| 亚洲男人的天堂狠狠| 可以免费在线观看a视频的电影网站| 国产片内射在线| 欧美最黄视频在线播放免费| 亚洲国产日韩欧美精品在线观看 | xxx96com| 亚洲国产日韩欧美精品在线观看 | 两性午夜刺激爽爽歪歪视频在线观看 | 亚洲中文字幕一区二区三区有码在线看 | 日本免费一区二区三区高清不卡| 99久久久亚洲精品蜜臀av| www.自偷自拍.com| 18禁黄网站禁片免费观看直播| 欧美在线黄色| 亚洲熟妇中文字幕五十中出| 国产伦人伦偷精品视频| 午夜亚洲福利在线播放| 国产成人精品久久二区二区91| 日韩大码丰满熟妇| 老司机午夜十八禁免费视频| 久久伊人香网站| 啦啦啦韩国在线观看视频| 国产aⅴ精品一区二区三区波| 脱女人内裤的视频| 热re99久久国产66热| 亚洲中文日韩欧美视频| 国语自产精品视频在线第100页| 曰老女人黄片| 日韩欧美 国产精品| 欧美日本亚洲视频在线播放| 日韩 欧美 亚洲 中文字幕| www.999成人在线观看| 女生性感内裤真人,穿戴方法视频| av视频在线观看入口| 久99久视频精品免费| 成人国语在线视频| 黄色丝袜av网址大全| 亚洲精品av麻豆狂野| 在线国产一区二区在线| 老熟妇乱子伦视频在线观看| 国产精品综合久久久久久久免费| 国产真人三级小视频在线观看| 亚洲av成人av| 国产精品久久视频播放| 国产欧美日韩一区二区精品| 成年免费大片在线观看| 男女那种视频在线观看| 亚洲专区字幕在线| 国产午夜福利久久久久久| 波多野结衣高清无吗| 男女下面进入的视频免费午夜 | 成熟少妇高潮喷水视频| 亚洲,欧美精品.| 欧美性猛交黑人性爽| 人成视频在线观看免费观看| 久久这里只有精品19| xxx96com| 婷婷精品国产亚洲av| 亚洲精品一区av在线观看| 巨乳人妻的诱惑在线观看| 男女做爰动态图高潮gif福利片| 国产午夜精品久久久久久| 亚洲aⅴ乱码一区二区在线播放 | 俺也久久电影网| 无人区码免费观看不卡| 正在播放国产对白刺激| 国产成人精品久久二区二区免费| 亚洲va日本ⅴa欧美va伊人久久| 亚洲国产欧美日韩在线播放| 亚洲真实伦在线观看| 日本精品一区二区三区蜜桃| 日本免费a在线| 久久九九热精品免费| 精品午夜福利视频在线观看一区| 中文字幕高清在线视频| 一边摸一边做爽爽视频免费| av天堂在线播放| 日本精品一区二区三区蜜桃| 搡老妇女老女人老熟妇| 精品欧美国产一区二区三| 中出人妻视频一区二区| 国产精品电影一区二区三区| 婷婷亚洲欧美| 精品国产超薄肉色丝袜足j| 99re在线观看精品视频| 午夜福利欧美成人| 免费电影在线观看免费观看| 久久人妻福利社区极品人妻图片| 国产av又大| 精品日产1卡2卡| 女人爽到高潮嗷嗷叫在线视频| 亚洲熟女毛片儿| 91成人精品电影| 精品乱码久久久久久99久播| 日韩av在线大香蕉| 欧美色欧美亚洲另类二区| 一级毛片精品| 熟女电影av网| 美女午夜性视频免费| 视频区欧美日本亚洲| 高清毛片免费观看视频网站| x7x7x7水蜜桃| 99久久99久久久精品蜜桃| 亚洲中文字幕日韩| 97人妻精品一区二区三区麻豆 | 香蕉av资源在线| 免费看美女性在线毛片视频| 禁无遮挡网站| 色综合欧美亚洲国产小说| 麻豆一二三区av精品| 久热爱精品视频在线9| 此物有八面人人有两片| 在线永久观看黄色视频| or卡值多少钱| 成人手机av| 国产精品亚洲av一区麻豆| 一级毛片精品| 日韩欧美 国产精品| 中文字幕人妻丝袜一区二区| 母亲3免费完整高清在线观看| 久久久精品国产亚洲av高清涩受| 国内精品久久久久精免费| 亚洲 欧美 日韩 在线 免费| 日韩欧美一区二区三区在线观看| a级毛片在线看网站| 婷婷丁香在线五月| 91九色精品人成在线观看| 国产亚洲精品久久久久久毛片| videosex国产| 国产一区二区激情短视频| 制服诱惑二区| 日本三级黄在线观看| 亚洲自偷自拍图片 自拍| 日韩精品青青久久久久久| 美女高潮到喷水免费观看| 两性夫妻黄色片| 成人18禁在线播放| 欧美精品亚洲一区二区| 精品久久久久久,| 叶爱在线成人免费视频播放| 午夜激情福利司机影院| 高潮久久久久久久久久久不卡| 欧美精品亚洲一区二区| 久热这里只有精品99| 欧美日韩乱码在线| 97超级碰碰碰精品色视频在线观看| 久久 成人 亚洲| 免费在线观看影片大全网站| 免费av毛片视频| 啪啪无遮挡十八禁网站| www.精华液| 岛国在线观看网站| 99国产极品粉嫩在线观看| 欧美黑人欧美精品刺激| 十八禁人妻一区二区| 国产人伦9x9x在线观看| 久久久久国产精品人妻aⅴ院| 村上凉子中文字幕在线| av在线天堂中文字幕| 欧美色视频一区免费| 看免费av毛片| 精品一区二区三区视频在线观看免费| 精品国产美女av久久久久小说| 久久天躁狠狠躁夜夜2o2o| 欧美国产日韩亚洲一区| 老熟妇乱子伦视频在线观看| 搡老岳熟女国产| 亚洲七黄色美女视频| 性欧美人与动物交配| 又黄又粗又硬又大视频| 国产三级在线视频| 亚洲无线在线观看| 黄色片一级片一级黄色片| 日日摸夜夜添夜夜添小说| 精品电影一区二区在线| 91成年电影在线观看| 国产成人一区二区三区免费视频网站| 一区二区三区国产精品乱码| 午夜视频精品福利| 在线天堂中文资源库| 欧美成狂野欧美在线观看| 一级毛片精品| 久久久久亚洲av毛片大全| 国产一区二区激情短视频| 久久久精品国产亚洲av高清涩受| 国产成人精品久久二区二区免费| 婷婷精品国产亚洲av在线| 国产亚洲欧美在线一区二区| 欧美色视频一区免费| 国产精品野战在线观看| 久久欧美精品欧美久久欧美| 国产精品自产拍在线观看55亚洲| 亚洲成a人片在线一区二区| 亚洲自偷自拍图片 自拍| 久久狼人影院| 国产精品久久久久久人妻精品电影| 成人三级做爰电影| 国产精品免费视频内射| 丰满的人妻完整版| 亚洲精品久久国产高清桃花| 成人18禁高潮啪啪吃奶动态图| 身体一侧抽搐| 久久精品影院6| 桃红色精品国产亚洲av| 成人18禁高潮啪啪吃奶动态图| 黄网站色视频无遮挡免费观看| 欧美黑人精品巨大| 中国美女看黄片| xxxwww97欧美| or卡值多少钱| 欧美日韩亚洲国产一区二区在线观看| www.熟女人妻精品国产| 亚洲精品色激情综合| 妹子高潮喷水视频| 久久久久久九九精品二区国产 | 精品少妇一区二区三区视频日本电影| 亚洲国产精品sss在线观看| 俄罗斯特黄特色一大片| 最新在线观看一区二区三区| av电影中文网址| www日本黄色视频网| 啪啪无遮挡十八禁网站| 黄色丝袜av网址大全| 丝袜在线中文字幕| 色综合站精品国产| 夜夜夜夜夜久久久久| 搡老熟女国产l中国老女人| 亚洲美女黄片视频| 一夜夜www| 久久久久精品国产欧美久久久| 少妇熟女aⅴ在线视频| 国产精品免费一区二区三区在线| 香蕉国产在线看| 国产激情偷乱视频一区二区| 亚洲自拍偷在线| 亚洲一码二码三码区别大吗| 91老司机精品| 一边摸一边做爽爽视频免费| 99精品欧美一区二区三区四区| 一本久久中文字幕| 窝窝影院91人妻| 日本五十路高清| 中亚洲国语对白在线视频| 日本一本二区三区精品| 天天躁夜夜躁狠狠躁躁| 美女高潮到喷水免费观看| 日本撒尿小便嘘嘘汇集6| 91在线观看av| 国内揄拍国产精品人妻在线 | 国产单亲对白刺激| 久久久久免费精品人妻一区二区 | 90打野战视频偷拍视频| 久久久久国产一级毛片高清牌| 一进一出好大好爽视频| 亚洲成av人片免费观看| 国产免费男女视频| 亚洲男人天堂网一区| 首页视频小说图片口味搜索| 精品一区二区三区视频在线观看免费| 听说在线观看完整版免费高清| 正在播放国产对白刺激| 日韩精品青青久久久久久| 俺也久久电影网| 两性午夜刺激爽爽歪歪视频在线观看 | 一边摸一边做爽爽视频免费| 91av网站免费观看| 中文在线观看免费www的网站 | 亚洲国产看品久久| 女警被强在线播放| 成人国产综合亚洲| 91av网站免费观看| 久久 成人 亚洲| 国产高清激情床上av| 热99re8久久精品国产| 亚洲真实伦在线观看| 成年免费大片在线观看| 丁香欧美五月| 正在播放国产对白刺激| 国产成年人精品一区二区| 亚洲在线自拍视频| 久久香蕉国产精品| 国产伦一二天堂av在线观看| 国产成人精品无人区| 老熟妇乱子伦视频在线观看| 日韩大码丰满熟妇| 欧美亚洲日本最大视频资源| bbb黄色大片| 欧美+亚洲+日韩+国产| aaaaa片日本免费| 久久久国产欧美日韩av| 亚洲男人的天堂狠狠| 久久99热这里只有精品18| 午夜久久久久精精品| 久久久久久久久久黄片| 日日夜夜操网爽| 亚洲av日韩精品久久久久久密| 9191精品国产免费久久| 国产成人精品久久二区二区91| 午夜激情av网站| 18禁裸乳无遮挡免费网站照片 | av欧美777| 欧美成人免费av一区二区三区| 久久精品夜夜夜夜夜久久蜜豆 | 免费一级毛片在线播放高清视频| 久久香蕉国产精品| 香蕉久久夜色| 两个人视频免费观看高清| 国产精品国产高清国产av| 免费人成视频x8x8入口观看| 亚洲三区欧美一区| 欧美乱妇无乱码| 国产99白浆流出| 99在线视频只有这里精品首页| 欧美又色又爽又黄视频| 久久精品国产99精品国产亚洲性色| 美女免费视频网站| 国产精品永久免费网站| 一二三四社区在线视频社区8| 久久青草综合色| av有码第一页| 欧美绝顶高潮抽搐喷水| 中国美女看黄片| 亚洲三区欧美一区| 久久午夜亚洲精品久久| 人人妻,人人澡人人爽秒播| 色综合欧美亚洲国产小说| 亚洲精品久久成人aⅴ小说| 高清在线国产一区| 亚洲中文av在线| 久久精品国产综合久久久| 中文字幕另类日韩欧美亚洲嫩草| 欧美日韩精品网址| 国产精华一区二区三区| 一区二区三区高清视频在线| 夜夜夜夜夜久久久久| 久久亚洲精品不卡| 天堂影院成人在线观看| 中文字幕人妻丝袜一区二区| 好看av亚洲va欧美ⅴa在| 黄频高清免费视频| 亚洲精品一卡2卡三卡4卡5卡| 久久九九热精品免费| 自线自在国产av| 国产日本99.免费观看| 亚洲国产看品久久| 日韩大尺度精品在线看网址| 99国产综合亚洲精品| 日韩精品中文字幕看吧| 国产欧美日韩一区二区精品| 99国产极品粉嫩在线观看| 成人亚洲精品av一区二区| 国产成年人精品一区二区| 国产激情偷乱视频一区二区| 亚洲av第一区精品v没综合| 中文亚洲av片在线观看爽| 麻豆国产av国片精品| 久久精品人妻少妇| 精品日产1卡2卡| 1024香蕉在线观看| 久久久久久久久中文| 69av精品久久久久久| 亚洲人成网站在线播放欧美日韩| 国产真实乱freesex| 人人妻,人人澡人人爽秒播| 国产亚洲欧美精品永久| 亚洲,欧美精品.| 欧美日本视频| 亚洲av第一区精品v没综合| 国产视频一区二区在线看| 美女 人体艺术 gogo| 色在线成人网| 国产精品 欧美亚洲| 久久香蕉国产精品| 欧美 亚洲 国产 日韩一| 99riav亚洲国产免费| 欧美+亚洲+日韩+国产| 老司机午夜福利在线观看视频| 国产私拍福利视频在线观看| 黑人操中国人逼视频| 大型av网站在线播放| 妹子高潮喷水视频| 老司机午夜福利在线观看视频| 少妇被粗大的猛进出69影院| 国产精品电影一区二区三区| 欧美在线一区亚洲| 日韩中文字幕欧美一区二区| 日本撒尿小便嘘嘘汇集6| 欧美午夜高清在线| 日日夜夜操网爽| 丝袜美腿诱惑在线| 精品国产亚洲在线| 在线观看日韩欧美| 欧美一区二区精品小视频在线| 久99久视频精品免费| 精华霜和精华液先用哪个| 国产成人啪精品午夜网站| 在线观看一区二区三区| 熟女少妇亚洲综合色aaa.| 十八禁网站免费在线| 亚洲最大成人中文| 国产亚洲av嫩草精品影院| 免费看a级黄色片| 久久人妻av系列| 久久久久久国产a免费观看| 色老头精品视频在线观看| 在线观看日韩欧美| 美女免费视频网站| 免费观看精品视频网站| 国产高清有码在线观看视频 | 午夜免费成人在线视频| 亚洲国产精品久久男人天堂| 亚洲人成77777在线视频| 天天一区二区日本电影三级| 99精品久久久久人妻精品| 国产在线观看jvid| 怎么达到女性高潮| 人成视频在线观看免费观看| 午夜福利高清视频| www.熟女人妻精品国产| 亚洲人成77777在线视频| 亚洲三区欧美一区| 老司机午夜福利在线观看视频| 国产精品久久久久久精品电影 | 亚洲五月天丁香| 成人国产一区最新在线观看| 久久伊人香网站| 欧美性猛交黑人性爽| 欧美成人午夜精品| 欧美黄色淫秽网站| 国产av不卡久久| 曰老女人黄片| 老司机午夜十八禁免费视频| 亚洲成人国产一区在线观看| 不卡av一区二区三区| 麻豆国产av国片精品| 亚洲精品久久成人aⅴ小说| 看片在线看免费视频| 一边摸一边抽搐一进一小说| 免费看a级黄色片| 久久人妻av系列| 黄频高清免费视频| 国产爱豆传媒在线观看 | 亚洲熟女毛片儿| 免费在线观看成人毛片| 国产精品,欧美在线| 在线观看舔阴道视频| 午夜福利视频1000在线观看| 欧美在线黄色| 黄色毛片三级朝国网站| 久久久久国内视频| 久久久国产欧美日韩av| 日日夜夜操网爽| 中亚洲国语对白在线视频| 91成人精品电影| 脱女人内裤的视频| 一级片免费观看大全| 国产三级在线视频| 午夜免费鲁丝| 亚洲五月天丁香| 最好的美女福利视频网| 看免费av毛片| 久久亚洲真实| 村上凉子中文字幕在线| 人妻丰满熟妇av一区二区三区| 欧美黑人巨大hd| 免费看a级黄色片| 精品国内亚洲2022精品成人| 一个人观看的视频www高清免费观看 | 亚洲第一欧美日韩一区二区三区| 久久久水蜜桃国产精品网| 侵犯人妻中文字幕一二三四区| 国产区一区二久久| 亚洲精品粉嫩美女一区| 少妇 在线观看| 欧美一级毛片孕妇| 色综合站精品国产| 国产99白浆流出| 精品久久久久久成人av|