• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    RF-Net:Unsupervised Low-Light Image Enhancement Based on Retinex and Exposure Fusion

    2023-12-12 15:51:14TianMaChenhuiFuJiayiYangJiehuiZhangandChuyangShang
    Computers Materials&Continua 2023年10期

    Tian Ma,Chenhui Fu,Jiayi Yang,Jiehui Zhang and Chuyang Shang

    College of Computer Science and Technology,Xi’an University of Science and Technology,Xi’an,710054,China

    ABSTRACT Low-light image enhancement methods have limitations in addressing issues such as color distortion,lack of vibrancy,and uneven light distribution and often require paired training data.To address these issues,we propose a two-stage unsupervised low-light image enhancement algorithm called Retinex and Exposure Fusion Network(RFNet),which can overcome the problems of over-enhancement of the high dynamic range and under-enhancement of the low dynamic range in existing enhancement algorithms.This algorithm can better manage the challenges brought about by complex environments in real-world scenarios by training with unpaired low-light images and regular-light images.In the first stage,we design a multi-scale feature extraction module based on Retinex theory,capable of extracting details and structural information at different scales to generate high-quality illumination and reflection images.In the second stage,an exposure image generator is designed through the camera response mechanism function to acquire exposure images containing more dark features,and the generated images are fused with the original input images to complete the low-light image enhancement.Experiments show the effectiveness and rationality of each module designed in this paper.And the method reconstructs the details of contrast and color distribution,outperforms the current state-of-the-art methods in both qualitative and quantitative metrics,and shows excellent performance in the real world.

    KEYWORDS Low-light image enhancement;multiscale feature extraction module;exposure generator;exposure fusion

    1 Introduction

    With the rapid development of artificial intelligence,low-light image-enhancement technology has been widely applied for pre-processing in advanced visual tasks.However,low-light images often suffer from detail degradation and color distortion due to the shooting environment and technical limitations.Balancing the image-enhancement effect and maintaining image realism are challenging problems in low-light image enhancement.These problems can significantly affect the performance of advanced downstream vision tasks.Therefore,improving visual quality and recovering image details have become important research topics.

    In the process of image enhancement,it is important to strike a balance between preserving image details and maintaining overall image quality.This requires preserving the original details in well-exposed areas,while appropriately brightening the underexposed areas to achieve a high-quality image.In addition,attention must be paid to balancing the brightness and contrast of the image during enhancement.If only the brightness is increased globally,the texture details in the image may be lost.Therefore,both brightness and contrast changes must be considered when enhancing an image to ensure its quality.Traditional methods [1–3] in the past often required a large amount of manual parameter adjustment to improve image quality;however,this method has significant limitations,as its effectiveness is largely based on assumptions regarding the threshold range.The use of low-light/normal-light images in supervised deep model training has become the main approach in algorithm research owing to advancements in deep learning.The accuracy of supervised learning methods depends on paired training datasets.However,it is technically difficult to obtain paired datasets from the same scene.In addition,the algorithm has poor generalization ability and cannot be effectively applied to real-scene images.In recent years,unsupervised image enhancement algorithms have emerged that eliminate reliance on paired datasets and achieve good enhancement results.For example,Deep Light Enhancement without Paired Supervision (EnlightenGAN) [4] uses unpaired datasets to train and implement low-light image enhancement techniques.Zero-Reference Deep Curve Estimation (Zero-DCE) [5] achieves enhancement using scene images with different illumination intensities.Although these methods eliminate the dependence of deep learning techniques on paired datasets,the quality of enhancement remains a challenge.The EnlightenGAN[4]method may produce artifacts,an overall uneven picture,and color-recovery errors when enhancing dark areas.Images enhanced using the Zero-DCE [5] method may exhibit whitish tones and less vibrant colors.These methods exhibit stronger generalization ability than supervised methods and reduce the requirements for dataset collection.

    To address these issues,we propose an unsupervised enhancement network called RF-Net,which combines Retinex with exposure fusion.The network comprises two stages: image decomposition and exposure fusion.In the first stage,to fully consider contextual and global information,we employed the powerful image-generation capabilities of a generative adversarial network and designed a multi-scale feature-extraction module to produce high-quality illumination and reflection images.Specifically,our network uses a multi-scale feature extraction module to perceptively capture differentscale features,preserve more detailed information,and avoid information loss between layers by using residual connections to transmit information from the current layer to the next layer.Most existing Retinex-based image enhancement methods obtain illumination and reflection component information matrices and generate enhanced images through calculations,which not only involve high computational complexity but also result in artifacts when processing shadow parts in dark areas.After obtaining the illumination and reflection images,an exposure image generator with correction coefficients was designed using the camera response function in the second stage to generate the exposure image and fuse it with the original low-light image to complete low-light image enhancement.The results obtained using the proposed method are shown in Fig.1.

    In summary,the main contributions of this paper are as follows:

    1.We devised a multi-scale feature-extraction module to produce high-quality illumination and reflection images.We incorporated a Coordinate Attention (CA) module that includes position-encoding information into the Markov discriminator.This module builds on channel attention and pays closer attention to the location information of the generated image,allowing for more accurate discrimination of texture details and improving the quality of the images generated.

    2.We improved the original Retinex formulation and designed an exposure image generator module with correction coefficients by referring to the camera response mechanism function.This module can generate images with different exposure levels while fusing illumination and reflection images.

    3.We proposed a novel unsupervised image enhancement method called RF-Net,which exhibits excellent performance in test results on several datasets and can be generalized to real-world low-light conditions.

    Figure 1:Representative enhancement results of RF-Net.Which can improve over-enhancement in the high dynamic range and under-enhancement in the low dynamic range

    2 Related Work

    In this section,we review research on low-light image enhancement using traditional and deep learning methods.

    2.1 Traditional Methods

    Histogram equalization is a classical image-enhancement method that enhances the contrast of an image by adjusting its brightness distribution.However,histogram equalization tends to cause image noise and over-enhancement problems.Some methods further increase the enhancement effect by setting a threshold to divide the image blocks [6],dividing the clipping points into chunks for processing[7],and combining them with adaptive gamma correction[8]to obtain a more reasonable S-shaped mapping function.However,these methods lead to problems of over-enhancement and amplification artifacts.The contrast was enhanced to an extent,but the details were lost.Retinex theory [9] posits that an image’s brightness is composed of two parts,reflection,and illumination,which enhance the image quality by separating these two parts.However,this approach is ineffective for images that are too dark or bright.Accordingly,researchers have proposed various improvement schemes[1–3],etc.This presupposes that spatial illumination changes slowly during implementation,but the processing is prone to halation and inaccurate color recovery.To reduce the computational cost in Retinex theory,researchers have proposed Low-Light Image Enhancement via Illumination Map Estimation(LIME)[10],the local illumination distribution of the image obtained by analyzing the local information.The local illumination distribution is applied to the reflection component to obtain the enhanced image.Compared with the Retinex method,LIME reduces the occurrence of halo artifacts during processing.However,LIME has a limited ability to distinguish between the foreground and background of an image,which can result in over-enhancement of the foreground and noise in the background.

    2.2 Deep Learning Methods

    Researchers have widely employed deep learning for image enhancement over the past decade,achieving promising results.In the following section,we review the current state of research on fully supervised,semi-supervised,and unsupervised approaches.

    2.2.1 Fully Supervised Methods

    The use of paired datasets to train network models has been widely adopted because of the oneto-one correspondence between the training data.RetinexNet[11],for the first time,combines Retinex theory with Convolutional Neural Networks to implement the low-illumination image enhancement problem by designing a decomposition module,and enhancement module.The feasibility of the Retinex application in deep learning was demonstrated for the first time.Kindling the Darkness(KinD)[12]designed global and local enhancement modules,where the global module extracts global luminance based on Retinex decomposition,and the local module enhances the image texture details.The global and local modules interacted through an adaptive mechanism,and the difference between the image generated by the global enhancement branch and the original image was used to calculate the weight of each pixel.These weights were then passed to the local enhancement branch to achieve contrast enhancement.KinD++[13]is based on KinD and improves the training speed and accuracy of the model by designing group learning and back-propagation mechanisms.GLobal Illuminationaware and Detail preserving Network (GLADNet) [14] generates a global before light by designing a global illumination estimation module that is then combined with the original input image to produce an enhanced image.Low-Light Image Enhancement with Normalizing Flow(LLFlow)[15]uses adaptive weights to control the effects of optical flow and global constraints;it also uses a deep learning model to learn the optical flow and the image to obtain an enhanced image.Self-Calibrated Illumination(SCI)[16]reduces computational costs by designing an adaptive correction illumination module that ensures the convergence of the results of each training phase to the final one.The literature[17]designs a generative adversarial network containing dual attention units that can effectively inhibit the artifacts and color reproduction bias generated during the enhancement.Transformer Photo Enhancement (TPE) [18] uses a pure transformer architecture to implement image enhancement based on multi-stage curve adjustment.Retinex based deep unfolding network (URetinex-Net) [19]decomposes the input image by designing a continuous optimization model with mutual feedback.To optimize the decomposition results,an implicit a priori regularization model was used,and a data initialization module,specific illumination intensity module,and denoising detail retention module were designed.Illumination Adaptive Transformer(IAT)[20]implements low-light enhancement by designing a lightweight transformer model that uses attention-query techniques to represent and adjust the parameters associated with the image signal processor(ISP).

    2.2.2 Semi-Supervised Methods

    These methods can learn better feature representations using both paired and unpaired data.First,researchers use paired datasets to train and obtain prior knowledge,and then they use the trained model as pre-training weights for unpaired data training.Based on this subdivision,Deep Recursive Band Network(DRBN)[21]introduces a recursive network architecture that uses the information of the highlight and shadow regions of the image and constructs a low-rank matrix and a sparse matrix to represent the brightness and structural information of the image,then inputs the two matrices into two branches of the network for feature extraction,and finally merges them to obtain an enhanced image.DRBN [22] utilizes a “band representation” technique to enhance low-light images.This method decomposes a low-light image into multiple bands using band representation and trains a neural network with a small amount of labeled data to learn how to enhance each band.Thus,this method retains the detail and texture information of a low-light image,thereby enhancing its quality.

    2.2.3 Unsupervised Methods

    Obtaining paired datasets can be difficult and using unsupervised methods has become the main approach for accomplishing image-enhancement tasks.This approach improves generality and applicability to many real-world scenarios.Exposure Correction Network (ExCNet) [23] is the first unsupervised enhancement method that uses the powerful learning ability of neural networks to estimate the most suitable“S”curve for low-light images and uses this curve directly to enhance the image.low-light image enhancement network(LEGAN)[24]is enhanced by a cleverly designed light perception module and a loss function that solves the overexposure problem.The EnlightenGAN[4]completed its first unsupervised image enhancement using unpaired datasets.This design overcomes the previous reliance on paired datasets by establishing unpaired mappings between low-light and nonmatching images and employing global-local discriminators and feature retention losses to constrain the feature distance between the enhanced and origin images for enhancement.Zero-DCE[5]uses a neural network to match a brightness mapping curve and then generates an enhanced image based on the curve.Unsupervised low-light image enhancement was achieved by designing a multi-stage highorder curve with a pixel-level dynamic range adjustment.Based on this,a lightweight version,Zero-DCE++[25],was developed.Generative adversarial network and Retinex (RetinexGAN) [26] uses a generative adversarial network based on Retinex to design a decomposition network with a simple two-layer convolution and achieve low-light image enhancement through image fusion.Restoration of Underexposed Images via Robust Retinex Decomposition (RRDNet) [27] achieves enhancement by designing a three-branch decomposition network and iterative loss functions to decompose the three components of reflection,illumination,and noise module.Retinex-inspired Unrolling with Architecture Search(RUAS)[28]used neural structure search to find an effective and lightweight set of networks for low-light enhancement.Retinex Deep Image Prior(RetinexDIP)[29]proposes a Retinexbased generation strategy that reduces the coupling between two components in the decomposition,making it easier to adjust the estimated illumination to perform an enhancement.

    3 Proposed Method

    In this section,the first module introduces the overall structure of the RF-Net network and provides a hierarchical representation of the first-and second-stage exposure-generation fusion networks.The second module describes the designed exposure image generator,and the third module describes the loss function.

    3.1 Network Architecture Design

    The proposed RF-Net is a two-stage network with the overall structure shown in Fig.2.In the first stage,two coupled-generator frameworks were used.The R network generates reflection images,whereas the L network generates illumination images.First,by cascading the maximum,minimum,and mean values of each channel of the original low-light image as inputs to the network,a multi-scale feature extraction module was designed to maintain the global consistency of good illumination and contextual information.After basic feature extraction,the resulting features are concatenated and mapped to high-dimensional information before being reduced in dimension through convolution,thereby improving image quality while learning complex features.For the discriminator,we used a VGG-based network structure in which the original Markov discriminator maps the input to an N×N matrix,such that each point in the matrix corresponds to the evaluation value of each region.To enhance the decomposition effect and discrimination accuracy of the network,we incorporated a CA attention mechanism with positional information[30].This further enhances the network’s ability to perceive and understand images and spatial information,learn useful features,and suppress irrelevant features,thereby improving the discriminator’s ability to accurately judge texture details.

    Figure 2:Overview of RF-Net.First,the low-light image and the corresponding maximum,minimum,and average grayscale images are input.The R and L have the same structure and are used to acquire the illumination and reflection components,respectively.Then the exposure image is acquired by the exposure generator.Finally,the low-light image is fused with the exposure image

    The second stage also consists of two coupled networks that use the original input image and the output of the first-stage network as inputs.At this stage,the output information from the first stage is processed by the exposure image generator module to create the initial exposure image.The details of this module are discussed in Section 2.The exposed image and original input image were then separately fed into two-branch networks[31].The two branches consist of a feature extraction module(FE),a super-resolution module(SR),and a feature fusion module(CF).The FE module consists of two convolution layers: SR uses the Convolutional Networks for Biomedical Image Segmentation(U-net) structure to learn more advanced features,and the first two modules are used to extract advanced features from the input low-dynamic-range images.RE was employed for super-resolution of the original input image before fusion to ensure the accurate extraction of high-level image features.The final module is the image fusion module that combines the super-resolution outputs of the two coupled networks and generates the output by weighting the super-resolution of the original image and the outputs of the two coupled blocks.

    3.2 Exposure Generation Module Design

    We employed a design that combines Retinex with the camera response mechanism to create an exposure image generator.In the original Retinex theory,the input imageSoutputis represented as the element-wise product of the illumination and reflectance components,which is expressed as Eq.(1):

    The input image is denoted as S,the reflected image as R,the illuminated image as L,andR×Ldenotes pixel-wise multiplication.However,numerous experiments have shown that the results obtained from the original Retinex equation are over-enhanced and lose detail owing to noise and uneven illumination.Therefore,we improved the original formula by first inverting the source illumination image using Eq.(2)to better utilize the content in the relatively overexposed region.

    The improved Retinex formula is represented by Eq.(3).

    where L and R denote the illuminated and reflected images,respectively,S denotes the original input low-light image,andSoudenotes the output result of the improved Retinex formula.

    To maintain a balance between brightness and contrast,we re-designed the improved Retinex formula using a camera response mechanism and proposed an exposure image generator.Here,we refer to the camera response function described in [32],where the model parameters of the camera response mechanism are determined by the camera’s parametersα,βand k.Parameter k is a correction factor that can be adjusted to obtain images with different exposure levels.As the k value increases,a brighter exposed image is acquired,and the details in the low-light areas become more significant;however,when the k value is too large,more detailed information is lost because of an exposure level that is too high.Therefore,we limited the value of k to a range of 2–6.The equation for generating the initial exposure image by combining the improved Retinex and camera response functions is expressed as Eq.(4).

    whereαandβare fixed parameters suitable for most cameras,withαset to -0.3293 andβset to 1.1258.Seorepresents the output exposure image,and different values of k directly affect the resulting output ofSeo.

    3.3 Loss Function

    Adversarial loss: In [11,12],by decomposing a low-light image into illumination and reflection components,the illumination components are approximately the same as those decomposed in a normally exposed image.With only differences in brightness,the reflection component is the same as the reflection component decomposed from the normal exposure image,which can be decomposed into high-quality reflection components by noise reduction.This means that the distribution of normally exposed images is very similar to that of the original images based on Retinex decomposition.Therefore,the original function [33] was used as an adversarial loss function to train the generator.In practical applications,the generated fake and real input samples are encoded as zero and one,respectively.Discriminators for the illumination and reflection maps were trained using squared error as the objective function.Our adversarial losses are defined by Eqs.(5)–(8).

    wheregLandgRdenote the reflected and illuminated images,respectively,denotes the average grey scale value of each channel.y represents the ground-truth reflection mapDLandDRrepresent the discriminators for the illumination and reflection maps,respectively.

    However, these aircraft had the familiar P-51 black paint with white stripes on the wings and were equipped with the wing tanks for extra range. Suddenly, they dropped their tanks just off to our right, and we looked around for German fighters in the area. We found them, when the whole formation of P-51s turned out to be Luftwaffe ME-109s that turned in to us with their cannons10 blazing! We narrowly missed being rammed11 by two of them that just barely passed over us.

    Perceptual loss: Using non-matching images for unsupervised image enhancement implies that the pixels in the training images are not one-to-one.The same pixel may have different semantics in different images.Therefore,we need a loss function to address the issue of non-corresponding pixel positions,and the perceptual loss function serves this purpose.This function is typically defined in the activation layer of a pre-trained network.By computing the distance between the activation layer features,it can effectively quantify the fundamental attributes of an image as well as the differences between its detailed features and high-level semantic information.This lays the foundation for generating high-quality images.Unlike common perceptual losses,this study adopted the concept of perceptual loss from[31].This implementation not only maintains the luminance consistency between the original and reference images but also recovers the details better.The perceptual loss function used in this paper is given by Eq.(9).

    where denotes the features extracted by the predefined network VGG19.CiHiWidenotes the number of channels and the height and width of the feature map in layer i.y denotes the unpaired real image information learned by the discriminator,g denotes the image generated by the generator.

    The total loss function of the network is shown in Eq.(10).

    4 Experiment

    4.1 Experimental Details

    We trained the RF-Net model on 914 randomly selected pairs of asymmetric datasets using the datasets provided in[4]and tested it on various datasets,including NPE[34],DICM[35],LIME[10],MEF[36],and VV.These datasets contain various low-light and unevenly exposed images from both indoor and outdoor settings.To demonstrate the performance of the enhancement algorithm better,we selected images with significant exposure differences from each test set as test set,which made the test more challenging.The deep learning framework used was PyTorch,and the hardware configuration was Tesla A100.

    To ensure that the training of the network could fully utilize the computing and storage resources of the computer and obtain better training results,the size of the training image was 640×400,and we randomly cropped the training data into a face slice of size 300×300.The batch size was set to one.To increase the data diversity,data expansion was performed,including random flipping,rotation,and cropping.This allowed the network to adapt better to various image scenes.The Adam optimizer was used to optimize the network,and the learning rate was set to 1e-4.Our network achieved better enhancement results with no more than 50 training iterations.

    4.2 Performance Evaluation

    To demonstrate the advantages of our proposed RF-Net,we compared our method with 10 other advanced methods:RetinexNet[11],EnlightenGAN[4],KinD[12],KinD++[13]RUAS[28],Zero-DCE [5],LLFlow [15],SCI [16],IAT [20],and GLADNet [14].Among these,RetinexNet,KinD,KinD++,RUAS,LLFlow,SCI,IAT,and GLADNet are supervised enhancement methods,whereas[4] and [5] are unsupervised enhancement methods.To ensure fairness,tests were conducted using the network parameters recommended in each study.Because we were unable to train the supervised models on unpaired datasets,we evaluated them using pre-trained models saved in the original papers.For unsupervised methods,if the original study used unpaired datasets for training,we used the datasets provided by [4].If the method used images with different exposures for training,we used the datasets provided by[5].Finally,the optimal model was selected for testing.These comparisons enabled us to evaluate the performance of the RF-Net method and demonstrate its competitiveness in image enhancement.

    4.2.1 Qualitative Comparison

    Figure 3:Qualitative comparison of RF-Net with other advanced algorithms.See the patch area for more detailed information

    Figure 4:(Continued)

    Figure 4:Qualitative comparison of RF-Net with other advanced algorithms.See the patch area for more detailed information

    Figure 5:Qualitative comparison of RF-Net with other advanced algorithms.See the patch area for more detailed information

    4.2.2 Quantitative Comparison

    The subjective evaluation may not be sufficient for determining the degree of detail retention during image enhancement.To demonstrate the feasibility of the proposed method further,we conducted quantitative comparisons.As we used unsupervised methods for model training,we could not evaluate the Peak Signal-to-Noise Ratio(PSNR)and Structural Similarity(SSIM)of the enhanced images to the ground truth,as with other supervised methods.Therefore,we used a non-referenced image quality assessment metric to compare the RF-Net method with other competitors.The metrics evaluated were Natural Image Quality Evaluator(NIQE)[37]and Blind/Referenceless Image Spatial QUality Evaluator(BRISQUE)[38].NIQE is a natural-image-based evaluation metric that compares algorithm processing results with a model calculated based on natural scenes.BRISQUE is an image-based no-reference quality score that is calculated based on natural scene images with similar distortions.The metrics used to evaluate the performance of RF-Net compared with the other algorithms are listed in Table 1.Based on the qualitative evaluation,the following issues were observed:RetinexNet[11]resulted in inaccurate color restoration with color bias;KinD[12]did not significantly enhance dark areas;KinD++[13]introduced artifacts while enhancing dark areas;RUAS[28]overenhanced the image,causing loss of information;LLFlow[15]produced incorrect color restoration;SCI[16]and IAT[20]over-enhanced the image and had insignificant enhancement in dark areas;Zero-DCE[5]had lower metrics compared to other unsupervised methods,possibly due to the whitish image produced by its enhancement results,as noted in the qualitative analysis.As shown in the graphs,the enhancement results of RF-Net validated this.In summary,because both NIQE and BRISQUE are methods based on local image statistical information,the impact of different algorithms on the small differences in the achieved metrics after image enhancement is better illustrated in the quantitative evaluation.

    Table 1:The NIQE(↓)and BRISQUE(↓)scores are shown,with lower scores indicating better image quality and richer information contained.The averages of the test image metrics are taken for each of the five datasets,and the five averages are eventually averaged again.The best result is shown in red and the second-ranked result is shown in blue

    4.2.3 Subjective Evaluation of People

    We also conducted a human subjective visual evaluation to further quantify the subjective quality of RF-Net compared to the other methods.We randomly selected 20 original low-light images from the test set(NPE[34],DICM[35],LIME[10],MEF[36],and VV)and applied four state-of-the-art methods (EnlightenGAN [4],Zero-DCE [5],SCI [16],and IAT [20]) to each image separately.We invited twelve reviewers to independently score the results of the five algorithms,including RF-Net.The reviewers primarily observed the following aspects:1.Whether the results contained artifacts in the over-or under-enhanced areas;2.Whether the color restoration in the results was accurate(e.g.,whether the colors were distorted);and 3.Whether the noise in dark areas was amplified,and whether there was an obvious loss of texture details in the results.As can be observed from the statistics in Fig.6,RF-Net achieved a higher subjective evaluation score for the reviewed images.

    Figure 6:Overview of people’s objective evaluation,in each graph,the x-axis indicates the observed quality of the five algorithms(1 for the best and 5 for the worst)and the y-axis indicates the number of good and bad images corresponding to m each algorithm.RF-Net shows the best performance

    4.3 Ablation Study

    To demonstrate the effectiveness of the modules used,the following ablation studies were conducted separately: three studies were designed in terms of CA removal,separation of inception[39]from residual connectivity[40],and direct fusion of illuminated and reflected components.

    In Ablation Study 1,we experimented with the multiscale module,residual connection,and CA attention separately,as shown in Fig.7,we can observe from the visual results that zooming in on the person above the house in the first image reveals that using only inception and the Markov discriminator leads to a blurred task,whereas using both inception and residual linking with a Markov discriminator produces more vivid colors and preserves more detailed information owing to residual linking.Furthermore,using inception,residual linking,and an improved Markov discriminator leads to a more realistic restoration of texture details,because the Markov discriminator has a clearer ability to distinguish true from false.The reddish hue in our image is due to the reflected image generated by the first stage of the network,which does not include knowledge of the illumination image or fusion module.

    Figure 7:Ablation study for each module.(a)Input,(b)w/o Multiscale,(c)w/o Residual,(d)w/o CA

    In Ablation Study 2,we compared our method with the direct integration of light and reflection components,as shown in Fig.8.The color of the sky in the second column of the first row was inaccurately restored,and the face of the person in the second column of the second row was overexposed.The proposed method in the third column performs significantly better than the direct fusion method.

    In Ablation Study 3,we tested the designed exposure generator module by adjusting the value of k to obtain images with different exposure levels and achieved exposure fusion.As shown in Fig.9,for different input images,we acquired images with varying exposure levels for fusion,with the k values set differently in different scenarios.For example,in images without exposed areas,k values are usually set between 4–6,while in images with exposed areas,k values are typically set between 2–4.This approach achieved the best enhancement when the exposed images were fused with the original lowlight images.Finally,we conducted a quantitative evaluation of the three sets of ablation studies.The results in Table 2 show that adding the residual block to the inception and residual block combinations produced superior performance.In addition,exposure fusion in the second stage of RF-Net exhibited advantages.A comparison of the second and fourth experimental rows demonstrates the effectiveness of RF-Net.

    Table 2:Quantitative review of ablation studies for each module

    Figure 8:Direct fusion of illuminated and reflected components with RF-Net ablation study

    Figure 9:Image decomposition and exposure generation fusion ablation study

    In Ablation Study 4,we use the low-light image and its corresponding Red,Green,Blue(RGB)channels with the maximum,minimum,and average values as inputs to our network.The final network input consists of six channels,with the first three channels used to generate the reflection map and the last three channels used to generate the illumination map.In the generated illumination images,the method used in this paper retains more details in the final generated illumination map compared with the input of only the original low illumination image.The qualitative comparison in Fig.10 shows that in the first column,there are artifacts near the shoulders of the person in the illumination map generated by the network with only low illumination input.In the second column,there is a noticeable loss of detail in the wall of the house on the right.In the third column,there are artifacts near the candle flame,and the grayscale boundary near the cup in the upper left corner is not clear;In the fourth column,there is an excessive amount of detail in the cave,while the detail in the trees outside the cave is lacking or insufficient.In Fig.11,the grayscale histogram results show that our method can generate a light map with smoother contrast and luminance changes,making the light and dark parts more visible and details more prominent.

    Figure 10:Qualitative comparison of illuminated images.(a) Input,(b) Low-light images as input,(c)Ours

    5 Conclusion

    In this study,we combined Retinex theory with exposure fusion for the first time to achieve unpaired low-light image enhancement.In the first stage,we designed a multi-scale generator by combining a residual network and inception.We also added a CA attention mechanism with position information to the discriminator network to obtain high-quality illumination and reflection components.By improving the original Retinex and camera response mechanism functions,we designed an exposure image generator with correction coefficients to solve the problems of illumination,reflection image fusion,and exposure image generation.Based on this,we realized low-light image enhancement with second-stage exposure fusion and proved the superiority of our method by comparing it with state-of-the-art methods.However,the algorithm has some limitations.On the one hand,it requires manual adjustment of the correction parameters for the exposure image generator based on the scene’s exposure level.On the other hand,Compared to other lightweight networks,RF-Net processes 640 × 400 images at a rate of 5 frames per second.In the future,our research will focus on lightweight the network structure and developing an adaptive low-light image enhancement method based on negative feedback control to solve the manual tuning problem of existing methods.We also aim to apply this model to enhance specific scenes,which will not only improve the generalization of the algorithm but also enhance the accuracy of other vision tasks.

    Acknowledgement:The authors also gratefully acknowledge the helpful comments and suggestions of the reviewers,which have improved the presentation.

    Funding Statement:This work was supported by the National Key Research and Development Program Topics (Grant No.2021YFB4000905),the National Natural Science Foundation of China(Grant Nos.62101432 and 62102309),and in part by Shaanxi Natural Science Fundamental Research Program Project(No.2022JM-508).

    Author Contributions:Study conception and design:Tian Ma,Jiayi Yang;data collection:Chenhui Fu;analysis and interpretation of results:Chenhui Fu,Jiehui Zhang,Chuyang Shang;draft manuscript preparation: Chenhui Fu.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data used in this paper can be requested from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    最近中文字幕高清免费大全6| 99久久人妻综合| 春色校园在线视频观看| 久久精品综合一区二区三区| 免费av毛片视频| av在线观看视频网站免费| 91午夜精品亚洲一区二区三区| 三级男女做爰猛烈吃奶摸视频| 久久精品国产亚洲av涩爱| 水蜜桃什么品种好| 深爱激情五月婷婷| 精品久久久久久久久亚洲| ponron亚洲| 亚洲国产高清在线一区二区三| 国内精品一区二区在线观看| 国产激情偷乱视频一区二区| 日本一本二区三区精品| 国产乱人偷精品视频| 女的被弄到高潮叫床怎么办| 啦啦啦啦在线视频资源| 99热全是精品| 国产精品熟女久久久久浪| 春色校园在线视频观看| 欧美成人一区二区免费高清观看| 亚洲欧美日韩卡通动漫| 日日干狠狠操夜夜爽| 国产精品精品国产色婷婷| 亚洲av电影在线观看一区二区三区 | 在线观看66精品国产| 亚洲最大成人中文| 看黄色毛片网站| 国产欧美日韩精品一区二区| 秋霞伦理黄片| 中文字幕久久专区| 联通29元200g的流量卡| 最近的中文字幕免费完整| 国产成人freesex在线| 亚洲国产精品合色在线| 黄色日韩在线| 成人亚洲欧美一区二区av| 好男人在线观看高清免费视频| 成人漫画全彩无遮挡| 亚洲18禁久久av| 精品一区二区免费观看| 午夜a级毛片| 久久精品国产自在天天线| 精品久久久久久久久亚洲| 国产真实伦视频高清在线观看| 大话2 男鬼变身卡| 亚洲丝袜综合中文字幕| 亚洲精品456在线播放app| 亚洲欧美日韩高清专用| av女优亚洲男人天堂| 亚洲四区av| 精品人妻熟女av久视频| 国产麻豆成人av免费视频| 国产真实乱freesex| 国产人妻一区二区三区在| 一本久久精品| 丰满少妇做爰视频| 卡戴珊不雅视频在线播放| 天美传媒精品一区二区| 天美传媒精品一区二区| 嫩草影院新地址| 国产高清有码在线观看视频| 免费观看人在逋| 我要看日韩黄色一级片| 精品午夜福利在线看| 国产高清视频在线观看网站| 日韩三级伦理在线观看| 日韩大片免费观看网站 | 国产成人91sexporn| 亚洲av免费在线观看| 亚洲一级一片aⅴ在线观看| 国产精品av视频在线免费观看| 亚洲欧美日韩无卡精品| 只有这里有精品99| 日产精品乱码卡一卡2卡三| 久久99蜜桃精品久久| 中文欧美无线码| 欧美日韩在线观看h| 精品久久久久久久久亚洲| 国内揄拍国产精品人妻在线| 国产一区二区在线观看日韩| 国内揄拍国产精品人妻在线| 亚洲av成人精品一二三区| 精华霜和精华液先用哪个| 久久精品夜夜夜夜夜久久蜜豆| 亚洲欧洲国产日韩| 国产伦精品一区二区三区视频9| 97人妻精品一区二区三区麻豆| 日本wwww免费看| www.色视频.com| 久久久成人免费电影| 久久久久性生活片| 在线免费十八禁| 乱系列少妇在线播放| 能在线免费观看的黄片| 最近2019中文字幕mv第一页| 最近2019中文字幕mv第一页| 青春草国产在线视频| 成人鲁丝片一二三区免费| 嘟嘟电影网在线观看| 嘟嘟电影网在线观看| 久久精品国产亚洲av天美| 国产大屁股一区二区在线视频| 久久久a久久爽久久v久久| 乱码一卡2卡4卡精品| 国产又黄又爽又无遮挡在线| 99热这里只有是精品50| 美女被艹到高潮喷水动态| 国产午夜精品一二区理论片| 中国美白少妇内射xxxbb| av在线亚洲专区| 小蜜桃在线观看免费完整版高清| 免费av观看视频| 人人妻人人看人人澡| 麻豆成人av视频| 国产黄片美女视频| 国产精品一区二区性色av| 亚洲精品乱码久久久久久按摩| 欧美又色又爽又黄视频| av在线观看视频网站免费| 精华霜和精华液先用哪个| 搡女人真爽免费视频火全软件| 久久久久久伊人网av| 久久人人爽人人爽人人片va| av又黄又爽大尺度在线免费看 | 狠狠狠狠99中文字幕| 国产精品精品国产色婷婷| av在线天堂中文字幕| 综合色av麻豆| 日本爱情动作片www.在线观看| 性插视频无遮挡在线免费观看| 卡戴珊不雅视频在线播放| 亚洲国产精品成人综合色| 免费搜索国产男女视频| 国产亚洲精品久久久com| 国产中年淑女户外野战色| 精华霜和精华液先用哪个| 男女下面进入的视频免费午夜| 国产伦精品一区二区三区视频9| 日韩制服骚丝袜av| 免费电影在线观看免费观看| 国产亚洲最大av| 国产精品女同一区二区软件| 女人十人毛片免费观看3o分钟| 国模一区二区三区四区视频| 亚洲国产色片| 桃色一区二区三区在线观看| 国产在线一区二区三区精 | 波野结衣二区三区在线| 久久精品91蜜桃| 久久欧美精品欧美久久欧美| 欧美日本亚洲视频在线播放| 搡女人真爽免费视频火全软件| 国产一区二区在线av高清观看| 日韩成人av中文字幕在线观看| 日本免费a在线| 禁无遮挡网站| 桃色一区二区三区在线观看| 国产av一区在线观看免费| 日韩一区二区视频免费看| 久久久久久久久久成人| 夫妻性生交免费视频一级片| 男的添女的下面高潮视频| 日韩中字成人| 国产av不卡久久| av在线播放精品| a级毛色黄片| 久久久久久久国产电影| 国产一区有黄有色的免费视频 | 男插女下体视频免费在线播放| 51国产日韩欧美| 午夜精品国产一区二区电影 | 中文天堂在线官网| 国内揄拍国产精品人妻在线| 日本猛色少妇xxxxx猛交久久| 两个人视频免费观看高清| 亚洲人与动物交配视频| 99久久人妻综合| 日韩在线高清观看一区二区三区| 人人妻人人澡人人爽人人夜夜 | 国产精品久久电影中文字幕| 夜夜爽夜夜爽视频| 国产毛片a区久久久久| 九草在线视频观看| 人体艺术视频欧美日本| 男女啪啪激烈高潮av片| 婷婷色综合大香蕉| 亚洲熟妇中文字幕五十中出| 一级二级三级毛片免费看| 亚洲高清免费不卡视频| 水蜜桃什么品种好| 蜜桃久久精品国产亚洲av| 免费av观看视频| 国产成人精品久久久久久| 在线观看一区二区三区| 中文字幕熟女人妻在线| 免费观看的影片在线观看| 国产69精品久久久久777片| 国产高清三级在线| 岛国毛片在线播放| 日韩精品青青久久久久久| 免费电影在线观看免费观看| 欧美zozozo另类| 非洲黑人性xxxx精品又粗又长| 久久久久久久亚洲中文字幕| 久久精品国产亚洲av涩爱| 亚洲激情五月婷婷啪啪| 黑人高潮一二区| 国产高清有码在线观看视频| 亚洲成人久久爱视频| 蜜桃久久精品国产亚洲av| 国产伦理片在线播放av一区| 大话2 男鬼变身卡| 亚洲美女视频黄频| av在线天堂中文字幕| 国产淫片久久久久久久久| 精品久久久久久久久久久久久| 欧美+日韩+精品| 日韩av不卡免费在线播放| 亚洲av成人av| 亚洲国产最新在线播放| 国产亚洲一区二区精品| 亚洲丝袜综合中文字幕| 你懂的网址亚洲精品在线观看 | 能在线免费观看的黄片| 日韩欧美国产在线观看| 日韩国内少妇激情av| 日韩人妻高清精品专区| 国产成人免费观看mmmm| 一区二区三区四区激情视频| 建设人人有责人人尽责人人享有的 | 黄片wwwwww| 亚洲精品国产成人久久av| 成人性生交大片免费视频hd| av黄色大香蕉| 成人高潮视频无遮挡免费网站| 亚洲真实伦在线观看| 床上黄色一级片| 最近手机中文字幕大全| 亚洲欧美精品综合久久99| 97超碰精品成人国产| 大话2 男鬼变身卡| 国产精品久久久久久精品电影| 日韩成人av中文字幕在线观看| 三级毛片av免费| av在线老鸭窝| 人人妻人人澡人人爽人人夜夜 | 久久久久国产网址| 日韩精品青青久久久久久| 免费电影在线观看免费观看| 少妇猛男粗大的猛烈进出视频 | 中文字幕av在线有码专区| 欧美最新免费一区二区三区| 99热这里只有是精品在线观看| 国产免费又黄又爽又色| 九九在线视频观看精品| 日本wwww免费看| 国产高清有码在线观看视频| 男人狂女人下面高潮的视频| 91精品一卡2卡3卡4卡| 成人毛片a级毛片在线播放| 久久久久久久久大av| 日日摸夜夜添夜夜爱| 中文字幕制服av| 亚洲国产精品合色在线| 亚洲精品,欧美精品| 国产精品日韩av在线免费观看| 男人的好看免费观看在线视频| 免费黄网站久久成人精品| 国产精品乱码一区二三区的特点| 人人妻人人澡欧美一区二区| 国产国拍精品亚洲av在线观看| 性色avwww在线观看| 午夜福利成人在线免费观看| 欧美+日韩+精品| 欧美日本亚洲视频在线播放| 亚洲最大成人av| 少妇人妻精品综合一区二区| 啦啦啦韩国在线观看视频| 亚洲中文字幕一区二区三区有码在线看| 久久99热这里只有精品18| 校园人妻丝袜中文字幕| 国产视频内射| 国产又色又爽无遮挡免| 精品一区二区三区人妻视频| 免费av毛片视频| 亚洲五月天丁香| 午夜福利成人在线免费观看| 天天一区二区日本电影三级| 日本色播在线视频| 一夜夜www| 久久草成人影院| 少妇裸体淫交视频免费看高清| av女优亚洲男人天堂| 夜夜爽夜夜爽视频| 国产成人精品久久久久久| 最近最新中文字幕免费大全7| 搡老妇女老女人老熟妇| 久久久久性生活片| 亚洲av中文字字幕乱码综合| 国产乱人视频| a级毛片免费高清观看在线播放| 日韩 亚洲 欧美在线| 联通29元200g的流量卡| 寂寞人妻少妇视频99o| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 精品免费久久久久久久清纯| 免费在线观看成人毛片| 国产在线男女| 久久久午夜欧美精品| 久久久精品欧美日韩精品| 日韩 亚洲 欧美在线| av专区在线播放| 日韩一本色道免费dvd| АⅤ资源中文在线天堂| 1024手机看黄色片| 国产淫语在线视频| 久久精品人妻少妇| 国产高清视频在线观看网站| 午夜精品在线福利| 成人欧美大片| ponron亚洲| 久久久精品94久久精品| 久99久视频精品免费| 少妇的逼水好多| 国产三级中文精品| 日韩欧美三级三区| 亚洲国产最新在线播放| 国产91av在线免费观看| 久久久久久久久久久免费av| 国产 一区精品| 欧美一区二区亚洲| 久热久热在线精品观看| www.av在线官网国产| 国产私拍福利视频在线观看| 日韩亚洲欧美综合| 欧美97在线视频| 亚洲av日韩在线播放| 美女xxoo啪啪120秒动态图| 国产精品蜜桃在线观看| 欧美+日韩+精品| 国产一区二区在线观看日韩| 亚洲av熟女| 中文字幕人妻熟人妻熟丝袜美| 亚洲av日韩在线播放| 亚洲欧洲日产国产| 国产成人aa在线观看| 在线播放无遮挡| 亚洲激情五月婷婷啪啪| 嫩草影院新地址| 欧美高清性xxxxhd video| 高清午夜精品一区二区三区| 黑人高潮一二区| 亚洲欧美中文字幕日韩二区| 久久久欧美国产精品| 国产一区二区三区av在线| 日日摸夜夜添夜夜爱| 精品人妻一区二区三区麻豆| 精品熟女少妇av免费看| 亚洲av成人av| 最近的中文字幕免费完整| 人人妻人人看人人澡| 亚洲成人中文字幕在线播放| 18禁在线无遮挡免费观看视频| 十八禁国产超污无遮挡网站| 男女啪啪激烈高潮av片| 日本av手机在线免费观看| 亚洲不卡免费看| 亚洲中文字幕一区二区三区有码在线看| 亚洲精品色激情综合| 国产真实乱freesex| 18+在线观看网站| 国产三级在线视频| 69人妻影院| av黄色大香蕉| 亚洲中文字幕一区二区三区有码在线看| 亚洲精品乱久久久久久| 久99久视频精品免费| 国产伦在线观看视频一区| 我的女老师完整版在线观看| 亚洲国产色片| 国产探花极品一区二区| 色吧在线观看| 午夜激情福利司机影院| 欧美+日韩+精品| 美女黄网站色视频| 国产乱来视频区| 在线a可以看的网站| 色哟哟·www| 亚洲美女搞黄在线观看| 日本黄大片高清| 观看美女的网站| 又爽又黄a免费视频| 久久精品人妻少妇| 色播亚洲综合网| 亚洲丝袜综合中文字幕| 国产 一区精品| 亚洲最大成人手机在线| 九草在线视频观看| 亚洲自偷自拍三级| 啦啦啦啦在线视频资源| 99视频精品全部免费 在线| 淫秽高清视频在线观看| 直男gayav资源| 91午夜精品亚洲一区二区三区| 少妇猛男粗大的猛烈进出视频 | 激情 狠狠 欧美| 久久久久久久久久久丰满| 男人舔奶头视频| 亚洲高清免费不卡视频| 亚洲精华国产精华液的使用体验| 五月伊人婷婷丁香| 久久久久网色| 成人高潮视频无遮挡免费网站| 黑人高潮一二区| 免费看av在线观看网站| 国产精品久久久久久久电影| 我要搜黄色片| 国产成人精品一,二区| 哪个播放器可以免费观看大片| 一本一本综合久久| 黄色配什么色好看| 亚洲国产欧美在线一区| 99久国产av精品| 身体一侧抽搐| 免费看美女性在线毛片视频| 男女那种视频在线观看| 国产精品福利在线免费观看| 丰满少妇做爰视频| 亚洲无线观看免费| 国产免费福利视频在线观看| 亚洲国产欧美人成| 九九爱精品视频在线观看| ponron亚洲| 久久韩国三级中文字幕| 日韩一区二区视频免费看| 热99在线观看视频| 麻豆精品久久久久久蜜桃| 国产精品久久久久久久电影| 女人十人毛片免费观看3o分钟| 成人欧美大片| 免费av毛片视频| 欧美3d第一页| 寂寞人妻少妇视频99o| 亚洲aⅴ乱码一区二区在线播放| 天堂av国产一区二区熟女人妻| 精品国产三级普通话版| 精品无人区乱码1区二区| 最近中文字幕2019免费版| 免费黄色在线免费观看| a级毛色黄片| 国产av码专区亚洲av| 我要搜黄色片| 亚洲在线自拍视频| 国产精品福利在线免费观看| 97热精品久久久久久| 日韩欧美在线乱码| 亚洲一级一片aⅴ在线观看| 美女高潮的动态| 1000部很黄的大片| 亚洲精品影视一区二区三区av| 国产欧美日韩精品一区二区| 国产伦精品一区二区三区视频9| 少妇熟女欧美另类| 国产精品久久久久久久电影| 日本爱情动作片www.在线观看| 国产日韩欧美在线精品| 亚洲精品成人久久久久久| 狂野欧美白嫩少妇大欣赏| 看十八女毛片水多多多| 舔av片在线| 国产精品永久免费网站| 蜜臀久久99精品久久宅男| 日本一二三区视频观看| 内射极品少妇av片p| 国产成人福利小说| 亚洲国产精品久久男人天堂| 欧美日本亚洲视频在线播放| 春色校园在线视频观看| 亚洲在久久综合| 又粗又硬又长又爽又黄的视频| 国产精品久久久久久久久免| 天天躁夜夜躁狠狠久久av| 一边摸一边抽搐一进一小说| 国产成人freesex在线| 一边亲一边摸免费视频| 国产综合懂色| 自拍偷自拍亚洲精品老妇| 在线免费十八禁| 日本免费在线观看一区| 毛片一级片免费看久久久久| 一个人看视频在线观看www免费| 亚洲国产欧美在线一区| 中文字幕亚洲精品专区| 汤姆久久久久久久影院中文字幕 | www日本黄色视频网| 少妇熟女欧美另类| 成人毛片a级毛片在线播放| 亚洲va在线va天堂va国产| 国产精品国产高清国产av| 久久综合国产亚洲精品| 久久久久久久久中文| 国产成人福利小说| 又黄又爽又刺激的免费视频.| eeuss影院久久| 亚洲一区高清亚洲精品| 纵有疾风起免费观看全集完整版 | 美女黄网站色视频| 精品久久久久久成人av| 亚洲精品亚洲一区二区| 亚洲国产精品成人久久小说| 夜夜看夜夜爽夜夜摸| 99国产精品一区二区蜜桃av| 国产成人一区二区在线| 久久久久久国产a免费观看| 国产探花在线观看一区二区| 精品国产三级普通话版| 久久草成人影院| 2022亚洲国产成人精品| 最近2019中文字幕mv第一页| 精品一区二区三区人妻视频| 国语对白做爰xxxⅹ性视频网站| 黄色一级大片看看| 超碰av人人做人人爽久久| 18+在线观看网站| 欧美一区二区国产精品久久精品| 久久人妻av系列| 免费av毛片视频| 亚洲图色成人| 久久久久精品久久久久真实原创| 日韩 亚洲 欧美在线| 欧美三级亚洲精品| 欧美成人a在线观看| 伦理电影大哥的女人| 成人漫画全彩无遮挡| 精品久久久噜噜| 免费播放大片免费观看视频在线观看 | 我的老师免费观看完整版| 亚洲最大成人av| 午夜视频国产福利| 亚洲av男天堂| 精品无人区乱码1区二区| 国产免费男女视频| 26uuu在线亚洲综合色| 天天躁日日操中文字幕| 一级av片app| 三级国产精品欧美在线观看| 99在线视频只有这里精品首页| 我的老师免费观看完整版| 禁无遮挡网站| 国产精品国产高清国产av| 卡戴珊不雅视频在线播放| 人妻制服诱惑在线中文字幕| 日韩成人av中文字幕在线观看| 国产伦在线观看视频一区| 免费观看人在逋| 精品少妇黑人巨大在线播放 | 国产在线男女| 精品久久久久久久久av| 99热精品在线国产| 中文精品一卡2卡3卡4更新| 国产又黄又爽又无遮挡在线| 亚洲av男天堂| 精品人妻一区二区三区麻豆| 国产精品av视频在线免费观看| 国产精品永久免费网站| 菩萨蛮人人尽说江南好唐韦庄 | 女人久久www免费人成看片 | 欧美日韩国产亚洲二区| 超碰av人人做人人爽久久| 日日摸夜夜添夜夜添av毛片| 九九热线精品视视频播放| 国产午夜精品一二区理论片| 我要看日韩黄色一级片| 色5月婷婷丁香| 校园人妻丝袜中文字幕| 中文字幕av在线有码专区| 亚洲av免费在线观看| 最近最新中文字幕大全电影3| 日韩欧美三级三区| 久久久a久久爽久久v久久| 国产成人免费观看mmmm| 真实男女啪啪啪动态图| 亚州av有码| 99热精品在线国产| 激情 狠狠 欧美| 国产探花极品一区二区| 成人午夜高清在线视频| 久久久久久久久久久免费av| 亚洲精品影视一区二区三区av| 老司机影院毛片| 亚洲精品国产av成人精品| 亚洲第一区二区三区不卡| 亚洲欧美中文字幕日韩二区| 国产爱豆传媒在线观看| 午夜激情欧美在线| 亚洲成人精品中文字幕电影| 亚洲欧美精品综合久久99| 只有这里有精品99| 成人性生交大片免费视频hd| 全区人妻精品视频| 亚洲av免费高清在线观看| 网址你懂的国产日韩在线| 18禁在线播放成人免费| 欧美激情国产日韩精品一区| 欧美激情在线99| 一级黄色大片毛片| 一本一本综合久久| 亚洲精品乱码久久久久久按摩| 成人鲁丝片一二三区免费| 最新中文字幕久久久久| 男女那种视频在线观看|