• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    RF-Net:Unsupervised Low-Light Image Enhancement Based on Retinex and Exposure Fusion

    2023-12-12 15:51:14TianMaChenhuiFuJiayiYangJiehuiZhangandChuyangShang
    Computers Materials&Continua 2023年10期

    Tian Ma,Chenhui Fu,Jiayi Yang,Jiehui Zhang and Chuyang Shang

    College of Computer Science and Technology,Xi’an University of Science and Technology,Xi’an,710054,China

    ABSTRACT Low-light image enhancement methods have limitations in addressing issues such as color distortion,lack of vibrancy,and uneven light distribution and often require paired training data.To address these issues,we propose a two-stage unsupervised low-light image enhancement algorithm called Retinex and Exposure Fusion Network(RFNet),which can overcome the problems of over-enhancement of the high dynamic range and under-enhancement of the low dynamic range in existing enhancement algorithms.This algorithm can better manage the challenges brought about by complex environments in real-world scenarios by training with unpaired low-light images and regular-light images.In the first stage,we design a multi-scale feature extraction module based on Retinex theory,capable of extracting details and structural information at different scales to generate high-quality illumination and reflection images.In the second stage,an exposure image generator is designed through the camera response mechanism function to acquire exposure images containing more dark features,and the generated images are fused with the original input images to complete the low-light image enhancement.Experiments show the effectiveness and rationality of each module designed in this paper.And the method reconstructs the details of contrast and color distribution,outperforms the current state-of-the-art methods in both qualitative and quantitative metrics,and shows excellent performance in the real world.

    KEYWORDS Low-light image enhancement;multiscale feature extraction module;exposure generator;exposure fusion

    1 Introduction

    With the rapid development of artificial intelligence,low-light image-enhancement technology has been widely applied for pre-processing in advanced visual tasks.However,low-light images often suffer from detail degradation and color distortion due to the shooting environment and technical limitations.Balancing the image-enhancement effect and maintaining image realism are challenging problems in low-light image enhancement.These problems can significantly affect the performance of advanced downstream vision tasks.Therefore,improving visual quality and recovering image details have become important research topics.

    In the process of image enhancement,it is important to strike a balance between preserving image details and maintaining overall image quality.This requires preserving the original details in well-exposed areas,while appropriately brightening the underexposed areas to achieve a high-quality image.In addition,attention must be paid to balancing the brightness and contrast of the image during enhancement.If only the brightness is increased globally,the texture details in the image may be lost.Therefore,both brightness and contrast changes must be considered when enhancing an image to ensure its quality.Traditional methods [1–3] in the past often required a large amount of manual parameter adjustment to improve image quality;however,this method has significant limitations,as its effectiveness is largely based on assumptions regarding the threshold range.The use of low-light/normal-light images in supervised deep model training has become the main approach in algorithm research owing to advancements in deep learning.The accuracy of supervised learning methods depends on paired training datasets.However,it is technically difficult to obtain paired datasets from the same scene.In addition,the algorithm has poor generalization ability and cannot be effectively applied to real-scene images.In recent years,unsupervised image enhancement algorithms have emerged that eliminate reliance on paired datasets and achieve good enhancement results.For example,Deep Light Enhancement without Paired Supervision (EnlightenGAN) [4] uses unpaired datasets to train and implement low-light image enhancement techniques.Zero-Reference Deep Curve Estimation (Zero-DCE) [5] achieves enhancement using scene images with different illumination intensities.Although these methods eliminate the dependence of deep learning techniques on paired datasets,the quality of enhancement remains a challenge.The EnlightenGAN[4]method may produce artifacts,an overall uneven picture,and color-recovery errors when enhancing dark areas.Images enhanced using the Zero-DCE [5] method may exhibit whitish tones and less vibrant colors.These methods exhibit stronger generalization ability than supervised methods and reduce the requirements for dataset collection.

    To address these issues,we propose an unsupervised enhancement network called RF-Net,which combines Retinex with exposure fusion.The network comprises two stages: image decomposition and exposure fusion.In the first stage,to fully consider contextual and global information,we employed the powerful image-generation capabilities of a generative adversarial network and designed a multi-scale feature-extraction module to produce high-quality illumination and reflection images.Specifically,our network uses a multi-scale feature extraction module to perceptively capture differentscale features,preserve more detailed information,and avoid information loss between layers by using residual connections to transmit information from the current layer to the next layer.Most existing Retinex-based image enhancement methods obtain illumination and reflection component information matrices and generate enhanced images through calculations,which not only involve high computational complexity but also result in artifacts when processing shadow parts in dark areas.After obtaining the illumination and reflection images,an exposure image generator with correction coefficients was designed using the camera response function in the second stage to generate the exposure image and fuse it with the original low-light image to complete low-light image enhancement.The results obtained using the proposed method are shown in Fig.1.

    In summary,the main contributions of this paper are as follows:

    1.We devised a multi-scale feature-extraction module to produce high-quality illumination and reflection images.We incorporated a Coordinate Attention (CA) module that includes position-encoding information into the Markov discriminator.This module builds on channel attention and pays closer attention to the location information of the generated image,allowing for more accurate discrimination of texture details and improving the quality of the images generated.

    2.We improved the original Retinex formulation and designed an exposure image generator module with correction coefficients by referring to the camera response mechanism function.This module can generate images with different exposure levels while fusing illumination and reflection images.

    3.We proposed a novel unsupervised image enhancement method called RF-Net,which exhibits excellent performance in test results on several datasets and can be generalized to real-world low-light conditions.

    Figure 1:Representative enhancement results of RF-Net.Which can improve over-enhancement in the high dynamic range and under-enhancement in the low dynamic range

    2 Related Work

    In this section,we review research on low-light image enhancement using traditional and deep learning methods.

    2.1 Traditional Methods

    Histogram equalization is a classical image-enhancement method that enhances the contrast of an image by adjusting its brightness distribution.However,histogram equalization tends to cause image noise and over-enhancement problems.Some methods further increase the enhancement effect by setting a threshold to divide the image blocks [6],dividing the clipping points into chunks for processing[7],and combining them with adaptive gamma correction[8]to obtain a more reasonable S-shaped mapping function.However,these methods lead to problems of over-enhancement and amplification artifacts.The contrast was enhanced to an extent,but the details were lost.Retinex theory [9] posits that an image’s brightness is composed of two parts,reflection,and illumination,which enhance the image quality by separating these two parts.However,this approach is ineffective for images that are too dark or bright.Accordingly,researchers have proposed various improvement schemes[1–3],etc.This presupposes that spatial illumination changes slowly during implementation,but the processing is prone to halation and inaccurate color recovery.To reduce the computational cost in Retinex theory,researchers have proposed Low-Light Image Enhancement via Illumination Map Estimation(LIME)[10],the local illumination distribution of the image obtained by analyzing the local information.The local illumination distribution is applied to the reflection component to obtain the enhanced image.Compared with the Retinex method,LIME reduces the occurrence of halo artifacts during processing.However,LIME has a limited ability to distinguish between the foreground and background of an image,which can result in over-enhancement of the foreground and noise in the background.

    2.2 Deep Learning Methods

    Researchers have widely employed deep learning for image enhancement over the past decade,achieving promising results.In the following section,we review the current state of research on fully supervised,semi-supervised,and unsupervised approaches.

    2.2.1 Fully Supervised Methods

    The use of paired datasets to train network models has been widely adopted because of the oneto-one correspondence between the training data.RetinexNet[11],for the first time,combines Retinex theory with Convolutional Neural Networks to implement the low-illumination image enhancement problem by designing a decomposition module,and enhancement module.The feasibility of the Retinex application in deep learning was demonstrated for the first time.Kindling the Darkness(KinD)[12]designed global and local enhancement modules,where the global module extracts global luminance based on Retinex decomposition,and the local module enhances the image texture details.The global and local modules interacted through an adaptive mechanism,and the difference between the image generated by the global enhancement branch and the original image was used to calculate the weight of each pixel.These weights were then passed to the local enhancement branch to achieve contrast enhancement.KinD++[13]is based on KinD and improves the training speed and accuracy of the model by designing group learning and back-propagation mechanisms.GLobal Illuminationaware and Detail preserving Network (GLADNet) [14] generates a global before light by designing a global illumination estimation module that is then combined with the original input image to produce an enhanced image.Low-Light Image Enhancement with Normalizing Flow(LLFlow)[15]uses adaptive weights to control the effects of optical flow and global constraints;it also uses a deep learning model to learn the optical flow and the image to obtain an enhanced image.Self-Calibrated Illumination(SCI)[16]reduces computational costs by designing an adaptive correction illumination module that ensures the convergence of the results of each training phase to the final one.The literature[17]designs a generative adversarial network containing dual attention units that can effectively inhibit the artifacts and color reproduction bias generated during the enhancement.Transformer Photo Enhancement (TPE) [18] uses a pure transformer architecture to implement image enhancement based on multi-stage curve adjustment.Retinex based deep unfolding network (URetinex-Net) [19]decomposes the input image by designing a continuous optimization model with mutual feedback.To optimize the decomposition results,an implicit a priori regularization model was used,and a data initialization module,specific illumination intensity module,and denoising detail retention module were designed.Illumination Adaptive Transformer(IAT)[20]implements low-light enhancement by designing a lightweight transformer model that uses attention-query techniques to represent and adjust the parameters associated with the image signal processor(ISP).

    2.2.2 Semi-Supervised Methods

    These methods can learn better feature representations using both paired and unpaired data.First,researchers use paired datasets to train and obtain prior knowledge,and then they use the trained model as pre-training weights for unpaired data training.Based on this subdivision,Deep Recursive Band Network(DRBN)[21]introduces a recursive network architecture that uses the information of the highlight and shadow regions of the image and constructs a low-rank matrix and a sparse matrix to represent the brightness and structural information of the image,then inputs the two matrices into two branches of the network for feature extraction,and finally merges them to obtain an enhanced image.DRBN [22] utilizes a “band representation” technique to enhance low-light images.This method decomposes a low-light image into multiple bands using band representation and trains a neural network with a small amount of labeled data to learn how to enhance each band.Thus,this method retains the detail and texture information of a low-light image,thereby enhancing its quality.

    2.2.3 Unsupervised Methods

    Obtaining paired datasets can be difficult and using unsupervised methods has become the main approach for accomplishing image-enhancement tasks.This approach improves generality and applicability to many real-world scenarios.Exposure Correction Network (ExCNet) [23] is the first unsupervised enhancement method that uses the powerful learning ability of neural networks to estimate the most suitable“S”curve for low-light images and uses this curve directly to enhance the image.low-light image enhancement network(LEGAN)[24]is enhanced by a cleverly designed light perception module and a loss function that solves the overexposure problem.The EnlightenGAN[4]completed its first unsupervised image enhancement using unpaired datasets.This design overcomes the previous reliance on paired datasets by establishing unpaired mappings between low-light and nonmatching images and employing global-local discriminators and feature retention losses to constrain the feature distance between the enhanced and origin images for enhancement.Zero-DCE[5]uses a neural network to match a brightness mapping curve and then generates an enhanced image based on the curve.Unsupervised low-light image enhancement was achieved by designing a multi-stage highorder curve with a pixel-level dynamic range adjustment.Based on this,a lightweight version,Zero-DCE++[25],was developed.Generative adversarial network and Retinex (RetinexGAN) [26] uses a generative adversarial network based on Retinex to design a decomposition network with a simple two-layer convolution and achieve low-light image enhancement through image fusion.Restoration of Underexposed Images via Robust Retinex Decomposition (RRDNet) [27] achieves enhancement by designing a three-branch decomposition network and iterative loss functions to decompose the three components of reflection,illumination,and noise module.Retinex-inspired Unrolling with Architecture Search(RUAS)[28]used neural structure search to find an effective and lightweight set of networks for low-light enhancement.Retinex Deep Image Prior(RetinexDIP)[29]proposes a Retinexbased generation strategy that reduces the coupling between two components in the decomposition,making it easier to adjust the estimated illumination to perform an enhancement.

    3 Proposed Method

    In this section,the first module introduces the overall structure of the RF-Net network and provides a hierarchical representation of the first-and second-stage exposure-generation fusion networks.The second module describes the designed exposure image generator,and the third module describes the loss function.

    3.1 Network Architecture Design

    The proposed RF-Net is a two-stage network with the overall structure shown in Fig.2.In the first stage,two coupled-generator frameworks were used.The R network generates reflection images,whereas the L network generates illumination images.First,by cascading the maximum,minimum,and mean values of each channel of the original low-light image as inputs to the network,a multi-scale feature extraction module was designed to maintain the global consistency of good illumination and contextual information.After basic feature extraction,the resulting features are concatenated and mapped to high-dimensional information before being reduced in dimension through convolution,thereby improving image quality while learning complex features.For the discriminator,we used a VGG-based network structure in which the original Markov discriminator maps the input to an N×N matrix,such that each point in the matrix corresponds to the evaluation value of each region.To enhance the decomposition effect and discrimination accuracy of the network,we incorporated a CA attention mechanism with positional information[30].This further enhances the network’s ability to perceive and understand images and spatial information,learn useful features,and suppress irrelevant features,thereby improving the discriminator’s ability to accurately judge texture details.

    Figure 2:Overview of RF-Net.First,the low-light image and the corresponding maximum,minimum,and average grayscale images are input.The R and L have the same structure and are used to acquire the illumination and reflection components,respectively.Then the exposure image is acquired by the exposure generator.Finally,the low-light image is fused with the exposure image

    The second stage also consists of two coupled networks that use the original input image and the output of the first-stage network as inputs.At this stage,the output information from the first stage is processed by the exposure image generator module to create the initial exposure image.The details of this module are discussed in Section 2.The exposed image and original input image were then separately fed into two-branch networks[31].The two branches consist of a feature extraction module(FE),a super-resolution module(SR),and a feature fusion module(CF).The FE module consists of two convolution layers: SR uses the Convolutional Networks for Biomedical Image Segmentation(U-net) structure to learn more advanced features,and the first two modules are used to extract advanced features from the input low-dynamic-range images.RE was employed for super-resolution of the original input image before fusion to ensure the accurate extraction of high-level image features.The final module is the image fusion module that combines the super-resolution outputs of the two coupled networks and generates the output by weighting the super-resolution of the original image and the outputs of the two coupled blocks.

    3.2 Exposure Generation Module Design

    We employed a design that combines Retinex with the camera response mechanism to create an exposure image generator.In the original Retinex theory,the input imageSoutputis represented as the element-wise product of the illumination and reflectance components,which is expressed as Eq.(1):

    The input image is denoted as S,the reflected image as R,the illuminated image as L,andR×Ldenotes pixel-wise multiplication.However,numerous experiments have shown that the results obtained from the original Retinex equation are over-enhanced and lose detail owing to noise and uneven illumination.Therefore,we improved the original formula by first inverting the source illumination image using Eq.(2)to better utilize the content in the relatively overexposed region.

    The improved Retinex formula is represented by Eq.(3).

    where L and R denote the illuminated and reflected images,respectively,S denotes the original input low-light image,andSoudenotes the output result of the improved Retinex formula.

    To maintain a balance between brightness and contrast,we re-designed the improved Retinex formula using a camera response mechanism and proposed an exposure image generator.Here,we refer to the camera response function described in [32],where the model parameters of the camera response mechanism are determined by the camera’s parametersα,βand k.Parameter k is a correction factor that can be adjusted to obtain images with different exposure levels.As the k value increases,a brighter exposed image is acquired,and the details in the low-light areas become more significant;however,when the k value is too large,more detailed information is lost because of an exposure level that is too high.Therefore,we limited the value of k to a range of 2–6.The equation for generating the initial exposure image by combining the improved Retinex and camera response functions is expressed as Eq.(4).

    whereαandβare fixed parameters suitable for most cameras,withαset to -0.3293 andβset to 1.1258.Seorepresents the output exposure image,and different values of k directly affect the resulting output ofSeo.

    3.3 Loss Function

    Adversarial loss: In [11,12],by decomposing a low-light image into illumination and reflection components,the illumination components are approximately the same as those decomposed in a normally exposed image.With only differences in brightness,the reflection component is the same as the reflection component decomposed from the normal exposure image,which can be decomposed into high-quality reflection components by noise reduction.This means that the distribution of normally exposed images is very similar to that of the original images based on Retinex decomposition.Therefore,the original function [33] was used as an adversarial loss function to train the generator.In practical applications,the generated fake and real input samples are encoded as zero and one,respectively.Discriminators for the illumination and reflection maps were trained using squared error as the objective function.Our adversarial losses are defined by Eqs.(5)–(8).

    wheregLandgRdenote the reflected and illuminated images,respectively,denotes the average grey scale value of each channel.y represents the ground-truth reflection mapDLandDRrepresent the discriminators for the illumination and reflection maps,respectively.

    However, these aircraft had the familiar P-51 black paint with white stripes on the wings and were equipped with the wing tanks for extra range. Suddenly, they dropped their tanks just off to our right, and we looked around for German fighters in the area. We found them, when the whole formation of P-51s turned out to be Luftwaffe ME-109s that turned in to us with their cannons10 blazing! We narrowly missed being rammed11 by two of them that just barely passed over us.

    Perceptual loss: Using non-matching images for unsupervised image enhancement implies that the pixels in the training images are not one-to-one.The same pixel may have different semantics in different images.Therefore,we need a loss function to address the issue of non-corresponding pixel positions,and the perceptual loss function serves this purpose.This function is typically defined in the activation layer of a pre-trained network.By computing the distance between the activation layer features,it can effectively quantify the fundamental attributes of an image as well as the differences between its detailed features and high-level semantic information.This lays the foundation for generating high-quality images.Unlike common perceptual losses,this study adopted the concept of perceptual loss from[31].This implementation not only maintains the luminance consistency between the original and reference images but also recovers the details better.The perceptual loss function used in this paper is given by Eq.(9).

    where denotes the features extracted by the predefined network VGG19.CiHiWidenotes the number of channels and the height and width of the feature map in layer i.y denotes the unpaired real image information learned by the discriminator,g denotes the image generated by the generator.

    The total loss function of the network is shown in Eq.(10).

    4 Experiment

    4.1 Experimental Details

    We trained the RF-Net model on 914 randomly selected pairs of asymmetric datasets using the datasets provided in[4]and tested it on various datasets,including NPE[34],DICM[35],LIME[10],MEF[36],and VV.These datasets contain various low-light and unevenly exposed images from both indoor and outdoor settings.To demonstrate the performance of the enhancement algorithm better,we selected images with significant exposure differences from each test set as test set,which made the test more challenging.The deep learning framework used was PyTorch,and the hardware configuration was Tesla A100.

    To ensure that the training of the network could fully utilize the computing and storage resources of the computer and obtain better training results,the size of the training image was 640×400,and we randomly cropped the training data into a face slice of size 300×300.The batch size was set to one.To increase the data diversity,data expansion was performed,including random flipping,rotation,and cropping.This allowed the network to adapt better to various image scenes.The Adam optimizer was used to optimize the network,and the learning rate was set to 1e-4.Our network achieved better enhancement results with no more than 50 training iterations.

    4.2 Performance Evaluation

    To demonstrate the advantages of our proposed RF-Net,we compared our method with 10 other advanced methods:RetinexNet[11],EnlightenGAN[4],KinD[12],KinD++[13]RUAS[28],Zero-DCE [5],LLFlow [15],SCI [16],IAT [20],and GLADNet [14].Among these,RetinexNet,KinD,KinD++,RUAS,LLFlow,SCI,IAT,and GLADNet are supervised enhancement methods,whereas[4] and [5] are unsupervised enhancement methods.To ensure fairness,tests were conducted using the network parameters recommended in each study.Because we were unable to train the supervised models on unpaired datasets,we evaluated them using pre-trained models saved in the original papers.For unsupervised methods,if the original study used unpaired datasets for training,we used the datasets provided by [4].If the method used images with different exposures for training,we used the datasets provided by[5].Finally,the optimal model was selected for testing.These comparisons enabled us to evaluate the performance of the RF-Net method and demonstrate its competitiveness in image enhancement.

    4.2.1 Qualitative Comparison

    Figure 3:Qualitative comparison of RF-Net with other advanced algorithms.See the patch area for more detailed information

    Figure 4:(Continued)

    Figure 4:Qualitative comparison of RF-Net with other advanced algorithms.See the patch area for more detailed information

    Figure 5:Qualitative comparison of RF-Net with other advanced algorithms.See the patch area for more detailed information

    4.2.2 Quantitative Comparison

    The subjective evaluation may not be sufficient for determining the degree of detail retention during image enhancement.To demonstrate the feasibility of the proposed method further,we conducted quantitative comparisons.As we used unsupervised methods for model training,we could not evaluate the Peak Signal-to-Noise Ratio(PSNR)and Structural Similarity(SSIM)of the enhanced images to the ground truth,as with other supervised methods.Therefore,we used a non-referenced image quality assessment metric to compare the RF-Net method with other competitors.The metrics evaluated were Natural Image Quality Evaluator(NIQE)[37]and Blind/Referenceless Image Spatial QUality Evaluator(BRISQUE)[38].NIQE is a natural-image-based evaluation metric that compares algorithm processing results with a model calculated based on natural scenes.BRISQUE is an image-based no-reference quality score that is calculated based on natural scene images with similar distortions.The metrics used to evaluate the performance of RF-Net compared with the other algorithms are listed in Table 1.Based on the qualitative evaluation,the following issues were observed:RetinexNet[11]resulted in inaccurate color restoration with color bias;KinD[12]did not significantly enhance dark areas;KinD++[13]introduced artifacts while enhancing dark areas;RUAS[28]overenhanced the image,causing loss of information;LLFlow[15]produced incorrect color restoration;SCI[16]and IAT[20]over-enhanced the image and had insignificant enhancement in dark areas;Zero-DCE[5]had lower metrics compared to other unsupervised methods,possibly due to the whitish image produced by its enhancement results,as noted in the qualitative analysis.As shown in the graphs,the enhancement results of RF-Net validated this.In summary,because both NIQE and BRISQUE are methods based on local image statistical information,the impact of different algorithms on the small differences in the achieved metrics after image enhancement is better illustrated in the quantitative evaluation.

    Table 1:The NIQE(↓)and BRISQUE(↓)scores are shown,with lower scores indicating better image quality and richer information contained.The averages of the test image metrics are taken for each of the five datasets,and the five averages are eventually averaged again.The best result is shown in red and the second-ranked result is shown in blue

    4.2.3 Subjective Evaluation of People

    We also conducted a human subjective visual evaluation to further quantify the subjective quality of RF-Net compared to the other methods.We randomly selected 20 original low-light images from the test set(NPE[34],DICM[35],LIME[10],MEF[36],and VV)and applied four state-of-the-art methods (EnlightenGAN [4],Zero-DCE [5],SCI [16],and IAT [20]) to each image separately.We invited twelve reviewers to independently score the results of the five algorithms,including RF-Net.The reviewers primarily observed the following aspects:1.Whether the results contained artifacts in the over-or under-enhanced areas;2.Whether the color restoration in the results was accurate(e.g.,whether the colors were distorted);and 3.Whether the noise in dark areas was amplified,and whether there was an obvious loss of texture details in the results.As can be observed from the statistics in Fig.6,RF-Net achieved a higher subjective evaluation score for the reviewed images.

    Figure 6:Overview of people’s objective evaluation,in each graph,the x-axis indicates the observed quality of the five algorithms(1 for the best and 5 for the worst)and the y-axis indicates the number of good and bad images corresponding to m each algorithm.RF-Net shows the best performance

    4.3 Ablation Study

    To demonstrate the effectiveness of the modules used,the following ablation studies were conducted separately: three studies were designed in terms of CA removal,separation of inception[39]from residual connectivity[40],and direct fusion of illuminated and reflected components.

    In Ablation Study 1,we experimented with the multiscale module,residual connection,and CA attention separately,as shown in Fig.7,we can observe from the visual results that zooming in on the person above the house in the first image reveals that using only inception and the Markov discriminator leads to a blurred task,whereas using both inception and residual linking with a Markov discriminator produces more vivid colors and preserves more detailed information owing to residual linking.Furthermore,using inception,residual linking,and an improved Markov discriminator leads to a more realistic restoration of texture details,because the Markov discriminator has a clearer ability to distinguish true from false.The reddish hue in our image is due to the reflected image generated by the first stage of the network,which does not include knowledge of the illumination image or fusion module.

    Figure 7:Ablation study for each module.(a)Input,(b)w/o Multiscale,(c)w/o Residual,(d)w/o CA

    In Ablation Study 2,we compared our method with the direct integration of light and reflection components,as shown in Fig.8.The color of the sky in the second column of the first row was inaccurately restored,and the face of the person in the second column of the second row was overexposed.The proposed method in the third column performs significantly better than the direct fusion method.

    In Ablation Study 3,we tested the designed exposure generator module by adjusting the value of k to obtain images with different exposure levels and achieved exposure fusion.As shown in Fig.9,for different input images,we acquired images with varying exposure levels for fusion,with the k values set differently in different scenarios.For example,in images without exposed areas,k values are usually set between 4–6,while in images with exposed areas,k values are typically set between 2–4.This approach achieved the best enhancement when the exposed images were fused with the original lowlight images.Finally,we conducted a quantitative evaluation of the three sets of ablation studies.The results in Table 2 show that adding the residual block to the inception and residual block combinations produced superior performance.In addition,exposure fusion in the second stage of RF-Net exhibited advantages.A comparison of the second and fourth experimental rows demonstrates the effectiveness of RF-Net.

    Table 2:Quantitative review of ablation studies for each module

    Figure 8:Direct fusion of illuminated and reflected components with RF-Net ablation study

    Figure 9:Image decomposition and exposure generation fusion ablation study

    In Ablation Study 4,we use the low-light image and its corresponding Red,Green,Blue(RGB)channels with the maximum,minimum,and average values as inputs to our network.The final network input consists of six channels,with the first three channels used to generate the reflection map and the last three channels used to generate the illumination map.In the generated illumination images,the method used in this paper retains more details in the final generated illumination map compared with the input of only the original low illumination image.The qualitative comparison in Fig.10 shows that in the first column,there are artifacts near the shoulders of the person in the illumination map generated by the network with only low illumination input.In the second column,there is a noticeable loss of detail in the wall of the house on the right.In the third column,there are artifacts near the candle flame,and the grayscale boundary near the cup in the upper left corner is not clear;In the fourth column,there is an excessive amount of detail in the cave,while the detail in the trees outside the cave is lacking or insufficient.In Fig.11,the grayscale histogram results show that our method can generate a light map with smoother contrast and luminance changes,making the light and dark parts more visible and details more prominent.

    Figure 10:Qualitative comparison of illuminated images.(a) Input,(b) Low-light images as input,(c)Ours

    5 Conclusion

    In this study,we combined Retinex theory with exposure fusion for the first time to achieve unpaired low-light image enhancement.In the first stage,we designed a multi-scale generator by combining a residual network and inception.We also added a CA attention mechanism with position information to the discriminator network to obtain high-quality illumination and reflection components.By improving the original Retinex and camera response mechanism functions,we designed an exposure image generator with correction coefficients to solve the problems of illumination,reflection image fusion,and exposure image generation.Based on this,we realized low-light image enhancement with second-stage exposure fusion and proved the superiority of our method by comparing it with state-of-the-art methods.However,the algorithm has some limitations.On the one hand,it requires manual adjustment of the correction parameters for the exposure image generator based on the scene’s exposure level.On the other hand,Compared to other lightweight networks,RF-Net processes 640 × 400 images at a rate of 5 frames per second.In the future,our research will focus on lightweight the network structure and developing an adaptive low-light image enhancement method based on negative feedback control to solve the manual tuning problem of existing methods.We also aim to apply this model to enhance specific scenes,which will not only improve the generalization of the algorithm but also enhance the accuracy of other vision tasks.

    Acknowledgement:The authors also gratefully acknowledge the helpful comments and suggestions of the reviewers,which have improved the presentation.

    Funding Statement:This work was supported by the National Key Research and Development Program Topics (Grant No.2021YFB4000905),the National Natural Science Foundation of China(Grant Nos.62101432 and 62102309),and in part by Shaanxi Natural Science Fundamental Research Program Project(No.2022JM-508).

    Author Contributions:Study conception and design:Tian Ma,Jiayi Yang;data collection:Chenhui Fu;analysis and interpretation of results:Chenhui Fu,Jiehui Zhang,Chuyang Shang;draft manuscript preparation: Chenhui Fu.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data used in this paper can be requested from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    18禁动态无遮挡网站| 久久久久久久大尺度免费视频| 日日啪夜夜爽| 亚洲av在线观看美女高潮| 国产色爽女视频免费观看| 日韩欧美一区视频在线观看 | 国产视频首页在线观看| 97在线视频观看| a级毛片免费高清观看在线播放| 亚洲av一区综合| 在线观看免费高清a一片| 国内精品宾馆在线| 麻豆精品久久久久久蜜桃| 免费黄网站久久成人精品| av黄色大香蕉| 国产综合懂色| 久久精品国产鲁丝片午夜精品| 久久精品熟女亚洲av麻豆精品| 亚洲精品中文字幕在线视频 | 在线观看av片永久免费下载| 国产高潮美女av| av播播在线观看一区| 亚洲av.av天堂| 丰满乱子伦码专区| 一区二区三区四区激情视频| 大香蕉97超碰在线| a级毛色黄片| 亚洲美女搞黄在线观看| 亚洲在线观看片| 欧美+日韩+精品| 欧美区成人在线视频| 亚洲激情五月婷婷啪啪| 最新中文字幕久久久久| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 亚洲av免费高清在线观看| 久久精品久久久久久久性| 特大巨黑吊av在线直播| 久久久久久久亚洲中文字幕| 中国美白少妇内射xxxbb| 免费av不卡在线播放| 欧美精品人与动牲交sv欧美| 3wmmmm亚洲av在线观看| 五月玫瑰六月丁香| 久久人人爽av亚洲精品天堂 | 麻豆乱淫一区二区| 亚洲图色成人| 亚洲av中文字字幕乱码综合| 精品酒店卫生间| 纵有疾风起免费观看全集完整版| 亚洲精品日韩av片在线观看| 亚洲自偷自拍三级| 国产综合精华液| 国产亚洲精品久久久com| 色视频www国产| 久久久a久久爽久久v久久| 国产伦理片在线播放av一区| 久久久久久久久久人人人人人人| 日韩一本色道免费dvd| 亚洲av成人精品一二三区| 国产精品三级大全| 国产成人a∨麻豆精品| 51国产日韩欧美| 亚洲美女搞黄在线观看| 国产又色又爽无遮挡免| 在线a可以看的网站| 久久久午夜欧美精品| 免费av毛片视频| 国产精品精品国产色婷婷| 亚洲综合精品二区| 美女高潮的动态| 又黄又爽又刺激的免费视频.| 麻豆成人av视频| 天堂中文最新版在线下载 | 午夜福利视频精品| 最近2019中文字幕mv第一页| 丝瓜视频免费看黄片| 亚洲人成网站在线观看播放| 夫妻性生交免费视频一级片| 国产黄色免费在线视频| 99热网站在线观看| 18禁在线播放成人免费| 毛片女人毛片| 18禁动态无遮挡网站| 少妇的逼水好多| 色网站视频免费| 国产精品成人在线| 精品午夜福利在线看| av国产精品久久久久影院| 亚洲欧美中文字幕日韩二区| 亚洲美女搞黄在线观看| 中国美白少妇内射xxxbb| 午夜福利在线观看免费完整高清在| 成人黄色视频免费在线看| 国产黄a三级三级三级人| av卡一久久| 99热6这里只有精品| 在线观看人妻少妇| 伊人久久精品亚洲午夜| 久久久久久九九精品二区国产| 日韩制服骚丝袜av| 欧美一区二区亚洲| 日韩亚洲欧美综合| 免费黄网站久久成人精品| 熟女av电影| 亚洲av日韩在线播放| 国产伦精品一区二区三区视频9| 国产视频首页在线观看| 亚洲四区av| 高清毛片免费看| 日日摸夜夜添夜夜添av毛片| 大话2 男鬼变身卡| 国产伦精品一区二区三区视频9| 深爱激情五月婷婷| 国产淫片久久久久久久久| av线在线观看网站| 亚洲,欧美,日韩| 成人毛片a级毛片在线播放| 精品久久久久久电影网| 麻豆久久精品国产亚洲av| 亚洲精品自拍成人| 亚洲av男天堂| 亚洲,一卡二卡三卡| 一级黄片播放器| 国产免费又黄又爽又色| 亚洲av二区三区四区| 日韩欧美精品v在线| 97在线视频观看| 男人和女人高潮做爰伦理| 亚洲不卡免费看| 搞女人的毛片| 亚洲激情五月婷婷啪啪| 欧美精品国产亚洲| av国产久精品久网站免费入址| 一级av片app| 国产精品不卡视频一区二区| 人妻夜夜爽99麻豆av| 精品一区二区免费观看| 精品少妇久久久久久888优播| 国产黄色视频一区二区在线观看| 久久久a久久爽久久v久久| 国产黄片美女视频| 深夜a级毛片| 国产成人福利小说| 大香蕉97超碰在线| 涩涩av久久男人的天堂| 大香蕉久久网| 少妇人妻一区二区三区视频| 久久久精品94久久精品| 国内精品宾馆在线| 国产大屁股一区二区在线视频| 看十八女毛片水多多多| 少妇人妻一区二区三区视频| 听说在线观看完整版免费高清| 亚洲国产精品成人久久小说| 国产综合懂色| 在线a可以看的网站| 又大又黄又爽视频免费| 天美传媒精品一区二区| 少妇人妻精品综合一区二区| 青春草视频在线免费观看| 国产av码专区亚洲av| 99久久精品热视频| 69av精品久久久久久| 成人鲁丝片一二三区免费| 另类亚洲欧美激情| 日韩一本色道免费dvd| 国产男女内射视频| 一本一本综合久久| 老师上课跳d突然被开到最大视频| 伦理电影大哥的女人| 免费观看a级毛片全部| 欧美一区二区亚洲| 日本黄色片子视频| 亚洲精华国产精华液的使用体验| 在线观看av片永久免费下载| 亚洲欧洲日产国产| 男人爽女人下面视频在线观看| 午夜精品一区二区三区免费看| 欧美区成人在线视频| 免费av毛片视频| 国模一区二区三区四区视频| 亚洲av在线观看美女高潮| 三级经典国产精品| 少妇被粗大猛烈的视频| 久久99热这里只频精品6学生| 韩国高清视频一区二区三区| 欧美日本视频| 午夜福利在线在线| 国产视频首页在线观看| 少妇熟女欧美另类| 小蜜桃在线观看免费完整版高清| 亚洲av欧美aⅴ国产| 一边亲一边摸免费视频| 久久久久精品性色| 午夜免费观看性视频| 久久热精品热| av一本久久久久| 最近的中文字幕免费完整| 99热这里只有是精品50| 天堂中文最新版在线下载 | 99热全是精品| 日日摸夜夜添夜夜添av毛片| 国产精品偷伦视频观看了| 搡老乐熟女国产| 少妇熟女欧美另类| 日日摸夜夜添夜夜添av毛片| 亚洲自拍偷在线| 久久久久久久精品精品| 嘟嘟电影网在线观看| www.av在线官网国产| 26uuu在线亚洲综合色| 日日摸夜夜添夜夜爱| 久久久久久久大尺度免费视频| 黄色一级大片看看| 成人国产麻豆网| 天天躁日日操中文字幕| av在线app专区| 国产熟女欧美一区二区| 一级毛片 在线播放| a级毛片免费高清观看在线播放| 大香蕉97超碰在线| 成人亚洲精品av一区二区| 大又大粗又爽又黄少妇毛片口| 国产免费福利视频在线观看| 中文精品一卡2卡3卡4更新| 亚洲精品日本国产第一区| 久久这里有精品视频免费| 亚洲精品亚洲一区二区| 18禁在线播放成人免费| 国产精品久久久久久久久免| 欧美激情久久久久久爽电影| 国产精品av视频在线免费观看| 精品久久久精品久久久| 男女那种视频在线观看| 日韩欧美 国产精品| 亚洲欧洲日产国产| 麻豆精品久久久久久蜜桃| 久久精品国产亚洲av涩爱| 视频区图区小说| 欧美xxxx黑人xx丫x性爽| videos熟女内射| 午夜福利在线观看免费完整高清在| 少妇的逼好多水| 亚洲人与动物交配视频| 97超碰精品成人国产| 国产 一区精品| 天天躁日日操中文字幕| 日本一本二区三区精品| 国产精品伦人一区二区| 人妻制服诱惑在线中文字幕| 草草在线视频免费看| 能在线免费看毛片的网站| tube8黄色片| 91久久精品电影网| 青春草亚洲视频在线观看| 又爽又黄无遮挡网站| 大香蕉久久网| 国产精品国产三级国产av玫瑰| 国产亚洲精品久久久com| 日韩欧美一区视频在线观看 | 赤兔流量卡办理| 国产黄频视频在线观看| 亚洲色图综合在线观看| 精品亚洲乱码少妇综合久久| 国产老妇女一区| 又粗又硬又长又爽又黄的视频| 最近2019中文字幕mv第一页| 特级一级黄色大片| 99热国产这里只有精品6| 午夜老司机福利剧场| 精品视频人人做人人爽| 免费高清在线观看视频在线观看| 午夜免费鲁丝| 亚洲图色成人| 国产女主播在线喷水免费视频网站| 一级毛片电影观看| 嘟嘟电影网在线观看| 成人亚洲欧美一区二区av| 亚洲自偷自拍三级| 99热6这里只有精品| 国产毛片a区久久久久| av免费在线看不卡| 狂野欧美激情性bbbbbb| 特级一级黄色大片| 免费av毛片视频| 国产精品一区www在线观看| 欧美+日韩+精品| 少妇人妻久久综合中文| 简卡轻食公司| av在线亚洲专区| 精品国产三级普通话版| 亚洲四区av| 久久亚洲国产成人精品v| 精品国产乱码久久久久久小说| 精品久久久精品久久久| 久热这里只有精品99| 亚洲天堂av无毛| 亚洲综合精品二区| 久久久欧美国产精品| 日韩精品有码人妻一区| 日本欧美国产在线视频| 91精品一卡2卡3卡4卡| 搡女人真爽免费视频火全软件| 亚洲最大成人手机在线| 欧美老熟妇乱子伦牲交| 国内揄拍国产精品人妻在线| 色婷婷久久久亚洲欧美| 欧美日本视频| 精品一区二区免费观看| 日韩 亚洲 欧美在线| 国产成人freesex在线| 18禁裸乳无遮挡动漫免费视频 | 日本与韩国留学比较| 欧美成人a在线观看| 欧美97在线视频| 日本午夜av视频| 毛片女人毛片| 久久久久久久久久成人| 亚洲综合色惰| 少妇熟女欧美另类| 老司机影院毛片| 欧美国产精品一级二级三级 | 欧美一区二区亚洲| 乱码一卡2卡4卡精品| 男女边摸边吃奶| 少妇猛男粗大的猛烈进出视频 | 丰满乱子伦码专区| 蜜桃久久精品国产亚洲av| 国产女主播在线喷水免费视频网站| 观看美女的网站| 又黄又爽又刺激的免费视频.| av在线观看视频网站免费| 好男人视频免费观看在线| 午夜视频国产福利| 免费播放大片免费观看视频在线观看| 精品一区二区三区视频在线| 免费大片黄手机在线观看| 国产国拍精品亚洲av在线观看| 如何舔出高潮| 纵有疾风起免费观看全集完整版| 国产av国产精品国产| 美女被艹到高潮喷水动态| 最新中文字幕久久久久| 最近2019中文字幕mv第一页| 亚洲av二区三区四区| 国产视频首页在线观看| av黄色大香蕉| 一区二区三区精品91| 亚洲va在线va天堂va国产| 国产精品不卡视频一区二区| 亚洲一级一片aⅴ在线观看| 99热这里只有是精品在线观看| 全区人妻精品视频| 久久人人爽av亚洲精品天堂 | 一个人看视频在线观看www免费| 国产一区亚洲一区在线观看| 美女高潮的动态| 亚洲av日韩在线播放| 国精品久久久久久国模美| 日韩一区二区视频免费看| 草草在线视频免费看| 久久久久性生活片| 精品熟女少妇av免费看| 国产黄片视频在线免费观看| 成年女人在线观看亚洲视频 | 尤物成人国产欧美一区二区三区| 日日啪夜夜撸| 2021少妇久久久久久久久久久| 国产伦在线观看视频一区| 黑人高潮一二区| 国产v大片淫在线免费观看| 国产免费一级a男人的天堂| 国精品久久久久久国模美| 国产伦精品一区二区三区四那| 久久久成人免费电影| 国产69精品久久久久777片| 亚洲电影在线观看av| 国产黄片视频在线免费观看| 国产伦精品一区二区三区四那| 亚洲国产精品专区欧美| 99热这里只有是精品在线观看| 欧美xxⅹ黑人| 卡戴珊不雅视频在线播放| 成人午夜精彩视频在线观看| 国产成人免费观看mmmm| 亚洲成人一二三区av| 1000部很黄的大片| 卡戴珊不雅视频在线播放| 97超碰精品成人国产| 精品午夜福利在线看| 熟妇人妻不卡中文字幕| 国产精品国产三级专区第一集| 亚洲四区av| 看十八女毛片水多多多| 99九九线精品视频在线观看视频| 欧美97在线视频| 久久韩国三级中文字幕| 成人一区二区视频在线观看| 亚洲国产精品专区欧美| 国产精品伦人一区二区| 男女啪啪激烈高潮av片| 日本一二三区视频观看| 亚洲国产日韩一区二区| 久久久精品欧美日韩精品| 夜夜爽夜夜爽视频| 亚洲美女视频黄频| 成人二区视频| 日韩人妻高清精品专区| 日本一二三区视频观看| 网址你懂的国产日韩在线| 超碰av人人做人人爽久久| 国产欧美日韩精品一区二区| av免费在线看不卡| 亚洲人成网站在线播| 自拍偷自拍亚洲精品老妇| 久久久久网色| 高清午夜精品一区二区三区| 少妇猛男粗大的猛烈进出视频 | 久久精品久久精品一区二区三区| 另类亚洲欧美激情| 自拍欧美九色日韩亚洲蝌蚪91 | av卡一久久| 一个人看视频在线观看www免费| 综合色av麻豆| 亚洲丝袜综合中文字幕| 免费不卡的大黄色大毛片视频在线观看| 国产精品蜜桃在线观看| 日日啪夜夜撸| 丝袜脚勾引网站| 熟女电影av网| 久久6这里有精品| 国产毛片在线视频| a级毛片免费高清观看在线播放| 免费少妇av软件| 国产成人免费观看mmmm| 蜜桃久久精品国产亚洲av| 成人黄色视频免费在线看| 国产精品国产三级国产专区5o| 交换朋友夫妻互换小说| 日韩免费高清中文字幕av| 国产精品.久久久| 校园人妻丝袜中文字幕| 午夜免费男女啪啪视频观看| 久久人人爽人人爽人人片va| 国产综合懂色| www.色视频.com| 内地一区二区视频在线| 日本色播在线视频| 天堂俺去俺来也www色官网| 精品人妻偷拍中文字幕| 亚洲最大成人手机在线| 欧美区成人在线视频| 王馨瑶露胸无遮挡在线观看| 色综合色国产| 久久97久久精品| 国产成人freesex在线| 免费大片18禁| 91精品一卡2卡3卡4卡| a级毛片免费高清观看在线播放| 精品视频人人做人人爽| 777米奇影视久久| tube8黄色片| 色吧在线观看| 新久久久久国产一级毛片| 精品久久久久久电影网| 欧美成人午夜免费资源| 99热国产这里只有精品6| 特大巨黑吊av在线直播| 成人黄色视频免费在线看| 综合色丁香网| av在线亚洲专区| 在线观看国产h片| 免费看不卡的av| 欧美少妇被猛烈插入视频| 国产探花在线观看一区二区| 777米奇影视久久| 免费少妇av软件| 在线观看三级黄色| 日本一本二区三区精品| 各种免费的搞黄视频| 在线观看人妻少妇| 最近最新中文字幕免费大全7| av线在线观看网站| 九九久久精品国产亚洲av麻豆| 九九爱精品视频在线观看| 日韩中字成人| 香蕉精品网在线| 欧美日韩精品成人综合77777| 国产v大片淫在线免费观看| av免费观看日本| 91精品国产九色| 精华霜和精华液先用哪个| 欧美成人精品欧美一级黄| 精华霜和精华液先用哪个| 欧美 日韩 精品 国产| 黄片wwwwww| 午夜精品国产一区二区电影 | 春色校园在线视频观看| av.在线天堂| 亚洲最大成人手机在线| av在线蜜桃| 国产亚洲精品久久久com| 亚洲欧美精品自产自拍| 五月天丁香电影| 成人毛片60女人毛片免费| 成年人午夜在线观看视频| 国产女主播在线喷水免费视频网站| 亚洲国产色片| 久久热精品热| 韩国av在线不卡| 日本与韩国留学比较| 男人狂女人下面高潮的视频| 久久国产乱子免费精品| 亚洲av不卡在线观看| 黄片wwwwww| 亚洲图色成人| 色吧在线观看| 亚洲在线观看片| 精品久久久久久久久av| 另类亚洲欧美激情| 亚洲av电影在线观看一区二区三区 | 欧美老熟妇乱子伦牲交| 一级毛片久久久久久久久女| 国产成人freesex在线| 丝袜美腿在线中文| av天堂中文字幕网| 亚洲欧美清纯卡通| 成年女人看的毛片在线观看| 中文乱码字字幕精品一区二区三区| 男的添女的下面高潮视频| 在线免费观看不下载黄p国产| 亚洲真实伦在线观看| 一级av片app| 日韩精品有码人妻一区| 新久久久久国产一级毛片| 亚洲自偷自拍三级| 99久久精品一区二区三区| 国产高清国产精品国产三级 | 免费观看在线日韩| 国产精品熟女久久久久浪| 交换朋友夫妻互换小说| 99热全是精品| 日本色播在线视频| 内地一区二区视频在线| 99久久中文字幕三级久久日本| 欧美一级a爱片免费观看看| 国产黄a三级三级三级人| 亚洲欧美成人精品一区二区| 嫩草影院新地址| 成年版毛片免费区| 能在线免费看毛片的网站| 在线看a的网站| 99热6这里只有精品| 亚洲天堂av无毛| 国产在线一区二区三区精| av又黄又爽大尺度在线免费看| 男人添女人高潮全过程视频| 国产视频首页在线观看| 伊人久久精品亚洲午夜| 国产免费一区二区三区四区乱码| 天堂中文最新版在线下载 | 日韩国内少妇激情av| 六月丁香七月| 国产免费福利视频在线观看| 久久久久国产精品人妻一区二区| 亚洲真实伦在线观看| 亚洲精品国产av蜜桃| 久久久久久久久大av| 国产探花在线观看一区二区| 嫩草影院精品99| 大香蕉久久网| 禁无遮挡网站| 亚洲精品自拍成人| 国产v大片淫在线免费观看| 少妇人妻久久综合中文| 亚洲伊人久久精品综合| 国产亚洲av片在线观看秒播厂| 麻豆国产97在线/欧美| 人人妻人人爽人人添夜夜欢视频 | 亚洲色图av天堂| 国产亚洲av嫩草精品影院| 22中文网久久字幕| 男人狂女人下面高潮的视频| 亚洲伊人久久精品综合| 欧美潮喷喷水| 视频中文字幕在线观看| av天堂中文字幕网| 下体分泌物呈黄色| 国产一区二区三区av在线| 97在线人人人人妻| 精品久久国产蜜桃| 国产探花极品一区二区| 亚洲婷婷狠狠爱综合网| 纵有疾风起免费观看全集完整版| 韩国高清视频一区二区三区| 亚洲怡红院男人天堂| 精品视频人人做人人爽| 天美传媒精品一区二区| 最近中文字幕2019免费版| 国产黄色视频一区二区在线观看| 免费观看a级毛片全部| 老女人水多毛片| 久久ye,这里只有精品| 国产白丝娇喘喷水9色精品| 中文乱码字字幕精品一区二区三区| 国产精品偷伦视频观看了| 亚洲av在线观看美女高潮| 2018国产大陆天天弄谢| 亚洲性久久影院| 日韩欧美精品免费久久| 午夜福利高清视频| 国产精品久久久久久久久免| 一本色道久久久久久精品综合| 午夜视频国产福利| 91久久精品国产一区二区成人|