• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Unsupervised Martian Dust Storm Removal via Disentangled Representation Learning

    2022-10-25 08:24:30DongZhaoJiaLiHongyuLiandLongXu

    Dong Zhao, Jia Li, Hongyu Li, and Long Xu

    1 State Key Laboratory of Virtual Reality Technology and Systems, School of Computer Science and Engineering, Beihang University, Beijing 100191, China jiali@buaa.edu.cn

    2 State Key Laboratory of Space Weather, National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China

    3 Peng Cheng Laboratory, Shenzhen 518000, China

    Abstract Mars exploration has become a hot spot in recent years and is still advancing rapidly.However,Mars has massive dust storms that may cover many areas of the planet and last for weeks or even months.The local/global dust storms are so influential that they can significantly reduce visibility, and thereby the images captured by the cameras on the Mars rover are degraded severely.This work presents an unsupervised Martian dust storm removal network via disentangled representation learning (DRL). The core idea of the DRL framework is to use the content encoder and dust storm encoder to disentangle the degraded images into content features (on domain-invariant space) and dust storm features(on domain-specific space). The dust storm features carry the full dust storm-relevant prior knowledge from the dust storm images.The“cleaned”content features can be effectively decoded to generate more natural,faithful,clear images.The primary advantages of this framework are twofold.First,it is among the first to perform unsupervised training in Martian dust storm removal with a single image, avoiding the synthetic data requirements. Second, the model can implicitly learn the dust storm-relevant prior knowledge from the real-world dust storm data sets,avoiding the design of the complicated handcrafted priors. Extensive experiments demonstrate the DRL framework’s effectiveness and show the promising performance of our network for Martian dust storm removal.

    Key words: planets and satellites: atmospheres – planets and satellites: general – planets and satellites: surfaces

    1. Introduction

    The Zhurong rover, China’s first Mars rover, went into hibernation mode on 2022 May 18, to cope with the reducing solar power generation capacity caused by a dust storm which was observed by the medium resolution images obtained from the orbiter of the Tianwen-1 probe on 2022 March 16 and April 30.That is one case of the Martian dust storms posing a threat to the scientific instruments (Alexandra Witze 2018). Unfortunately, Martian dust storms are common, which may last for weeks and cover many areas of the planet(Clarke 2018).These dust storms are so influential that they can change the climate of Mars (Leovy 2001; Liu et al. 2022) and degrade signal/image qualities taken by cameras on the Mars rover. For example, as shown in Figure 1, the images captured under the dust storm weather suffer from contrast degradation, color attenuation, and poor visibility.Therefore,removing the dust storm from a single image becomes crucial for studying the Martian atmosphere(Banfield et al.2020)and Martian geology(Chaffin et al.2021).

    To this end,we have recalled many related studies on image enhancement and restoration algorithms. Among them, we focus on the haze removal task because it is very similar to dust storm removal, where the main difference between them is the suspended particles in the atmosphere.These image restoration approaches are categorized into five groups as follows.

    Conventional image enhancement techniques, such as Auto Levels (AL), Retinex model (Jobson et al. 1997) or contrast limited adaptive histogram equalization(CLAHE)(Reza 2004)can improve the visual quality of the Martian dust storm images. These methods have a lower computational cost, yet some of their common drawbacks are that some parameters should be manually configured. Thus, they cannot automatically adapt to different images captured under different sizes of dust storms.

    Physical prior-based method is another kind of approach that might work.For example,inspired by the success of the Earth’s atmospheric scattering model(Narasimhan 2000)that has been leveraged in natural image dehazing by introducing specific priors (e.g., dark channel prior (DCP) (He et al. 2011), colorlines prior (Fattal 2014), color attenuation prior (Zhu et al.2015),and non-local prior(NLP)(Berman et al.2018),Li et al.(2018) directly utilized the DCP for Martian dust storm removal. However, the sizes, types, and concentrations in the space of the aerosol particles on Mars are very different from the ones on Earth;therefore,the atmospheric scattering models of the two planets are different (Egan & Foreman 1971). For example, typically, Rayleigh scattering occurs on Earth, while Mie scattering on Mars(Collienne et al.2013).That means the existing priors proposed, especially for nature image dehazing on Earth, are unable to fit all conditions on Mars and may produce unwanted artifacts because the priors are invalid.

    Figure 1. The motivation of the DRL for unsupervised Martian dust storm removal.

    In contrast, supervised learning-based approaches can learn the latent prior knowledge from masses of paired contaminated images and their clear counterparts and produce visually appealing results. For example, in the multi-scale CNNs(MSCNN) (Ren et al. 2020) and enhanced pix2pix dehazing networks (EPDN) (Qu et al. 2019), the hazy samples are often synthesized by applying the atmospheric scattering model.Despite their effectiveness on the open-access data sets,supervised learning-based methods have suffered from the domain shift issues of the synthetic data sets.Training on such data sets,these methods are probably to overfit.Therefore,they are less able to generalize well to real-world images.

    To improve the performances in the real-world conditions,some semi-supervised learning-based models, such as semisupervised dehazing network (SSDN) (Li et al. 2019) and domain adaptation dehazing network (DADN) (Shao et al.2020),have been exploited to use both synthetic and real-world data sets. However, as to our task, such paired dust storm images and clear images are impractical to collect under realworld conditions on Mars.

    Unlike the above learning-based models, unsupervised learning-based models are solely trained on the real-world data set, avoiding labor-intensive collecting data and dealing with the domain shift problem. However, due to the lack of prior knowledge, the existing unsupervised models inevitably exploit the properties of clear images via predetermined prior losses. For example, the performances of deep dark channel prior(DDCP)(Golts et al.2019)and zero-shot image dehazing(ZID) (Li et al. 2020) are very dependent on the dark channel prior (He et al. 2011) loss. Like the physical prior-based methods, using the handcrafted priors as the objections may lead to the networks being less robust to the conditions that the priors are invalid.

    By discussing the approaches above, we found that the unsupervised learning-based method is more fit for the Martian dust storm removal task for two reasons. First, this method does not need paired clear and dust storm images. Second, it can implicitly learn the dust storm-relevant prior knowledge from real-world dust storm data sets. To this end, we develop an unsupervised Martian dust storm removal framework that learns the dust storm-relevant prior knowledge via disentangled representation learning(DRL).The motivation of the DRL can be succinctly illustrated in Figure 1, where the “disentanglement” module aims to encode the input dust storm image into intermediate representations, i.e., content features (domaininvariant cues) and dust storm features (domain-specific cues).Because the dust storm features are “taken away” from the original image,the remained content features can be effectively decoded into a clear image. The dust storm features carry the full dust storm-relevant prior knowledge of the original image,and they can be fused to another clear image’s content features to generate a new dust storm image. Concretely, the proposed framework consists of two content encoders to extract content features from unpaired clear and dust storm images, one dust storm encoder to disentangle dust storm features,two generators to reconstruct clear and dust storm images, and two discriminators to align the generated images into their corresponding domains.We use the adversarial loss during the training to align the content features to the ones of clear image. In addition, the cross-cycle consistency loss is used to guarantee the content consistency between the original and the reconstructed images.Finally, we also use a latent reconstruction loss to encourage bidirectional mapping of the intermediate representations.Extensive experiments show the promising performance of our Martian dust storm removal framework using DRL.

    Figure 2.The latent space assumption in the DRL.The clear and dust storm images belong in two different domainsC andD,respectively.They can be mapped to content features cc/cd in domain-invariant space, while dust storm features d in domain-specific space can be disentangled from domainD.

    This paper is structured as follows. In Section 2, the details of our unsupervised Martian dust storm removal network using disentangled representation learning are introduced. We then introduce our data sets in Section 3. We further report the experimental results and the ablation studies in Section 4. In Section 5, we summarize our approach.

    2. Method

    The DRL can be used to model one or more factors of domain variations while the other factors remain relatively invariant. Specifically, in the Martian dust storm removal task,there are two domains: the clear image domainC and dust storm image domainD.The goal of the DRL for unsupervised dust storm removal is to learn the domain-specific factors (i.e.,dust storm features d)from domainD and the domain-invariant factors (content features ccand cd)from domainC and domain D, as depicted in Figure 2. To do so, the DRL model in our work consists of one dust storm encoder Edsto extract the dust storm features; two content encodersEcctandEdctto extract the content features ccand cdfrom clear and dust storm images,respectively; and two image generators Gcand Gdto map the inputs onto clear and dust storm image domains, respectively.Details on the network architectures and training objectives are introduced in the following subsections.

    2.1. Network Architecture

    2.1.1. Overview

    The overall disentangled representation learning-based framework for the Martian dust storm removal consists of three parts:forward translations,backward reconstructions,and self-reconstructions,as shown in Figure 3.The notations in this figure are summarized in Table 1.

    Forward Translations:The forward translation contains two branches:one for the forward dust storm image translation and the other for the forward clear image translation.

    Figure 3. Overview of the unsupervised Martian dust storm removal network using disentangled representation learning.

    For backward dust storm image reconstruction, we have:

    For backward clear image reconstruction, we have:

    Self-Reconstructions: To facilitate the training process, we apply self-reconstructions as illustrated in Figure 3(b). With encoded content and dust storm features,the generators Gcand Gdshould decode them back to original inputs Icand Id,respectively. They can be formulated by:

    Table 1 A Summary of the Notations used in Our Unsupervised Martian Dust Storm Removal Framework

    2.1.2. Details

    We adapt the above architecture for the unsupervised framework with the following configurations.

    Identity Shortcut Connection and Long Skip Connection:

    The details of each layer in the encoder and generator are illustrated in Figure 4. Each basic layer in the encoders/generators consists of a convolutional/deconvolution for down/up-sampling followed by a ResNet (He et al. 2016)block to avoid the notorious exploding or vanishing gradients that may occur during the training. One ResNet block is realized by inserting shortcut connections into conventional convolutional layers, as shown in Figure 4. Note that the shortcut connections introduce neither extra parameter nor computation complexity.

    We also equip long skip connections between encoders and generators, referred to as U-shape networks (U-Net) (Ronneberger et al. 2015), as shown in Figure 4. The long skip connections allow the network to propagate structure (at lowlevel) and semantic (at high-level) information of inputs to deeper layers, preserving the spatial information lost during downsampling in the encoders.

    Sparse Switchable Normalization and Adaptive Instance Normalization: We use the sparse switchable normalization(SSN)(Shao et al.2019)in the basic layer instead of the batch normalization (BN) (Ioffe & Szegedy 2015) and instance normalization (IN) (Ulyanov et al. 2017) that are widely used in natural image generation networks. Our earlier explorations found that using BN and IN alone would lead to“spot”artifacts in the generated images, as illustrated in Figure 8. One reason might stem from challenges inherent in the Martian dust storm removal networks where different layers should behave differently. SSN can address this issue by adaptively selecting the optimal normalizers among BN, IN, and layer normalization (LN) (Ba et al. 2016) at different convolutional layers.

    In our model, the dust storm image generator Gdfuses the content features and dust storm features via adaptive instance normalization (AdaIN) (Huang & Belongie 2017). The clear content features ccand dust storm features d are fed into the AdaIN layer. The AdaIN layer aims to align the content features’mean and variance to the ones of dust storm features:

    where μ(·) and δ(·) are the mean and standard deviation operations.

    2.2. Training Objective

    The overall loss function of the proposed network is:

    where Ladv, Lccc, Lsrcand Llrcare the adversarial loss, crosscycle consistency loss, self-reconstruction loss and latent reconstruction loss, respectively. λadv, λccc, λsrcand λlrcare the corresponding hyper-parameters to control the importance of each term.

    Adversarial Loss: Here, we apply the adversarial loss(Goodfellow et al. 2014) between the generated images and corresponding target domains to force them to have as similar data distributions as possible.For the dust storm image domain,we define the adversarial loss as:

    where Ddis the discriminator for the dust storm image domain.E represents the mean operation of training samples in a batch.Our discriminator module is “PatchGAN” (Isola et al. 2017)classifiers.

    Figure 4. Details of the encoders (Ecct, Eds, and Edct) and generators (Gd and Gc) (taking the forward translation as an example). ?: elemental-wise subtraction; ?:elemental-wise division; ?: elemental-wise multiplication; ⊕: elemental-wise addition.

    Figure 5. Examples of the available samples in the data sets.

    Similarly,for the clear image domain,the adversarial loss is defined as:

    where Dcis the discriminator for clear image domain.

    Cross-Cycle Consistency Loss: We use the cross-cycle consistency loss to guarantee that the clear images Icfwdcan be reconstructed to the dust storm domain, and the dust storm image Idfwdcan be reconstructed back to the clear domain.Specifically, the cross-cycle consistency loss is defined as:

    Latent Reconstruction Loss: To encourage the invariant representation learning between the clear and the dust storm space, we apply a latent reconstruction (including content and dust storm features reconstruction) loss similar to Lin et al.(2018) and Zhu et al. (2017) as follows:

    whereωcc,ωcdand ωdare weights of the features for corresponding reconstructions, respectively. Note that the content features are more easily learned than the dust storm features.It is because the formers can be aligned by the long skip connections,delivering multi-scales of content information from encoders to the corresponding generators. In contrast, the dust storm features are only learned at the single high-level layer.To address this discrepancy, the ωdis set larger thanωccandωcd.

    3. Dataset

    The Mars32k Dataset(Dominik Schmidt 2018)collects 32,368 560×500 samples captured by the Curiosity rover between 2012 and 2018 on Mars. All samples are provided by NASA/JPLCaltech.This data set involves a variety of geographic features of Mars, including rocky terrain, sand dunes, and mountains.However, we found that the raw Mars32k data set cannot be directly served as the training data for our task accounts for two aspects. First, the raw data set is not curated, containing many anomalous samples. Second, it is difficult for some scenes to determine whether they have a dust storm or not. To build the available data sets, we first artificially eliminate the anomalous samples and select the available samples that are confirmed to have dust storms or are clear.After going through the above procedures,we obtain 1925 dust storm samples and 1432 clear samples, as shown in Figure 5. Then, we randomly selected 65 dust storm samples as the testing data sets and remained samples were used for training data augmentation. We augment the samples by flipping, randomly cropping, and randomly rotating. Finally, the data set contains 13,016 dust storms and 8985 clear unpaired samples for model training.Considering the limits of memory and fast training, we resize the input samples to 256×256. While during the testing, the inputs are resized to 512×512.

    4. Experiments

    4.1. Metrics

    Because it is impractical to acquire Martian dust storm images and their clear counterparts,we use the following three no-reference image quality metrics(NRIQM)in our evaluation experiments.

    Geometric Mean of the Ratios of Visibility Level (GMRVL)

    (Hautiere et al. 2008).The GMRVL is a function of the ratio between the visibility levels of restored and original images. It is used to measure the quality of the contrast restoration.

    Blind/Referenceless Image Spatial Quality Evaluator(BRISQUE) (Mittal et al. 2012).The BRISQUE is a natural scene statistic-based NRIQM calculated in the spatial domain. It can quantify the quality of distortion and naturalness using locally normalized luminance. We use this metric to evaluate whether the dust storm images have been satisfactorily restored with good visual perception.

    Perception-based Image Quality Evaluator (PIQE) (Venkatanath et al. 2015).The PIQE extracts local features for measuring patch-wise restoration quality. It is an opinionunaware methodology and thus does not need any training data.It can measure whether the images are satisfactorily restored:insufficient restored or over-restored results would lead to a higher PIQE score.

    4.2. Implemental Details

    We develop our model on PyTorch deep learning package.We adopt Adam (Kingma & Ba 2015) as the optimization algorithm during the training, with a batch size of 8. In the experiments, our model is trained for 20 epochs, taking about 36 hr on two NVIDIA GeForce GTX 3090 Ti GPU.We set the initial learning rate is 0.0002 for all of the encoders,generators and discriminators, and the learning rate decay is γ = 0.5 for every 10 epochs. The parameters of the hybrid loss functions are set as: (λadv, λccc, λsrc, λlrc) = (0.2, 3, 1, 0.5). Weights of the dust storm features are set as: (ωcc,ωcd, ωd) = (1, 1, 10).The model will be available as open source at https://github.com/phoenixtreesky7/DRL_UMDSR.

    4.3. Comparisons with Other Image Enhancement/Dehazing Methods

    We compare our model with the following methods: (1)Conventional Image Enhancement methods (AL and CLAHE Reza 2004); (2) Physical Prior-based methods (DCP He et al.2011 and NLP Berman et al. 2018); (3) Supervised Learningbased method (MSCNN Ren et al. 2020 and EPDN Qu et al.2019);(4)Semi-supervised Learning-based methods(SSDN Li et al. 2019 and DADN Shao et al. 2020); (5) Unsupervised Learning-based methods (ZID Li et al. 2020). Qualitative and quantitative results are illustrated in Figure 6 and Table 2,respectively.

    Qualitative Comparisons:All algorithms can be observed to restore the dust storm images with good visual perception to some extent. However, the visual results of AL and DADN(Shao et al. 2020) contribute false color, more like the photos taken on Earth. The CLAHE (Reza 2004) and SSDN (Li et al.2019) seem to fail to remove the dust storm effectively,producing poor visual results in distant scenes. Although the physical prior-based methods DCP (He et al. 2011) and NLP(Berman et al. 2018) can effectively remove dust storms from the scenes, they tend to yield over-enhanced visual artifacts,especially in the sky regions. These results also demonstrate that these priors are invalid for Martian dust storm images.The results of the MSCNN (Ren et al. 2020) and EPDN (Qu et al.2019) maintain plausible visual details, achieving better performance than the non-learning methods. However, some visual results are not satisfactory because some scenes are oversaturation,as illustrated in the first,the fifth,and the last images in Figure 6 (g). The restored images of ZID (Li et al. 2020)have severe color distortion in the sky regions.By contrast,our model achieves remarkable performance. Note that our visual results may not be as colorful as the ones of AL,DCP(He et al.2011), MSCNN (Ren et al. 2020), and DADN (Shao et al.2020). However, our DRL framework extracts the dust storm factors from the dust storm images, and the remaining content features are aligned to the ones of real-world Martian clear images. It enables the restored image to display consistent and faithful colors. Moreover, our model generates high-quality restored results with fewer artifacts than all the compared methods,as illustrated in Figure 7 where the red blocks are the noticeable artifacts masks detected by PIQE(Venkatanath et al.2015) metric.

    Quantitative Comparisons: We further use the GMRVL(Hautiere et al.2008),BRISQUE(Mittal et al.2012)and PIQE(Venkatanath et al. 2015) to quantitatively evaluate the effectiveness of our model. The results are listed in Table 2.It can be found that the conventional image enhancement models can generate high contrast results with higher GMRVL than learning-based methods. However, the other two metrics of them are worse, indicating that the results are unnatural and also demonstrating that these methods are unable to cope with complex dust storm conditions. The physical prior-based methods can generally obtain better GMRVL scores than other methods. However, they are more likely to lead to unnatural results with more artifacts as they get high BRISQUE and PIQE. The reasons are that these priors are invalid on Mars.The supervised and semi-supervised methods successfully refrain from over-saturation, resulting in smaller BRISQUE and PIQE than the non-learning methods. However, despite their superiorities, their BRISQUE and PIQW scores are still higher (worse) than ours. Thanks to the DRL, our model achieved better visual quality and obtained the lowest scores of the BRISQUE and PIQE. As Table 2 demonstrates, our model surpasses the second-best methods with gains of 1.123 on BRISQUE and 5.517 on PIQE.Note that our GMRVL is lower than some of the compared methods. The reason is that these methods tend to outcome over-enhanced results, which may increase the GMRVL to a certain extent.

    Figure 7.The noticeable artifacts masks detected by PIQE(Venkatanath et al.2015)metric for(b)AL,(c)CLAHE(Reza 2004),(d)DCP(He et al.2011),(e)NLP(Berman et al.2018),(f)MSCNN(Ren et al.2020),(g)EPDN(Qu et al.2019),(h)SSDN(Li et al.2019),(i)DADN(Shao et al.2020),(j)ZID(Li et al.2020)and(k) ours results. (a) The dust storm image.

    Table 2 Quantitative Evaluations of Different Methods using GMRVL (Hautiere et al. 2008), BRISQUE (Mittal et al. 2012) and PIQE (Venkatanath et al. 2015)

    4.4. Ablation Studies

    To evaluate the effectiveness of each configuration in our model, we compare the following models:(1)“BN”:using the batch normalization in the entire networks; (2) “IN”: using the instance normalization in the entire networks;(3)“LLS”:using only one long skip connection that connects low-level stage,i.e., the encoder block behind the feature extraction and the generator block before the final output one; (4) “CNN”: using the vanilla convolutional layers, i.e., removing the identity shortcut connections in each ResNet blocks; (5) CycleGAN:using CycleGAN framework, i.e., excluding the dust storm encoder Eds.All of the above models are trained for ten epochs for fast evaluations.

    As shown in Table 3,both of the results of models“BN”and“IN”reduce the performances of dust storm removal compared with our model with SSN.Furthermore,it verifies that different layers of the networks behave differently; thus, the normalization should also be changed at each layer. Moreover, they would generate “spot” artifacts as labeled by the red rectangle illustrated in Figures 8(b) and (c). The model U-Net uses skip connections between the encoders and generators at each stage.Compared with the “LLS” where skip connections are only used at the low-level stage, U-Net shaped network increases GMRVL by 0.157, and it significantly reduces BRISQUE and PIQE by 1.179 and 8.226, respectively. These are because the skip-connections at different levels preserve the generators’structure(low-level features)and semantic(high-level features)information. The model “CNN” replaces the ResNet blocks with two stacked conventional convolutional layers. However,it reduces by 0.083 on GMRVL and increases by 0.769 on BRISQUE and 0.403 on PIQE. We further compare the DRL with the CycleGAN, another widely used framework for unsupervised learning. As illustrated in Table 3, the DRL framework contributes 0.453 improvements on GMRVL,8.491 improvements on BRISQUE, and 10.509 improvements on PIQE,compared with the CycleGAN.These evaluations verify the effectiveness of using the DRL framework.

    Figure 8.The visual results of the models in our ablation studies. As we can find that artifacts are generated in the results of the models of “BN”,“IN”,“LLS”and“CycleGAN”, demonstrating the inefficiency of these models.

    Table 3 Ablation Study: Evaluations on the SSN, U-Net Shape, ResNet Block and DRL (at 10th epoch)

    5. Conclusion and Discussion

    In this work,we propose an unsupervised Martian dust storm removal network via DRL. The network is formed with three parts,i.e.,the forward translation,the backward reconstruction,and self-reconstruction, to achieve powerful representation learning. Additionally, we enforce the adversarial loss, cycleconsistency loss, and latent reconstruction loss to train the model.

    Advantages.There are several significant advantages of our model. First, our dust storm removal model with DRL can be solely trained on the real-world unpaired dust storm and clear images. Thus, we can circumvent the expansive and timeconsuming data set collection and address the domain shift issue. Second, the model can implicitly learn the dust stormrelevant prior knowledge from the dust storm data sets,generating a high-quality image without relying on any handcrafted priors as training objects. Extensive experiments demonstrate that our model exhibits good generalization performances on real-world Martian dust storm removal.

    Properties.It is worthy to note that our work is an image restoration model, not image enhancement. On the technical side, our model translates the Martian dust storm images into clear images.During this translation,the discriminator acts as a style classifier that can align the restored images and the clear images onto the same domain. In our data sets, the clear samples are collected from real, clear Martian images; as a result,our restored images are more looked like to be taken on a clear day on Mars.Because of this,unlike the goals of image enhancement algorithms that yield visually pleasant results with vivid color and high contrast, our goal is to restore the Martian dust storm image to a real, natural, and faithful one.

    Beyond Martian dust storm removal.With the rapid development of space technology,numerous optical telescopes have been built (in outer space and on the ground), and massive amounts of astronomical image data have been captured. It enables us to develop data-driven-based approaches for astronomical image restoration.Unfortunately, it is impractical to simultaneously obtain the paired clear and degraded astronomical images. As discussed above, the advantages and properties of the proposed DRL network enable effective unsupervised learning, having huge potential for many other data-driven-based astronomy image restoration, where the paired unclear and clear images are unavailable.

    Evaluation of the dust storm.We further propose a new approach for the Martian dust storm evaluations with respect to the level and spatial distribution in the scenes.Specifically,we compute the difference between the recovered image and the dust storm image by the following function:

    Figure 9. Evaluation of the dust storm. ΔI is the map of the difference between the recovered and dust storm images. As we can see, it can describe the spatial distribution of the dust.ΔI is the global average of ΔI. Larger ΔI indicates more severe dust storms.

    Acknowledgments

    This work was supported by the National Key R&D Program of China (No. 2021YFA1600504) and the National Natural Science Foundation of China (NSFC) (Nos. 11790305,62132002 and 61922006).

    中出人妻视频一区二区| 亚洲精品中文字幕在线视频| 自拍欧美九色日韩亚洲蝌蚪91| 人人妻人人澡人人爽人人夜夜| 色在线成人网| 久久国产亚洲av麻豆专区| 在线播放国产精品三级| 精品免费久久久久久久清纯 | 嫁个100分男人电影在线观看| videos熟女内射| 黄色怎么调成土黄色| 午夜精品久久久久久毛片777| 国产色视频综合| 国产精品久久久久久人妻精品电影| 极品教师在线免费播放| 国产一区二区三区综合在线观看| 精品人妻在线不人妻| 日韩 欧美 亚洲 中文字幕| 乱人伦中国视频| 天天躁夜夜躁狠狠躁躁| 老熟妇乱子伦视频在线观看| 咕卡用的链子| 国产免费男女视频| 午夜91福利影院| 国产免费男女视频| 9191精品国产免费久久| 啦啦啦视频在线资源免费观看| 在线观看舔阴道视频| 精品国内亚洲2022精品成人 | 黄色怎么调成土黄色| 久久人人爽av亚洲精品天堂| 午夜久久久在线观看| 久久久国产成人精品二区 | 黄片大片在线免费观看| av超薄肉色丝袜交足视频| 青草久久国产| 别揉我奶头~嗯~啊~动态视频| 18禁黄网站禁片午夜丰满| 一进一出抽搐动态| 精品熟女少妇八av免费久了| 99热国产这里只有精品6| 波多野结衣一区麻豆| 一边摸一边抽搐一进一小说 | 欧美黄色片欧美黄色片| 亚洲 国产 在线| 岛国在线观看网站| 91老司机精品| 少妇的丰满在线观看| 久久午夜亚洲精品久久| 亚洲欧美一区二区三区黑人| 国产色视频综合| 免费看十八禁软件| 亚洲欧美激情综合另类| 一进一出抽搐动态| 免费不卡黄色视频| 国产在线精品亚洲第一网站| 久久精品国产综合久久久| 精品国产亚洲在线| av线在线观看网站| ponron亚洲| 欧美+亚洲+日韩+国产| 久久精品熟女亚洲av麻豆精品| 亚洲 国产 在线| 99热只有精品国产| 欧美激情极品国产一区二区三区| 超色免费av| 国产激情久久老熟女| 国产99久久九九免费精品| 精品高清国产在线一区| 国内久久婷婷六月综合欲色啪| 国产亚洲欧美精品永久| 国产亚洲欧美98| 欧美人与性动交α欧美精品济南到| 一级,二级,三级黄色视频| 成人手机av| 免费人成视频x8x8入口观看| 亚洲aⅴ乱码一区二区在线播放 | 国产99久久九九免费精品| 中文字幕最新亚洲高清| 亚洲精品国产色婷婷电影| 搡老乐熟女国产| 黑人欧美特级aaaaaa片| 久久久久国产一级毛片高清牌| 亚洲精品国产一区二区精华液| 国产精品亚洲av一区麻豆| 国产成人av教育| 91国产中文字幕| 久久香蕉激情| 成人黄色视频免费在线看| tocl精华| 久久天躁狠狠躁夜夜2o2o| cao死你这个sao货| 亚洲av欧美aⅴ国产| 免费人成视频x8x8入口观看| 中亚洲国语对白在线视频| 久久天躁狠狠躁夜夜2o2o| 亚洲五月婷婷丁香| 大香蕉久久成人网| 成人黄色视频免费在线看| 99热网站在线观看| 狠狠婷婷综合久久久久久88av| 国产不卡av网站在线观看| 久久精品熟女亚洲av麻豆精品| 亚洲va日本ⅴa欧美va伊人久久| 中文字幕av电影在线播放| 国产日韩一区二区三区精品不卡| 久久午夜亚洲精品久久| 国产精品久久久久久人妻精品电影| 日韩熟女老妇一区二区性免费视频| 19禁男女啪啪无遮挡网站| 欧美日韩中文字幕国产精品一区二区三区 | 国产精品 欧美亚洲| av一本久久久久| 50天的宝宝边吃奶边哭怎么回事| 亚洲 欧美一区二区三区| 色综合婷婷激情| 老司机亚洲免费影院| 国产精品98久久久久久宅男小说| 国产成人欧美| 国产欧美日韩一区二区三区在线| 亚洲专区中文字幕在线| 每晚都被弄得嗷嗷叫到高潮| 国产欧美日韩一区二区三区在线| 999久久久精品免费观看国产| 黑人巨大精品欧美一区二区mp4| 757午夜福利合集在线观看| 中文字幕高清在线视频| 人人妻人人添人人爽欧美一区卜| 色尼玛亚洲综合影院| 狂野欧美激情性xxxx| 久久这里只有精品19| 欧美精品一区二区免费开放| 亚洲精品自拍成人| 亚洲黑人精品在线| 精品免费久久久久久久清纯 | 久久青草综合色| 国产成人啪精品午夜网站| 成熟少妇高潮喷水视频| 国产男女内射视频| 丰满人妻熟妇乱又伦精品不卡| 法律面前人人平等表现在哪些方面| 日韩欧美在线二视频 | 国产淫语在线视频| 亚洲第一av免费看| 精品亚洲成a人片在线观看| 国产淫语在线视频| 一区二区日韩欧美中文字幕| 少妇被粗大的猛进出69影院| 美女高潮喷水抽搐中文字幕| 国产精品自产拍在线观看55亚洲 | 精品少妇久久久久久888优播| 最近最新中文字幕大全电影3 | av网站免费在线观看视频| 1024视频免费在线观看| 高清毛片免费观看视频网站 | 日本欧美视频一区| 精品乱码久久久久久99久播| 女人久久www免费人成看片| 99精国产麻豆久久婷婷| 啪啪无遮挡十八禁网站| 色精品久久人妻99蜜桃| 久久性视频一级片| 国产成人欧美在线观看 | 又黄又爽又免费观看的视频| 人妻丰满熟妇av一区二区三区 | 国产精品综合久久久久久久免费 | 亚洲av熟女| 免费不卡黄色视频| 9色porny在线观看| 欧美日韩瑟瑟在线播放| 国产成人免费无遮挡视频| 法律面前人人平等表现在哪些方面| 看免费av毛片| 国产单亲对白刺激| 熟女少妇亚洲综合色aaa.| 久久久精品国产亚洲av高清涩受| 久久精品国产99精品国产亚洲性色 | a在线观看视频网站| 美女 人体艺术 gogo| 欧美黄色淫秽网站| 高清在线国产一区| 免费一级毛片在线播放高清视频 | 欧美日韩亚洲综合一区二区三区_| 亚洲精华国产精华精| 极品人妻少妇av视频| av片东京热男人的天堂| 波多野结衣一区麻豆| 性少妇av在线| 91麻豆av在线| 欧美大码av| 人妻丰满熟妇av一区二区三区 | 一级黄色大片毛片| 一区二区三区激情视频| 精品一区二区三区视频在线观看免费 | 91av网站免费观看| 欧美成狂野欧美在线观看| 精品久久久久久电影网| 色综合欧美亚洲国产小说| 久久影院123| 黄色毛片三级朝国网站| 又黄又爽又免费观看的视频| 丰满饥渴人妻一区二区三| 欧美亚洲日本最大视频资源| 婷婷精品国产亚洲av在线 | 大码成人一级视频| 咕卡用的链子| 飞空精品影院首页| 国产不卡一卡二| 亚洲精品粉嫩美女一区| 欧美精品人与动牲交sv欧美| 18禁黄网站禁片午夜丰满| 亚洲精品国产色婷婷电影| 欧美乱码精品一区二区三区| 午夜精品国产一区二区电影| 一区二区日韩欧美中文字幕| 99re6热这里在线精品视频| 亚洲人成77777在线视频| 51午夜福利影视在线观看| 嫁个100分男人电影在线观看| 老司机午夜十八禁免费视频| 在线av久久热| 免费av中文字幕在线| 黄色a级毛片大全视频| 男男h啪啪无遮挡| 免费在线观看视频国产中文字幕亚洲| 亚洲专区国产一区二区| 老熟妇仑乱视频hdxx| 狂野欧美激情性xxxx| 国产精品一区二区在线不卡| 亚洲欧美一区二区三区久久| 天天躁夜夜躁狠狠躁躁| 精品乱码久久久久久99久播| 亚洲精品国产精品久久久不卡| 免费人成视频x8x8入口观看| 别揉我奶头~嗯~啊~动态视频| 一级片'在线观看视频| 99精品在免费线老司机午夜| 午夜激情av网站| 老司机靠b影院| 国产一区在线观看成人免费| 天堂动漫精品| 高清黄色对白视频在线免费看| 三级毛片av免费| 日本一区二区免费在线视频| 国产单亲对白刺激| 欧美中文综合在线视频| 美女 人体艺术 gogo| 多毛熟女@视频| 91在线观看av| 亚洲成人免费电影在线观看| 巨乳人妻的诱惑在线观看| 午夜福利免费观看在线| 超色免费av| 老司机午夜福利在线观看视频| 精品久久久久久久久久免费视频 | 在线观看午夜福利视频| 欧美黄色淫秽网站| 国产1区2区3区精品| 老熟妇乱子伦视频在线观看| 女人高潮潮喷娇喘18禁视频| 99国产精品免费福利视频| 免费观看人在逋| 免费观看人在逋| 国产精品久久视频播放| 久久精品91无色码中文字幕| 欧美亚洲 丝袜 人妻 在线| 99热网站在线观看| 亚洲成国产人片在线观看| 美女福利国产在线| 咕卡用的链子| 999久久久精品免费观看国产| 不卡一级毛片| 亚洲精品国产区一区二| 国产高清激情床上av| 亚洲av熟女| 女人精品久久久久毛片| 免费一级毛片在线播放高清视频 | 天堂中文最新版在线下载| 中文欧美无线码| 亚洲国产中文字幕在线视频| 高清欧美精品videossex| 老熟女久久久| 女性生殖器流出的白浆| 国产黄色免费在线视频| 久久久久精品国产欧美久久久| 成年版毛片免费区| 亚洲 国产 在线| 亚洲av电影在线进入| 亚洲黑人精品在线| 操美女的视频在线观看| 精品国产乱子伦一区二区三区| 热99国产精品久久久久久7| 婷婷成人精品国产| 老司机午夜福利在线观看视频| 好看av亚洲va欧美ⅴa在| 久久香蕉精品热| 波多野结衣av一区二区av| 午夜福利,免费看| 国产成人精品久久二区二区91| 久99久视频精品免费| 欧美大码av| 欧美不卡视频在线免费观看 | 国产黄色免费在线视频| 大片电影免费在线观看免费| 亚洲自偷自拍图片 自拍| 亚洲美女黄片视频| 免费女性裸体啪啪无遮挡网站| 久久这里只有精品19| 真人做人爱边吃奶动态| 免费看a级黄色片| 国产精品二区激情视频| 亚洲一码二码三码区别大吗| 国产97色在线日韩免费| 看黄色毛片网站| 亚洲全国av大片| 人妻一区二区av| 久久精品91无色码中文字幕| 在线av久久热| netflix在线观看网站| 国产深夜福利视频在线观看| 免费高清在线观看日韩| 亚洲人成伊人成综合网2020| 丁香六月欧美| 国产无遮挡羞羞视频在线观看| 两个人免费观看高清视频| 少妇裸体淫交视频免费看高清 | 精品第一国产精品| 少妇 在线观看| 久久亚洲精品不卡| 啪啪无遮挡十八禁网站| 高清av免费在线| 国产乱人伦免费视频| 999久久久精品免费观看国产| 午夜福利,免费看| 99国产综合亚洲精品| 老司机午夜福利在线观看视频| 精品久久久久久久久久免费视频 | 午夜精品久久久久久毛片777| 91麻豆av在线| 王馨瑶露胸无遮挡在线观看| 亚洲五月色婷婷综合| 婷婷精品国产亚洲av在线 | 国产在线精品亚洲第一网站| 成人特级黄色片久久久久久久| 热99久久久久精品小说推荐| 一边摸一边抽搐一进一小说 | 国产日韩欧美亚洲二区| 极品教师在线免费播放| 国产区一区二久久| 亚洲自偷自拍图片 自拍| 亚洲国产精品sss在线观看 | 亚洲国产毛片av蜜桃av| 亚洲国产精品一区二区三区在线| 国产精品亚洲av一区麻豆| 韩国av一区二区三区四区| 日韩欧美国产一区二区入口| 国产免费现黄频在线看| av片东京热男人的天堂| 激情在线观看视频在线高清 | 亚洲五月天丁香| 午夜成年电影在线免费观看| 中文字幕最新亚洲高清| 亚洲专区字幕在线| 精品人妻1区二区| 国产一区二区三区在线臀色熟女 | 欧美成狂野欧美在线观看| 欧美久久黑人一区二区| 麻豆成人av在线观看| 国产精品国产av在线观看| 久久天躁狠狠躁夜夜2o2o| av一本久久久久| 亚洲精品成人av观看孕妇| 高清视频免费观看一区二区| 国产精品成人在线| 9色porny在线观看| 午夜视频精品福利| 亚洲av电影在线进入| 777米奇影视久久| 国产亚洲av高清不卡| 亚洲久久久国产精品| 青草久久国产| 成人精品一区二区免费| 亚洲一卡2卡3卡4卡5卡精品中文| 大码成人一级视频| 岛国在线观看网站| 黄片小视频在线播放| 亚洲人成电影免费在线| 无限看片的www在线观看| 精品电影一区二区在线| 90打野战视频偷拍视频| 亚洲精品乱久久久久久| 777久久人妻少妇嫩草av网站| 亚洲久久久国产精品| 18禁黄网站禁片午夜丰满| 无限看片的www在线观看| 国产精品香港三级国产av潘金莲| 电影成人av| 淫妇啪啪啪对白视频| 久久婷婷成人综合色麻豆| 亚洲精品国产精品久久久不卡| aaaaa片日本免费| 亚洲国产精品合色在线| 久久精品国产清高在天天线| 国产精品成人在线| 国产成人av激情在线播放| 999久久久国产精品视频| 久久香蕉国产精品| 首页视频小说图片口味搜索| 曰老女人黄片| 欧美激情 高清一区二区三区| 亚洲精品中文字幕一二三四区| 国产精品一区二区在线不卡| 变态另类成人亚洲欧美熟女 | 中文字幕人妻熟女乱码| 夜夜躁狠狠躁天天躁| 丰满饥渴人妻一区二区三| 亚洲精品美女久久久久99蜜臀| 十八禁人妻一区二区| 在线观看舔阴道视频| 亚洲精品在线观看二区| 亚洲人成77777在线视频| 免费在线观看亚洲国产| 九色亚洲精品在线播放| 丰满的人妻完整版| 久久午夜亚洲精品久久| 在线国产一区二区在线| 亚洲欧美激情在线| 大香蕉久久网| 精品久久久久久电影网| 一本大道久久a久久精品| 精品视频人人做人人爽| a在线观看视频网站| 国产精品电影一区二区三区 | 亚洲一区二区三区欧美精品| av天堂久久9| 国产高清视频在线播放一区| 免费观看人在逋| 国产激情久久老熟女| 一二三四在线观看免费中文在| 精品人妻在线不人妻| 丝袜人妻中文字幕| 久久人妻av系列| 亚洲 国产 在线| 大型av网站在线播放| 怎么达到女性高潮| 悠悠久久av| 露出奶头的视频| 国产成人啪精品午夜网站| 国产成人欧美| 欧美激情高清一区二区三区| 欧美中文综合在线视频| 热re99久久精品国产66热6| 国产亚洲欧美在线一区二区| 成人三级做爰电影| 久久久精品区二区三区| 天天添夜夜摸| 国产精品 欧美亚洲| videos熟女内射| 别揉我奶头~嗯~啊~动态视频| 九色亚洲精品在线播放| 精品免费久久久久久久清纯 | 亚洲美女黄片视频| 丝袜人妻中文字幕| 亚洲av成人一区二区三| 少妇被粗大的猛进出69影院| 国产国语露脸激情在线看| www日本在线高清视频| 免费少妇av软件| 黄片大片在线免费观看| 日韩免费高清中文字幕av| 他把我摸到了高潮在线观看| 亚洲成a人片在线一区二区| 天堂俺去俺来也www色官网| 俄罗斯特黄特色一大片| 精品久久久久久久久久免费视频 | 色播在线永久视频| 岛国毛片在线播放| 精品少妇久久久久久888优播| 国产精品香港三级国产av潘金莲| 亚洲成a人片在线一区二区| 久久国产精品影院| 国产一区二区三区在线臀色熟女 | 成人影院久久| 波多野结衣av一区二区av| 久久香蕉激情| 精品国产超薄肉色丝袜足j| 视频在线观看一区二区三区| 色婷婷久久久亚洲欧美| 国产精品一区二区在线观看99| 久久人妻av系列| 黑人巨大精品欧美一区二区蜜桃| 久久精品成人免费网站| 咕卡用的链子| 亚洲成人国产一区在线观看| 国产97色在线日韩免费| 妹子高潮喷水视频| av超薄肉色丝袜交足视频| 91精品国产国语对白视频| 久久精品国产亚洲av香蕉五月 | 变态另类成人亚洲欧美熟女 | 精品国内亚洲2022精品成人 | 久久人妻熟女aⅴ| 人人妻人人添人人爽欧美一区卜| 两个人免费观看高清视频| 中文字幕人妻丝袜制服| 中文字幕av电影在线播放| 国产精品国产av在线观看| 亚洲成a人片在线一区二区| 精品一区二区三区视频在线观看免费 | 亚洲色图 男人天堂 中文字幕| 超碰成人久久| 欧美激情极品国产一区二区三区| www.熟女人妻精品国产| 国产在线观看jvid| 国产色视频综合| 成年人午夜在线观看视频| 成人亚洲精品一区在线观看| 午夜免费鲁丝| 欧美老熟妇乱子伦牲交| 一级片'在线观看视频| 婷婷精品国产亚洲av在线 | 亚洲国产精品合色在线| 成人免费观看视频高清| 亚洲欧美日韩高清在线视频| 国产欧美日韩一区二区三| 人妻久久中文字幕网| 国产日韩欧美亚洲二区| 精品熟女少妇八av免费久了| 一边摸一边做爽爽视频免费| 精品免费久久久久久久清纯 | 欧美日本中文国产一区发布| 人妻久久中文字幕网| 18禁观看日本| 如日韩欧美国产精品一区二区三区| 不卡av一区二区三区| 国产一区二区三区综合在线观看| 欧洲精品卡2卡3卡4卡5卡区| 亚洲国产精品sss在线观看 | 日韩欧美三级三区| 亚洲精品乱久久久久久| 欧美日韩亚洲高清精品| 1024视频免费在线观看| 女人爽到高潮嗷嗷叫在线视频| 日本黄色日本黄色录像| 色尼玛亚洲综合影院| 国产日韩欧美亚洲二区| 日本一区二区免费在线视频| 欧美成人午夜精品| 午夜福利在线观看吧| 香蕉久久夜色| 国产有黄有色有爽视频| 国产1区2区3区精品| 国产欧美亚洲国产| 精品少妇久久久久久888优播| 99热只有精品国产| 久久久久精品国产欧美久久久| 国产不卡一卡二| 国产亚洲欧美精品永久| 国产成人系列免费观看| 两人在一起打扑克的视频| 免费看a级黄色片| 99热网站在线观看| videos熟女内射| 免费不卡黄色视频| 国产在线一区二区三区精| 久99久视频精品免费| 好男人电影高清在线观看| 人人妻人人澡人人看| 国产精品影院久久| 国产单亲对白刺激| 国产不卡一卡二| 国产男女内射视频| av欧美777| 中出人妻视频一区二区| 999久久久精品免费观看国产| 亚洲欧美一区二区三区黑人| 亚洲专区字幕在线| 国产精品免费一区二区三区在线 | 在线播放国产精品三级| 精品一品国产午夜福利视频| av网站在线播放免费| 欧美av亚洲av综合av国产av| 国产单亲对白刺激| 欧美精品高潮呻吟av久久| 免费不卡黄色视频| 十八禁高潮呻吟视频| 日韩一卡2卡3卡4卡2021年| 三上悠亚av全集在线观看| 欧美色视频一区免费| 女人被躁到高潮嗷嗷叫费观| 欧美日韩福利视频一区二区| 12—13女人毛片做爰片一| 精品熟女少妇八av免费久了| 中文字幕制服av| 久久久精品国产亚洲av高清涩受| 久久人人97超碰香蕉20202| 操美女的视频在线观看| 日韩成人在线观看一区二区三区| 欧美黄色淫秽网站| 巨乳人妻的诱惑在线观看| 欧美日韩亚洲国产一区二区在线观看 | 每晚都被弄得嗷嗷叫到高潮| 亚洲一卡2卡3卡4卡5卡精品中文| 精品电影一区二区在线| 日本vs欧美在线观看视频| 免费黄频网站在线观看国产| 亚洲第一欧美日韩一区二区三区| 亚洲美女黄片视频| 老司机福利观看| 久久精品熟女亚洲av麻豆精品| 亚洲avbb在线观看| 丰满的人妻完整版| 欧美久久黑人一区二区| 亚洲熟妇熟女久久| 美女福利国产在线|