• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Source free unsupervised domain adaptation for electro-mechanical actuator fault diagnosis

    2023-05-19 03:39:34JianyuWANGHengZHANGQiangMIAO
    CHINESE JOURNAL OF AERONAUTICS 2023年4期

    Jianyu WANG, Heng ZHANG, Qiang MIAO

    College of Electrical Engineering, Sichuan University, Chengdu 610065, China

    KEYWORDSData privacy;Electro-mechanical actuator;Pseudo-label clustering;Nearest centroid filtering;Unsupervised domain adaptation

    AbstractA common necessity for prior unsupervised domain adaptation methods that can improve the domain adaptation in unlabeled target domain dataset is access to source domain dataset and target domain dataset simultaneously.However, data privacy makes it not always possible to access source domain dataset and target domain dataset in actual industrial equipment simultaneously,especially for aviation component like Electro-Mechanical Actuator(EMA)whose dataset are often not shareable due to the data copyright and confidentiality.To address this problem,this paper proposes a source free unsupervised domain adaptation framework for EMA fault diagnosis.The proposed framework is a combination of feature network and classifier.Firstly,source domain datasets are only applied to train a source model.Secondly,the well-trained source model is transferred to target domain and classifier is frozen based on source domain hypothesis.Thirdly,nearest centroid filtering is introduced to filter the reliable pseudo labels for unlabeled target domain dataset, and finally, supervised learning and pseudo label clustering are applied to fine-tune the transferred model.In comparison with several traditional unsupervised domain adaptation methods,case studies based on low-and high-frequency monitoring signals on EMA indicate the effectiveness of the proposed method.

    1.Introduction

    Electro-Mechanical Actuators(EMAs)have drawn wide attentions in modern aviation aircrafts like A-380 and Boeing 787.1,2Mechanical component in EMA is directly driven by electrical motor, and thereby, it can relieve complex structure and avoid oil leakage compared with traditional integrated actuator package and electro-hydraulic actuator.3To avoid catastrophic accidents caused by unexpected failures, health monitoring system is one of the most important systems in airplane, especially for the monitoring of EMA due to its critical role in aileron and landing gear.4As one of the key approaches in health monitoring system, fault diagnosis can identify failures based on monitoring signal.

    In recent years,deep learning draws wide attention in actuator fault diagnosis because of its ability to automatically mine features from original signal and map them into different failure labels.It is easy to generalize it directly to different actuators, no longer labor intensive in developing observer and extracting features in traditional model- and feature-based methods,5respectively.Until now,many popular deep learning models such as deep auto-encoder,6Convolution Neural Network (CNN),7long-short term memory network,8and etc.,have been validated to be more effective than traditional feature-based methods in actuator fault diagnosis.However,a common limitation for those methods deployed in previous studies is that they only perform well in testing dataset with the same distribution as that in training dataset, while their performance will decrease a lot when facing a testing dataset under new working condition.Actually,EMA is usually driven by different commands with varying amplitudes and frequencies, and also, works in complicated loading conditions and various velocity environments.9Hence, all of those variable working conditions bring a data distribution discrepancy challenge for the application of aforementioned methods.

    To cope with the distribution discrepancy challenge, transfer learning is introduced to enhance the model generalization ability in testing dataset under new working condition.10The key idea of transfer learning model is to take advantages of knowledge in source domain(i.e.,training dataset in old working condition)to serve for target domain(i.e.,testing dataset in new working condition) so as to reduce distribution discrepancy between these two domains and improve domain adaptation ability of transferred model in target domain dataset.11The domain adaptation task in transfer learning can be divided into two categories,12i.e., supervised domain adaptation and unsupervised domain adaptation, depending on whether there are labels in target domain dataset.For supervised domain adaptation, instance-weight samples13and fine-tuning network14are usually applied to achieve domain adaptation in target domain dataset.The former, by enhancing the weight of similar samples in source domain, can implement the data augmentation for target domain samples.The latter aims to fine-tune the parameters of model trained by source domain dataset, and then, accommodate it adapt well in target domain.Labeled target domain samples are required for these two efforts.However, it is generally labor-intensive and timeconsuming to label those samples, and even impossible to acquire the labels for them in actual industrial equipment.

    In comparison with supervised domain adaptation, unsupervised domain adaptation (i.e., without labels in target domain dataset)is more realistic for mechanical fault diagnosis on actual industrial equipment.There are two popular unsupervised domain adaptation strategies,15i.e.,feature alignment and adversarial learning.The feature alignment introduces an additional distribution metric criterion in deep learning model to reduce the distribution discrepancy between two domains.For example,Lu et al.16utilized the Maximum Mean Discrepancy (MMD) to reduce the distribution distance in high dimensional space of deep learning network, which could address the distribution discrepancy in bearing and gear fault diagnosis, respectively.The correlation alignment loss can align the second order statistics of two domains, and then, it was introduced to validate its feasibility in bearing fault diagnosis under variable working conditions.17Subsequently,some improved distribution metric criterions like multi-kernel MMD, joint MMD, and so on, are also explored in different efforts.15,18,19In virtue of adversarial learning in generative adversarial network, the generator in unsupervised domain adaptation can create indistinguishable features for two domains.20Chen et al.21conducted the Domain Adversarial Network (DAN) in bearing and gear fault diagnosis, which indicated that it can perform well in unseen working conditions.Since only inter-domain clustering is concerned in DAN, further, Han et al.22combined the triplet loss with adversarial learning to strengthen the intra-domain clustering for fault diagnosis of rotating machinery.In addition, more variants inspired by DAN are also investigated in different efforts.23,24Those efforts have made advanced progress in addressing the distribution discrepancy challenge.However,the access to source domain dataset for feature alignment or adversarial learning approach is inevitable.In other words,source domain dataset and target domain dataset need to be involved in the model training process simultaneously.Once source domain dataset are absent, these two unsupervised domain adaptation approaches may not work properly.

    Monitoring dataset in some industrial equipment are unavailable for researchers due to some laws and regulation constraints on data.Hence, a noteworthy phenomenon in actual equipment is that data privacy or potential conflict of interest from data are concerned for many owners.Many owners just want researchers to utilize them in their own interests but not share them,which is due to the fact that data copyright and confidentiality are of the utmost importance for different equipment, even for the data sampled from the same equipment but with different working conditions.Therefore, the data privacy of source domain dataset will greatly impede the application of traditional unsupervised domain adaptation methods in industrial equipment.In recent years, federated learning draws wide attention in protecting the data privacy.One idea in federated learning is to upload different private data to the central server and utilize encryption techniques to protect them.25However, data leakage risk still remains when central server is attacked or improperly managed.The other idea is to train different sub-models for different local dataset, and then, parameters from those models are aggregated to a global model to make it adapt well in different local dataset.26However, labeled samples are necessary to train those sub-models.As the challenge mentioned in supervised domain adaptation, the labels collection in actual industrial equipment is time-consuming and labor-intensive, or even impossible.

    Hence, how to take advantages of knowledge in source domain dataset and protect their data privacy while still improving the domain adaptation for unlabeled target domain dataset is concerned in this paper.In other words, source domain dataset and target domain dataset would not been involved in model training process together.Similar to the fine-tuning network in supervised domain adaptation, Source Free Unsupervised Domain Adaptation(SFUDA)27can transfer the source model and fine-tune it to adapt well in unlabeled target domain dataset.Actually, the transferred deep learning model is a black box that is not interpretable for original dataset,28so it can protect the data privacy for original dataset while storing its knowledge in well-trained model.Until now,some researchers have noticed the SFUDA challenge in semantic segmentation29and object detection.30The key idea of fine-tuning model in target domain is to explore the reliable and valuable target samples for supervised learning.Therefore,pseudo labels will be estimated by the transferred model, and then, two available approaches,27i.e., pseudo label clustering and pseudo label filtering, are applied to enhance the domain adaptation based on supervised learning.Liang et al.31proposed a nearest classifier to minimize the centroid shift and improve clustering effect for pseudo labels.Further, information maximization and supervised pseudo labeling are introduced in target domain to improve the pseudo label clustering of transferred model.32For pseudo label filtering,a source domain hypothesis assumes that the pseudo labels from similar target domain dataset are close to the actual labels trained in source domain, and then, the similar samples will be applied to guide the training process of transferred model.A valuable phenomenon founded by Kim et al.33is that the target samples with low self-entropy will have higher reliability pseudo labels.Thereby, those samples with low-entropy will be filtered to improve the domain adaptation ability.In addition, contrastive learning is applied to give different weights for different target samples, and those samples with high confidence will be equipped with heavy weights.34Despite the advances of SFUDA aforementioned, there are still relatively limited researches focusing on the mechanical fault diagnosis to protect data privacy of source domain and improve the domain adaptation in unlabeled target domain, especially for the one-dimensional time series monitoring signal.For the monitoring dataset sampled from aviation airplane, it is usually critical to monitor the state of key component and their data privacy can not be ignored.

    For the EMA, dozens of sensors with different sampling frequencies are applied to monitor its state.Different time series signals from different sensors may reflect different failure information.EMA is usually driven by command with different amplitudes,and also,works in variable working conditions like various loadings and speeds.The distribution discrepancy challenge in EMA fault diagnosis is inevitable.In consideration of its special application scenarios like in aileron or landing gear of aircraft, labeling new failure samples is a difficult task.Hence, the dataset collected by different users sampling from different working conditions also have data copyright and confidentiality,especially for those valuable labeled source domain dataset.Based on those preliminary hypothesis, the source domain dataset and target domain dataset can not meet each other in model training process.The source domain dataset from one working condition are labeled while their data privacy need to be protected.The target domain dataset from the other working condition are unlabeled while improving domain adaptation of transferred model is concerned.

    It is still not clear whether SFUDA method can adapt well in domain adaptation task of EMA,because different methods may adapt well in different dataset and some traditional domain adaptation methods may bring a negative transfer learning phenomenon.Hence, a comparison between the SFUDA method and several traditional domain adaptation methods has been explored in two case studies of EMA fault diagnosis.The traditional domain adaptation methods need to access the source domain and target domain simultaneously,while the SFUDA framework is applied to address the unsupervised domain adaptation challenge and protect the data privacy of source domain.The model in the proposed framework is a combination of two networks, i.e., feature network and classifier.The proposed model follows the source domain hypothesis,32i.e., predicted encoding pseudo labels (one-hot encodings) from target domain dataset are similar to those of source domain dataset based on frozen classifier.Source domain dataset are only applied to train the source model firstly.Then, the well-trained source model will be transferred to target domain but the parameters in classifier are not allowed to be fine-tuned.Further, pseudo label filtering(guided by nearest centroid filtering) is applied to filter the noise pseudo labels, and then, supervised learning (guided by pseudo labels) and pseudo label clustering (guided by mutual information maximizing) are applied to fine-tune the feature network.Finally,the fine-tuned model can be applied in unlabeled testing target domain dataset.The main contributions of this paper are summarized as follows:

    (1) Benefitted from two separate training process of the proposed framework in source domain dataset and target domain dataset respectively,the data privacy of source domain dataset can be protected and the generalization ability of transferred model can be improved by several fine-tuning strategies in unlabeled target domain dataset.

    (2) Multiple monitoring signals sampled by different sampling frequencies but for the same failures are applied to illustrate the effectiveness of the proposed method in low- and high-frequency case studies.

    (3) This paper gives a comparison about the proposed method and traditional unsupervised domain adaptation methods in unsupervised domain adaptation task of EMA fault diagnosis, which can provide a benchmark for future studies.

    The rest of this paper are organized as follows.The details of proposed method are introduced in Section 2.Experiment descriptions including EMA dataset and implementation details are introduced in Section 3.Case studies are explored in Section 4.Conclusion is summarized in Section 5.

    2.Proposed method

    Although many unsupervised domain adaptation methods have been applied in mechanical fault diagnosis, the access to source domain dataset in domain adaptation stage is contrary to the data privacy.Hence, a SFUDA framework is applied to protect the data privacy of source domain dataset while improving the unsupervised domain adaptation in target domain dataset for EMA fault diagnosis.Section 2.1 illustrates the overview of the proposed method that can be divided into source model generation and target domain adaptation.In Section 2.2, the details of source model generation are introduced.In Section 2.3, the details of target model adaptation are introduced.In Section 2.4, model parameters and flow chart of the proposed method are illustrated.

    2.1.Overview of proposed method

    where Llossis a loss function.However, due to the noise pseudo labels existing in Eq.(1), the domain adaptation optimization of the model FTmay be impeded.

    Following the basic idea in SFUDA, the proposed model FTis a combination of two networks including feature network fE(?;θE)and classifier fC(?;θC), as shown in Fig.1.Feature network is a one-dimensional convolutional network that can mine the robust features from original time series signal.Classifier is a combination of fully connected layers and softmax function, which is applied to predict labels.Two stages are included in the proposed method,i.e.,source model generation and target domain adaptation.In source model generation, only source domain dataset with their labels are applied to train those two sub-networks.The source domain hypothesis used in this paper assumes that the pseudo labels(i.e., one-hot encodings) from target domain dataset are similar to those of source domain dataset based on the well-trained classifier.35In target domain adaptation, the well-trained model is transferred to target domain and classifier would be fixed based on source domain hypothesis.Later, the proposed method utilizes the nearest centroid filtering to filter the noise pseudo labels.The supervised learning through Eq.(2) and pseudo label clustering through mutual information maximization, and then, are applied to fine-tune the feature network.Finally,the fine-tuned network FTcan be applied to diagnosis the monitoring signal in EMA.

    2.2.Source model generation

    Pooling layers usually follow the convolutional layers,which can down sample the features and help to reduce the number of parameters in model.Two types of pooling layers,i.e., max pooling layer and average pooling layer, are usually adopted to acquire maximum value and mean value, respectively.With the help of pooling layer, the activation features in (l+1)-th layer can be described as follows:

    The fully connected layers transform multi-kernel features into one-dimensional features.Then, a softmax function is applied to predict the class labels based on the onedimensional features.The CNN learns to make class prediction by minimizing the cross-entropy loss Lsrcbetween the predicted and true labels.

    Fig.1 Framework of the proposed unsupervised domain adaptation model with source free dataset.

    where α is a smoothing parameter and is set to 0.1 in this paper; K stands for the number of classes.

    2.3.Target domain adaptation

    The key idea of SFUDA in this paper is to fine-tune the welltrained source model based on unlabeled target domain dataset.Benefitted from the blank box characteristic of deep learning model, the data privacy of source domain dataset can be protected by transferring only source model to target domain.The source domain hypothesis assumes that the pseudo labels predicted by classifier in transferred model are similar to the labels in source domain dataset.Hence, the classifier is fixed in target domain adaptation process while feature network is fine-tuned by pseudo labels predicted by classifier.To filter the noise pseudo labels,pseudo label filtering(i.e.,nearest centroid filtering) is applied to filter the reliable pseudo labels for supervised learning and pseudo label clustering.Here, the pseudo label clustering based on mutual information maximization is introduced firstly,and then,the necessity of pseudo label filtering is further explained.Finally, the network finetuning process in target domain is concluded.

    2.3.1.Pseudo label clustering

    For traditional domain adaptation methods, it is easy to measure the distribution discrepancy between two domains when source domain dataset are not absent in domain adaptation stages.However, the source domain dataset are unavailable in target domain adaptation stage due to the data privacy.Based on the source domain hypothesis, the ideal outputs of target datasets would be similar to the one-hot encoding vectors predicted in source domain while belong to different classes.To reinforce the above hypothesis,mutual information maximization38is introduced to make the pseudo one-hot encoding vectors more individually decisive and globally fair.

    The general idea of mutual information maximization is to retain as much information of input samples in labels as possible.Hence, the mutual information L(c;x )between the input samples x and the labels c can be described as follows:

    For the SFUDA, since the classifier is fixed, a good domain adaptation model for target domain should extract robust features from the feature network, and then, make the pseudo labels from classifier more individually certain and follow an even distribution between the classes.Based on the Eq.(10), thereby, maximizing mutual information LMIin the target domain of the proposed method can be illustrated as follows:

    where FT=fTE(?;θE)+fTC(?;θC); Ext?Xt[?] is an average operation.

    2.3.2.Pseudo label filtering

    The mutual information maximization can be applied to strengthen the pseudo label clustering,however,some samples are still matched to a wrong clustering center due to the effect of noise pseudo labels.Hence, this phenomenon will misguide the fine-tuning process, and even bring a negative effect on domain adaptation model.To alleviate such negative effect of noise pseudo labels, the nearest centroid filtering is applied to filter the pseudo labels before maximizing the mutual information.

    2.3.3.Network fine-tuning

    Based on such double nearest centroid filtering mechanisms from Eq.(12) to (15), the pseudo labels are reliable enough to relieve the noise labels in them, which is beneficial to strengthen the pseudo label clustering effect.Furthermore,those filtering pseudo labels will also be applied for the supervised learning in target domain dataset40.

    2.4.Model parameters and flow chart

    The architecture and parameters of the proposed model includes feature network and classifier, as shown in Table 1.The feature network includes four convolutional layers (Convolution 1–4), four batch normalization layers (Batchnorm 1–4), two pooling layers (Maxpooling and Averagepooling)and one fully connected layer.Three fully connected layers and softmax function are included in classifier.Parametric Rectified Linear Unit (PReLU) is regarded as the activation function in two networks, and two dropout strategies are applied to relieve the over-fitting effect in classifier.

    The main idea of the proposed model has been introduced from Section 2.1 to Section 2.3, the detailed flow chat of the proposed model is illustrated as follows:

    3.Experiment descriptions

    In this section,experiment descriptions including EMA dataset and implementation details are introduced.The EMA dataset based on two different monitoring signals with different sampling frequencies is introduced in Section 3.1.The implementation details of the proposed method and several comparison methods are illustrated in Section 3.2.

    3.1.Electro-mechanical actuator dataset

    The NASA flyable EMA testbed41is shown in Fig.2 (a).In Fig.2 (b), there are three actuators installed in the test bed.Actuator X is applied to simulate different failures.Actuator Y is treated as a comparison without injecting any failures.Actuator Z is applied to simulate the external load and apply force to the first two actuators.Three failures, i.e., ballscrew jam,sensor fault and spall fault,and normal state are regarded as four labels in this paper.Since unsupervised domain adaptation task is considered in this paper, eight working conditions considering different commands, amplitudes,frequencies, loads and velocities, are considered in this paper.Here, they are divided into two groups according to the sinusoidal and trapezoidal commands, as shown in Table 2.

    Two types of monitoring signals, i.e., low sampling frequency (100 Hz) signals and high sampling frequency(20 kHz) signals, are applied to monitor the state of actuator,which may reflect different failure information.Since there is big sampling frequency difference between them, it is difficult to combine them directly.Hence, two case studies are conducted in this paper,respectively.For low sampling frequency signals, four types of signals, i.e., motor voltage, motor current, motor temperature, and nut temperature, are regarded as the low-frequency case study.For high sampling frequencysignals,acceleration signals from nut accelerometer in x,y and z directions are regarded as the high-frequency case study.

    Table 1 Parameters of the proposed method.

    Fig.2 FLEA of NASA.

    Table 2 Experiment dataset under variable working conditions.

    3.2.Implementation details

    The input size of the low- and high-frequency case studies in the proposed model are different.For low-frequency signal,a sliding window with size 256 is applied to capture points from four signals.Therefore, the size of each sample under low-frequency is a 1×1024 vector reshaped by 4×256 points.For high-frequency signals, a sliding window with a size of 1024 is applied to capture points from three signals in turns.Therefore, the size of each sample under high-frequency is a 1×3072 vector reshaped by 3×1024 points.In each working condition,there are total 600 samples with 150 samples in each label.50 % samples in source domain and target domain are set as the training data,while the other 50%samples in target domain are regarded as testing data.12 cross-domain adaptation tasks within the same command are constructed in following experiment.The unsupervised domain adaptation task based on low-frequency signal is defined by A, while the task based on high-frequency signal is defined by B.For instance,12 domain adaptation of task A under sinusoidal command will be grouped as A1-A2, A1-A3, A1-A4, A2-A1, A2-A3,A2-A4, A3-A1, A3-A2, A3-A4, A4-A1, A4-A2, and A4-A3.The balancing-factor λ1and λ2in Eq.(16) are set to 1 and 0.1 before exploring its influence in Section 4.2 and Section 4.3,respectively.Adam optimizer with learning rate 0.01 is adopted.Batch size and iterative epoch are 64 and 30, respectively.The average classification accuracy of ten repeated experiment is considered as an evaluation criterion in all case studies.In addition, the basic CNN and other six traditional feature alignment and adversarial learning models are regarded as the comparison methods in this paper.It needs to be pointed out that DeepCoral, DeepMMD, DAN and DAN + MMD only consider reducing the global interdomain distance between two domains, while DANT-S and DANT-ST add the triplet loss to help to reduce the intraclass domain distance based on DAN.

    (1)CNN:It is a baseline network trained by source domain dataset.

    (2)DeepCoral17:Correlation alignment is applied to reduce second-order statistics distance of two domains in feature network.

    (3)DeepMMD16:MMD is applied to reduce mean distance of Hilbert space of two domains in feature network.

    (4) DAN20: It is a domain adversarial network using Kullback-Leibler divergence in discriminator.

    (5) DAN + MMD23: It is a combination of DAN and MMD.

    (6)DANT-S22:It is a combination of DAN and triplet loss,however, only source domain samples appear in triplet loss.

    (7) DANT-ST24: Based on the DANT-S method, target domain samples with pseudo labels are further applied in triplet loss.

    4.Case study

    In this section, the effectiveness of the proposed method is investigated on several case studies.In Section 4.1,the comparison results of traditional unsupervised domain adaptation methods and the proposed method are conducted on lowfrequency and high-frequency case studies, respectively.In Section 4.2, the influence of supervised learning in target domain adaptation is discussed.In Section 4.3, the influence of pseudo label clustering in target domain adaptation is further discussed.In Section 4.4, the negative transfer phenomenon of the proposed method in high-frequency case study under trapezoidal command is discussed.

    4.1.Performance on unsupervised domain adaptation

    Two case studies are designed for low-and high-frequency signals, respectively.For each case study, 12 domain adaptation tasks are further explored under sinusoidal and trapezoidal command, respectively.

    4.1.1.Low-frequency case study

    The comparison results of eight methods under sinusoidal command are shown in Table 3 and Fig.3.Among the 12 domain adaptation tasks, the proposed method obtains the best average accuracy 92.45 %.Except for the DANT-S,DANT-ST and the proposed method, negative transfer phenomenon exists in most domain adaptation tasks for the other four domain adaptation methods compared to CNN, especially for DeepMMD.Therefore, the average accuracy of DeepCoral,DeepMMD,DAN,and DAN+MMD are lower than that average accuracy 87.22%in basic CNN.With introducing the triplet loss to enhance intra-domain clustering based on pseudo labels, DANT-S and DANT-ST achieve better average accuracy than CNN, which indicates that intraclass domain clustering is beneficial to improve the domain adaptation.However, it is still necessary to acquire the source domain dataset and make them participate in unsupervised domain adaptation for those traditional domain adaptation methods.For the proposed method, only source model is transferred to target domain, which can protect the data privacy of source domain.Even though the source domain dataset are not available for target domain, the proposed method obtains the better performance than DANT-S and DANTST in most domain adaptation tasks.

    The comparison results of eight methods under trapezoidal command are shown in Table 4 and Fig.4.The average accuracy drops a lot among eight methods indicate that the domain adaptation tasks under trapezoidal command are more challenging than that under sinusoidal command.Only DANTST and the proposed method obtain the better average accuracy than CNN.However, severe negative transfer occurring in other five domain adaptation methods (DeepCoral,DeepMMD, DAN, DAN + MMD, and DANT-S) result in their low average accuracy, which may be due to the fact that only reducing the global inter-domain distance is considered in these efforts.In other words, although the source domain dataset are involved in domain adaptation training process,the improvement of those methods are still limited,which indicate that the EMA dataset bring a challenge for traditional domain adaptation methods.The proposed method only transfers the source model,and then,fine-tunes the transferred model to make it adapt well in target domain.It obtains the best accuracy in most domain adaptation tasks and also obtains the best average accuracy 86.67 %.For the lowfrequency monitoring signal, to conclude, the proposed method can achieve the stable performance and obtain the best average accuracy in domain adaptation tasks under sinusoidal command and trapezoidal command, respectively.

    4.1.2.High-frequency case study

    The comparison results of eight methods under sinusoidal command are shown in Table 5 and Fig.5.The CNN in high-frequency case study obtains the better average accuracy than that in Table 3 of low-frequency case study, which illustrates that different monitoring signals contain the different failure information.Different from the good performance of DANT-ST in Table 3 and Table 4, the negative transfer phenomenon exists in DANT-ST in this experiment.Only DAN + MMD and the proposed method obtain the better average accuracy than CNN.The DAN + MMD obtainsthe best accuracy in B2-B1, B2-B4 and B3-B4 tasks, however,the difference of the results between the proposed method and DAN+MMD under these three tasks are very small.In addition,the CNN obtains the best accuracy in five tasks, i.e., B1-B2,B2-B3,B3-B2,B4-B2,and B4-B3,which indicates there are great similarities between source domain dataset and target domain dataset.Hence, negative transfer learning occurs in all seven domain adaptation methods among those five similar domain adaptation tasks.However, the proposed method still achieves positive domain adaptation results under other seven domain adaptation tasks and obtains the best average accuracy (90.61 %) among those comparison methods.

    Table 3 Unsupervised domain adaptation accuracy of low-frequency signals under sinusoidal command.

    Fig.3 Unsupervised domain adaptation accuracy histogram of Table 3.

    Table 4 Unsupervised domain adaptation accuracy of low-frequency signals under trapezoidal command.

    Fig.4 Unsupervised domain adaptation accuracy histogram of Table 4.

    Table 5 Unsupervised domain adaptation accuracy of high-frequency signals under sinusoidal command.

    Fig.5 Unsupervised domain accuracy histogram of Table 5.

    The comparison results of eight methods under trapezoidal command are shown in Table 6 and Fig.6.Like the same inference from Table 4 and Fig.4, the domain adaptation tasks of high-frequency monitoring signals under trapezoidal command are also challenging.The CNN in high-frequency case study obtains the lower average accuracy than that in Table 4 and Table 5.Only DANT-S and DANT-ST obtain the better average accuracy than that in CNN.However, the proposed method does not obtains the impressive performance like in above three case studies.It gets better accuracy than CNN under B1-B4, B3-B1, B3-B2 and B3-B4 tasks, but the performance on other eight tasks not drop a lot compared to Deep-Coral, DeepMMD, DAN, and DAN + MMD.Further, the possible reason for its unsatisfactory performance under this case study will be discussed in Section 4.4.For the highfrequency monitoring signals, to conclude, there are different failures information between high sampling frequency and low sampling frequency signals, and also, the proposed method obtains the best average accuracy in domain adaptation tasks under sinusoidal command.

    4.2.Influence of supervised learning

    Since the supervised learning based on pseudo labels is applied to fine-tune the network in Eq.(16), the influence of the supervised learning is explored in this section.Hence, supervisedlearning in Eq.(16)by letting the λ1is equal to 0.The comparison results based on two types of monitoring signal are shown in Table 7 and Table 8, respectively.In addition, to make the comparison results more vividly,the sector diagram of Table 7 and Table 8 are shown in Fig.7 and Fig.8, respectively.For low-frequency case study, supervised learning is beneficial to improve the accuracy in almost all tasks, except for the A1-A4 and A3-A4 under trapezoidal command.The improvement of average accuracy in sinusoidal command and trapezoidal command are about 6 % and 3 %, respectively.For highfrequency case study, supervised learning is still beneficial for most domain adaptation tasks, except for the B2-B1, B2-B3 and B2-B4 task in trapezoidal command.The improvement of average accuracy in sinusoidal command and trapezoidal command are about 4 % and 7 %, respectively.In a word,when supervised learning based on pseudo labels is absent in target domain, negative transfer phenomenon is inevitable in most domain adaptation tasks.

    Table 6 Unsupervised domain adaptation accuracy of high-frequency signals under trapezoidal command.

    Fig.6 Unsupervised domain accuracy histogram of Table 6.

    4.3.Influence of pseudo label clustering

    Since the pseudo label clustering based on mutual information maximization is applied to enhance the clustering effects in Eq.(16), the influence of the pseudo label clustering is further discussed in this section.Hence,λ2is set to different values,from nonparticipation to a big weight in Eq.(16)with 0,0.1,0.3,0.5,0.7,and 1.0,respectively.The comparison results based on two case studies are shown in Table 9 and Table 10,respectively.In addition, the histograms of those two tables are drawn in Fig.9 and Fig.10,respectively.For low-frequency case study,when λ2is set to 0,the average accuracy drops about 1%than that in previous experiments with default value 0.1,which indicates that the pseudo label clustering is still necessary to improve the domain adaptation ability.However, when λ2is set to a big value 1, the accuracy of all domain adaptation tasks drops a lot, as shown in Table 9 and Fig.9.Hence, the proposed method is sensitive to select λ2and low values like 0.1 or 0.3 are beneficial to improve the domain adaptation ability for most tasks.For high-frequency case study, the application of pseudo label clustering still improves the domain adaptation ability in most tasks, especially for the experiment under trapezoidal command.It is preferred to select the low values for λ2, for instance, the performance of 0.1 and 0.3 are also more stable in Table 10 and Fig.10.The big value 1 for λ2brings a negative transfer effect on domain adaptation task.In a word, it is necessary to take the pseudo label clustering into account of the proposed method and λ2with a low weight value is recommended in Eq.(16).

    4.4.Discussion

    The proposed method obtains the best average accuracy from Table 3 to Table 5, however, the negative transfer phenomenon occurs in most domain adaptation tasks in Table 6.The potential reason for this phenomenon is discussed in thissection.The generalization ability of the well-trained model is critical to its performance on testing dataset,and also,a robust well-trained model also performs well in training dataset.Hence, the performance of training dataset under source model and target model are discussed.The accuracy of source domain training dataset (S_train) and target domain training dataset(T_train)are explored in Table 11 and Fig.11.For S_-train, the first three experiments (Table 3 to Table 5 corre-sponding to Sinusoidal low-frequency, Trapezoidal lowfrequency and Sinusoidal high-frequency, respectively) obtain the average accuracy 99.72 %, 98.92 %, and 94.52 %, respectively.The performance of last experiment (Table 6 corresponding to Trapezoidal high-frequency) drops a lot, with only average accuracy 88.64%in source domain training data-set.Later,these source models are fine-tuned by target training dataset, however, the performance in target domain training dataset will not exceed that in source domain training dataset,which are 92.55 %, 87.16 %, 91.09 %, and 75.65 % for four experiments, respectively.Hence, the generalization ability of source model is critical to deciding the performance on target domain dataset.That is because the source domain hypothesis assumes that classifier in the well-trained source model is fixed in target domain.Hence, the poor performance of source model results in the poor domain adaptation ability in highfrequency case study under trapezoidal command.

    Table 7 Influence of supervised learning in low-frequency case study.

    Table 8 Influence of supervised learning in high-frequency case study.

    Fig.7 Unsupervised domain accuracy sector diagram of Table 7.

    Fig.8 Unsupervised domain accuracy sector diagram of Table 8.

    Table 9 Influence of pseudo label clustering in low-frequency case study.

    Table 10 Influence of pseudo label clustering in high-frequency case study.

    Fig.9 Unsupervised domain accuracy histogram of Table 9.

    Fig.10 Unsupervised domain accuracy histogram of Table 10.

    Table 11 Accuracy of training dataset in source domain and target domain.

    Fig.11 Accuracy histogram of Table 11.

    5.Conclusions

    In this paper, a source free unsupervised domain adaptation framework for EMA fault diagnosis is proposed, which can improve the domain adaptation ability in unlabeled target dataset while source domain dataset are no longer needed to participate in domain adaptation process due to their data privacy.(A)Under the distribution discrepancy challenge of variable working conditions of EMA, two case studies based on monitoring signal with different sampling frequencies but for the same failures are applied to show the effectiveness of the proposed method.(B) Further, the influence of supervised learning and pseudo label clustering are explored in different experiment, respectively, which indicate they are beneficial to improve the unsupervised domain adaptation.(C)The possible negative transfer reasons for unsupervised domain adaptation accuracy of high-frequency signals under trapezoidal command are discussed to indicate that the generalization ability of source model is critical to the proposed method.(D) Only reducing the global inter-domain distance between two domains may be not a good choice for EMA fault diagnosis.These comparison results can also provide a benchmark for future studies on unsupervised domain adaptation for EMA fault diagnosis.(E) In the future, how to fuse the multiple monitoring signals with different sampling frequencies, search optimal hyper-parameters for those methods and explore more effectiveness methods are noteworthy research hotspots in domain adaptation task of EMA fault diagnosis.

    Declaration of Competing Interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    Acknowledgements

    This work was partially supported by the National Natural Science Foundation of China (No.52075349), the Aeronautical Science Foundation of China (No.201905019001), and the China Scholarship Council (No.202106240078).

    国产黄频视频在线观看| 国产国语露脸激情在线看| 国产极品天堂在线| 国产极品天堂在线| 男人操女人黄网站| 22中文网久久字幕| 亚洲成人av在线免费| 精品国产一区二区三区久久久樱花| 国产在线免费精品| 在线播放无遮挡| h视频一区二区三区| 五月开心婷婷网| 欧美日韩综合久久久久久| 国产伦理片在线播放av一区| 午夜视频国产福利| 国产欧美日韩一区二区三区在线 | 国产欧美日韩一区二区三区在线 | 国产免费一级a男人的天堂| 日韩人妻高清精品专区| 国产极品天堂在线| 成人黄色视频免费在线看| 看非洲黑人一级黄片| 观看av在线不卡| 亚洲欧美成人综合另类久久久| 一区二区av电影网| 国产精品国产三级国产av玫瑰| 欧美少妇被猛烈插入视频| www.色视频.com| 亚洲欧美日韩卡通动漫| 高清毛片免费看| 黄色配什么色好看| 九九久久精品国产亚洲av麻豆| 精品一区在线观看国产| 97精品久久久久久久久久精品| 男人添女人高潮全过程视频| 亚洲精品自拍成人| 欧美日韩视频高清一区二区三区二| 亚洲人与动物交配视频| 麻豆成人av视频| 亚洲精品日韩在线中文字幕| 欧美日韩视频高清一区二区三区二| 久久99热6这里只有精品| 日韩亚洲欧美综合| 如何舔出高潮| 色94色欧美一区二区| 久久精品国产自在天天线| 国产日韩一区二区三区精品不卡 | 免费观看av网站的网址| 免费久久久久久久精品成人欧美视频 | 欧美另类一区| 欧美丝袜亚洲另类| 成人免费观看视频高清| 黑人高潮一二区| 美女cb高潮喷水在线观看| 久久精品国产鲁丝片午夜精品| 女性生殖器流出的白浆| 最近最新中文字幕免费大全7| 欧美日韩国产mv在线观看视频| 伊人久久国产一区二区| 看十八女毛片水多多多| 国产精品秋霞免费鲁丝片| 中文字幕av电影在线播放| 香蕉精品网在线| 亚洲无线观看免费| 亚洲精品,欧美精品| 中文字幕最新亚洲高清| 51国产日韩欧美| 久久久精品免费免费高清| 亚洲av欧美aⅴ国产| 久久久国产一区二区| 亚洲国产精品成人久久小说| 欧美激情 高清一区二区三区| 亚洲激情五月婷婷啪啪| 欧美日韩视频高清一区二区三区二| 成人二区视频| 久久久久国产网址| 一本色道久久久久久精品综合| 日韩免费高清中文字幕av| 中文字幕人妻熟人妻熟丝袜美| 欧美日韩综合久久久久久| 色吧在线观看| 国产精品一区www在线观看| 午夜视频国产福利| 五月伊人婷婷丁香| 午夜老司机福利剧场| 欧美成人午夜免费资源| 日本-黄色视频高清免费观看| 大码成人一级视频| 少妇人妻精品综合一区二区| 高清视频免费观看一区二区| 九九久久精品国产亚洲av麻豆| 一区二区三区乱码不卡18| 久久久a久久爽久久v久久| 欧美日韩一区二区视频在线观看视频在线| 午夜激情久久久久久久| 麻豆精品久久久久久蜜桃| 人人妻人人爽人人添夜夜欢视频| 91aial.com中文字幕在线观看| 观看美女的网站| 特大巨黑吊av在线直播| 插逼视频在线观看| 男女高潮啪啪啪动态图| 麻豆精品久久久久久蜜桃| 97超碰精品成人国产| 成人手机av| 日韩强制内射视频| 亚洲国产欧美日韩在线播放| 精品熟女少妇av免费看| 中文字幕最新亚洲高清| 亚洲av不卡在线观看| 黄色一级大片看看| 在线观看三级黄色| 国产精品一二三区在线看| 人妻制服诱惑在线中文字幕| 日本vs欧美在线观看视频| 2021少妇久久久久久久久久久| 夜夜爽夜夜爽视频| 男女啪啪激烈高潮av片| 9色porny在线观看| 新久久久久国产一级毛片| 精品熟女少妇av免费看| 制服人妻中文乱码| 丝袜在线中文字幕| 国产av码专区亚洲av| av又黄又爽大尺度在线免费看| 国产精品熟女久久久久浪| av在线老鸭窝| 久久久久久久久久久久大奶| 在线观看美女被高潮喷水网站| 午夜激情av网站| 在线播放无遮挡| 3wmmmm亚洲av在线观看| 日韩亚洲欧美综合| 成年人免费黄色播放视频| 国产精品秋霞免费鲁丝片| 国内精品宾馆在线| 麻豆成人av视频| 草草在线视频免费看| 一级毛片我不卡| 欧美一级a爱片免费观看看| 亚洲精品亚洲一区二区| 成人无遮挡网站| 夜夜爽夜夜爽视频| 精品久久久久久久久亚洲| 一区二区三区免费毛片| h视频一区二区三区| 国产午夜精品久久久久久一区二区三区| av在线播放精品| av黄色大香蕉| 最新的欧美精品一区二区| 天美传媒精品一区二区| 国产精品麻豆人妻色哟哟久久| 伊人亚洲综合成人网| 丰满迷人的少妇在线观看| 中文字幕亚洲精品专区| 永久免费av网站大全| 久久久久人妻精品一区果冻| 国产在线一区二区三区精| 国产成人精品久久久久久| 激情五月婷婷亚洲| 久久精品国产a三级三级三级| 在线看a的网站| 中国国产av一级| 中文字幕制服av| 一级毛片黄色毛片免费观看视频| 美女xxoo啪啪120秒动态图| 狠狠精品人妻久久久久久综合| 欧美日韩综合久久久久久| 男女边吃奶边做爰视频| 精品亚洲乱码少妇综合久久| 久久99蜜桃精品久久| 美女福利国产在线| 少妇的逼水好多| 精品久久久久久久久亚洲| 免费观看的影片在线观看| 亚洲av成人精品一区久久| 亚洲精品自拍成人| 欧美性感艳星| 午夜影院在线不卡| av福利片在线| 热99国产精品久久久久久7| 亚洲无线观看免费| 亚洲色图 男人天堂 中文字幕 | 国产永久视频网站| 哪个播放器可以免费观看大片| 男女高潮啪啪啪动态图| 男人爽女人下面视频在线观看| 日本欧美视频一区| 大陆偷拍与自拍| 成人毛片a级毛片在线播放| 日韩三级伦理在线观看| 成人无遮挡网站| 久热久热在线精品观看| 人妻 亚洲 视频| av有码第一页| 国产精品 国内视频| 少妇人妻久久综合中文| 高清不卡的av网站| 91精品国产九色| 欧美激情国产日韩精品一区| 2022亚洲国产成人精品| 亚洲内射少妇av| 亚洲综合精品二区| 丝瓜视频免费看黄片| 成年人免费黄色播放视频| 国产一区二区三区av在线| 美女xxoo啪啪120秒动态图| 在线免费观看不下载黄p国产| 天天影视国产精品| 人妻夜夜爽99麻豆av| 国产精品人妻久久久影院| 国产精品一区www在线观看| 大香蕉久久网| 午夜久久久在线观看| 国产精品蜜桃在线观看| 一级,二级,三级黄色视频| 精品少妇内射三级| 尾随美女入室| 一边亲一边摸免费视频| 久久久久久久久久久丰满| 精品人妻偷拍中文字幕| 国语对白做爰xxxⅹ性视频网站| 热99久久久久精品小说推荐| av又黄又爽大尺度在线免费看| 美女视频免费永久观看网站| 亚洲欧美一区二区三区黑人 | 亚洲国产欧美在线一区| 亚洲人成网站在线播| 国产精品 国内视频| 美女内射精品一级片tv| 另类亚洲欧美激情| 国产国拍精品亚洲av在线观看| 午夜精品国产一区二区电影| 欧美精品一区二区免费开放| 人妻少妇偷人精品九色| 狠狠婷婷综合久久久久久88av| 免费大片18禁| 欧美变态另类bdsm刘玥| 久久久久视频综合| 久久精品国产a三级三级三级| 亚洲精品久久久久久婷婷小说| 国产在视频线精品| 日韩熟女老妇一区二区性免费视频| 国国产精品蜜臀av免费| 天天影视国产精品| 桃花免费在线播放| 丝瓜视频免费看黄片| 超色免费av| 国产精品久久久久成人av| 国产精品99久久久久久久久| 五月玫瑰六月丁香| 精品亚洲成a人片在线观看| 只有这里有精品99| 亚洲欧美中文字幕日韩二区| 国产熟女欧美一区二区| 精品少妇内射三级| 国产成人精品福利久久| 午夜免费鲁丝| 久久毛片免费看一区二区三区| 最近最新中文字幕免费大全7| 91精品国产九色| 久久免费观看电影| 在线 av 中文字幕| 色94色欧美一区二区| 一级二级三级毛片免费看| 又黄又爽又刺激的免费视频.| 午夜av观看不卡| 美女国产高潮福利片在线看| 免费观看av网站的网址| 日韩欧美精品免费久久| 亚洲久久久国产精品| 大片电影免费在线观看免费| 涩涩av久久男人的天堂| 国产国语露脸激情在线看| 国产视频内射| 亚洲精品乱码久久久久久按摩| 婷婷色综合www| av有码第一页| 亚洲情色 制服丝袜| 少妇精品久久久久久久| 少妇被粗大猛烈的视频| 亚洲精品中文字幕在线视频| 亚洲一区二区三区欧美精品| 九色亚洲精品在线播放| 永久免费av网站大全| 日韩欧美精品免费久久| 伊人久久精品亚洲午夜| 人人妻人人添人人爽欧美一区卜| 国产成人一区二区在线| 亚洲国产精品一区三区| 亚洲人成网站在线观看播放| 九色成人免费人妻av| 亚洲少妇的诱惑av| 精品国产一区二区久久| 亚洲av不卡在线观看| 大香蕉久久成人网| 天堂俺去俺来也www色官网| 国产免费一级a男人的天堂| 能在线免费看毛片的网站| 岛国毛片在线播放| 啦啦啦中文免费视频观看日本| 国产又色又爽无遮挡免| 一级毛片电影观看| 赤兔流量卡办理| 久久久久视频综合| 亚洲欧美色中文字幕在线| 菩萨蛮人人尽说江南好唐韦庄| 免费看av在线观看网站| 91aial.com中文字幕在线观看| 亚洲人与动物交配视频| 日韩av在线免费看完整版不卡| 国产精品嫩草影院av在线观看| 国产成人精品在线电影| 一区二区三区四区激情视频| 不卡视频在线观看欧美| 人人澡人人妻人| 免费看不卡的av| 一级a做视频免费观看| 亚洲av成人精品一区久久| 国产一区二区在线观看av| 丝袜在线中文字幕| 国产一区二区在线观看日韩| 亚洲精品国产av蜜桃| 在线观看免费日韩欧美大片 | 免费播放大片免费观看视频在线观看| 欧美成人精品欧美一级黄| 亚洲精品久久久久久婷婷小说| 国产 精品1| 免费观看在线日韩| 高清欧美精品videossex| 国产高清不卡午夜福利| 一区二区三区乱码不卡18| 亚洲成色77777| 中文字幕av电影在线播放| 91精品三级在线观看| 狂野欧美激情性xxxx在线观看| 久久这里有精品视频免费| 国产一区亚洲一区在线观看| 国产一区有黄有色的免费视频| 亚洲av.av天堂| 黑人巨大精品欧美一区二区蜜桃 | 91久久精品国产一区二区成人| 一本一本久久a久久精品综合妖精 国产伦在线观看视频一区 | 日本黄大片高清| 美女cb高潮喷水在线观看| 最近中文字幕2019免费版| 欧美激情极品国产一区二区三区 | 亚洲国产欧美日韩在线播放| 日本午夜av视频| 亚洲av综合色区一区| 这个男人来自地球电影免费观看 | 少妇人妻 视频| 如何舔出高潮| 亚洲av成人精品一二三区| 精品99又大又爽又粗少妇毛片| 日本色播在线视频| 在线亚洲精品国产二区图片欧美 | 午夜福利视频在线观看免费| av又黄又爽大尺度在线免费看| 又黄又爽又刺激的免费视频.| 成人国语在线视频| 免费少妇av软件| 精品午夜福利在线看| 久久久久久久亚洲中文字幕| 美女国产视频在线观看| 老女人水多毛片| 亚洲综合色惰| 亚洲欧美色中文字幕在线| 各种免费的搞黄视频| 精品亚洲成国产av| 久久久久视频综合| av专区在线播放| 国产亚洲午夜精品一区二区久久| 国产又色又爽无遮挡免| 国产综合精华液| 伦精品一区二区三区| 欧美日韩视频高清一区二区三区二| kizo精华| 国产成人av激情在线播放 | 男女高潮啪啪啪动态图| 欧美老熟妇乱子伦牲交| 色视频在线一区二区三区| 一个人看视频在线观看www免费| 黑人欧美特级aaaaaa片| 欧美激情极品国产一区二区三区 | 久久精品国产鲁丝片午夜精品| 亚洲伊人久久精品综合| 老女人水多毛片| 日韩中文字幕视频在线看片| 99久久精品一区二区三区| 亚洲内射少妇av| 久热这里只有精品99| 精品久久久久久久久av| 国产成人午夜福利电影在线观看| 久久综合国产亚洲精品| 精品酒店卫生间| 男男h啪啪无遮挡| 女性生殖器流出的白浆| 插阴视频在线观看视频| 三级国产精品欧美在线观看| 51国产日韩欧美| 汤姆久久久久久久影院中文字幕| 国产69精品久久久久777片| 午夜av观看不卡| 中文字幕制服av| 日本wwww免费看| 伊人久久国产一区二区| 欧美精品国产亚洲| 国产不卡av网站在线观看| 久久人人爽av亚洲精品天堂| 国产欧美亚洲国产| 91久久精品国产一区二区成人| 欧美老熟妇乱子伦牲交| 亚洲av福利一区| 韩国高清视频一区二区三区| 久久久久国产网址| 国产精品一区二区在线不卡| 美女中出高潮动态图| 在线观看三级黄色| 精品亚洲乱码少妇综合久久| 久久精品久久久久久噜噜老黄| 三级国产精品片| 日韩人妻高清精品专区| 一本大道久久a久久精品| 久久国产精品大桥未久av| 在线观看免费高清a一片| 中国美白少妇内射xxxbb| 一边亲一边摸免费视频| videos熟女内射| 亚洲美女视频黄频| 热re99久久精品国产66热6| 国产永久视频网站| 色婷婷久久久亚洲欧美| 美女脱内裤让男人舔精品视频| 一二三四中文在线观看免费高清| 久久久欧美国产精品| 一本—道久久a久久精品蜜桃钙片| 日韩成人av中文字幕在线观看| 久久久久久久亚洲中文字幕| 久久精品国产自在天天线| 少妇人妻精品综合一区二区| 欧美三级亚洲精品| 国产高清有码在线观看视频| 国产色婷婷99| 亚洲国产av新网站| 亚洲五月色婷婷综合| 国产av码专区亚洲av| 国产av国产精品国产| 午夜视频国产福利| 国产男人的电影天堂91| 99久久人妻综合| 制服诱惑二区| 少妇人妻精品综合一区二区| 欧美精品一区二区大全| 亚洲五月色婷婷综合| 69精品国产乱码久久久| 精品亚洲乱码少妇综合久久| 又粗又硬又长又爽又黄的视频| 97超碰精品成人国产| 亚洲,一卡二卡三卡| 夫妻性生交免费视频一级片| 免费观看的影片在线观看| a级毛片在线看网站| 蜜桃国产av成人99| 亚洲国产av影院在线观看| 国产色婷婷99| 亚洲av电影在线观看一区二区三区| 国产在线免费精品| 久久国内精品自在自线图片| 国产永久视频网站| 欧美亚洲日本最大视频资源| 亚洲综合色惰| 99久久精品一区二区三区| 在线观看美女被高潮喷水网站| 亚洲成色77777| 成年人午夜在线观看视频| 2022亚洲国产成人精品| 国产爽快片一区二区三区| 国产欧美另类精品又又久久亚洲欧美| 久久久欧美国产精品| 亚洲精品av麻豆狂野| 日本-黄色视频高清免费观看| 亚洲欧美日韩另类电影网站| 国产欧美日韩综合在线一区二区| 国产色爽女视频免费观看| a级毛片免费高清观看在线播放| 欧美人与善性xxx| 欧美日韩国产mv在线观看视频| 伊人亚洲综合成人网| 免费看av在线观看网站| 亚洲综合色惰| 考比视频在线观看| 久久综合国产亚洲精品| 亚洲成人av在线免费| 最近2019中文字幕mv第一页| 男人操女人黄网站| av在线老鸭窝| 亚洲av电影在线观看一区二区三区| 国产免费现黄频在线看| 丝袜脚勾引网站| 国产 一区精品| 国产色爽女视频免费观看| 国产精品一区二区在线不卡| 亚洲成色77777| 午夜激情av网站| 欧美日韩精品成人综合77777| 又黄又爽又刺激的免费视频.| 久久午夜综合久久蜜桃| .国产精品久久| 欧美精品亚洲一区二区| 26uuu在线亚洲综合色| 欧美97在线视频| 国产成人av激情在线播放 | 校园人妻丝袜中文字幕| 免费观看av网站的网址| 丰满饥渴人妻一区二区三| 久久久久久伊人网av| 午夜免费男女啪啪视频观看| 亚洲精品av麻豆狂野| 国产免费一级a男人的天堂| 国产精品久久久久久精品古装| 一本—道久久a久久精品蜜桃钙片| 日韩在线高清观看一区二区三区| 久久午夜福利片| 亚洲精品视频女| 亚洲不卡免费看| 黑人猛操日本美女一级片| 成人免费观看视频高清| 水蜜桃什么品种好| 最新中文字幕久久久久| 一级毛片aaaaaa免费看小| 啦啦啦中文免费视频观看日本| 永久网站在线| 亚洲成人一二三区av| 国产欧美另类精品又又久久亚洲欧美| 亚洲,欧美,日韩| 欧美 亚洲 国产 日韩一| 欧美日韩成人在线一区二区| 精品亚洲成a人片在线观看| 中国国产av一级| 国产有黄有色有爽视频| 天堂俺去俺来也www色官网| 成人国语在线视频| 99re6热这里在线精品视频| 国产精品熟女久久久久浪| 哪个播放器可以免费观看大片| 成人综合一区亚洲| 最近中文字幕高清免费大全6| 黑丝袜美女国产一区| 国产乱来视频区| 日韩大片免费观看网站| 亚洲久久久国产精品| 五月天丁香电影| 欧美bdsm另类| 在线观看免费视频网站a站| 国产成人a∨麻豆精品| 日韩av在线免费看完整版不卡| 亚洲精品一区蜜桃| 欧美性感艳星| 国产免费福利视频在线观看| 在线亚洲精品国产二区图片欧美 | www.av在线官网国产| 亚洲国产色片| 国产精品秋霞免费鲁丝片| 日韩在线高清观看一区二区三区| 亚洲精品国产av成人精品| 国产精品一国产av| 精品久久久噜噜| 日日摸夜夜添夜夜爱| 视频在线观看一区二区三区| 国产乱人偷精品视频| 亚洲av电影在线观看一区二区三区| 狠狠婷婷综合久久久久久88av| 欧美成人精品欧美一级黄| 亚洲欧洲日产国产| 大陆偷拍与自拍| 婷婷色av中文字幕| 欧美精品国产亚洲| 亚洲经典国产精华液单| 一级黄片播放器| 国产成人免费观看mmmm| 99国产精品免费福利视频| 国产精品熟女久久久久浪| 久久久久精品久久久久真实原创| 少妇被粗大的猛进出69影院 | 免费黄色在线免费观看| 国产成人91sexporn| 欧美人与善性xxx| 简卡轻食公司| 亚洲av.av天堂| 国产一区二区三区综合在线观看 | 欧美变态另类bdsm刘玥| 最近手机中文字幕大全| 国产 一区精品| 午夜视频国产福利| 一本大道久久a久久精品| 看非洲黑人一级黄片| 国产不卡av网站在线观看| 热99国产精品久久久久久7| 少妇被粗大猛烈的视频| 国产不卡av网站在线观看| 一本一本综合久久| 一区二区av电影网| 人妻一区二区av| 国产老妇伦熟女老妇高清| 亚洲美女视频黄频| 欧美人与善性xxx| 水蜜桃什么品种好| 18禁裸乳无遮挡动漫免费视频| 精品酒店卫生间| 久久99热6这里只有精品| 少妇的逼好多水| 欧美最新免费一区二区三区| 国产亚洲一区二区精品| 精品久久久久久久久av|