• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    LCF:A Deep Learning-Based Lightweight CSI Feedback Scheme for MIMO Networks

    2022-08-23 02:20:34KyuhaengLee
    Computers Materials&Continua 2022年6期

    Kyu-haeng Lee

    Dankook University,Yongin-si,Gyeonggi-do,16890,Korea

    Abstract:Recently,as deep learning technologies have received much attention for their great potential in extracting the principal components of data,there have been many efforts to apply them to the Channel State Information(CSI)feedback overhead problem,which can significantly limit Multi-Input Multi-Output(MIMO)beamforming gains.Unfortunately,since most compression models can quickly become outdated due to channel variation,timely model updates are essential for reflecting the current channel conditions,resulting in frequent additional transmissions for model sharing between transceivers.In particular, the heavy network models employed by most previous studies to achieve high compression gains exacerbate the impact of the overhead,eventually cancelling out the benefits of deep learning-based CSI compression.To address these issues,in this paper,we propose Lightweight CSI Feedback(LCF), a new lightweight CSI feedback scheme.LCF fully utilizes autoregressive Long Short-Term Memory(LSTM)to generate CSI predictions and uses them to train the autoencoder, so that the compression model could work effectively even in highly dynamic wireless channels.In addition, 3D convolutional layers are directly adopted in the autoencoder to capture diverse types of channel correlations in three dimensions.Extensive experiments show that LCF achieves a lower CSI compression error in terms of the Mean Squared Error (MSE), using only about 10% of the overhead of existing approaches.

    Keywords:CSI;MIMO;autoencoder

    1 Introduction

    Wireless communication systems have significantly benefited from utilizing Channel State Information(CSI)at the transmitter.As one indicator of CSI,the Signal-to-Interference-plus-Noise Ratio(SINR) has been used to enable intelligent transmission functionalities such as dynamic data rate adaptation, admission control, and load balancing since the early days of wireless communication.With the advent of the Multi-Input Multi-Output (MIMO) method, which has now become a core underlying technology in most current systems,such as 5G and Wi-Fi[1,2],the importance of CSI at the transmitter has been highlighted,since proper MIMO beamforming weights can be calculated only through CSI values that accurately reflect the attenuation of the actual channel between a transmitter and a receiver.For this reason,both cellular and Wi-Fi systems already operate their own CSI feedback protocols to allow the transmitter to acquire channel information a priori.

    A CSI feedback process is essential for modern communication systems, yet it faces a critical overhead issue that could greatly limit the potential of MIMO beamforming gains.The amount of CSI that needs to be sent to the transmitter is basically proportional to the number of transmitting and receiving antennas;in Orthogonal Frequency Division Multiplexing(OFDM)systems,the number of subchannels also contributes to an increase in the feedback size, since CSI for every subchannel is required for OFDM transmission.Moreover,for reliable CSI feedback,CSI is typically transmitted at low data rates(e.g.,6.5 Mbps over Wi-Fi),which further exacerbates the impact of the overhead.According to a study,the overhead can reach up to 25×the data transmission for a 4×4 160 MHz Wi-Fi MIMO channel[3],and such substantial overhead will not only limit the network capacity but also prevent the realization of advanced MIMO functionalities,such as massive MIMO,distributed MIMO,and optimal user selection[4–7].

    Thus far,numerous schemes have been proposed to address the CSI feedback overhead problem.One widely accepted idea is to exploit the diverse types of channel correlation that could be readily observed in the temporal,frequency,and spatial domains.The channel coherence time has been used to eliminate unnecessary CSI feedback transmissions in many studies[3,8–10],and similarly adaptive subcarrier grouping is employed to reduce the feedback size in OFDM systems [3,10].A rich body of literature focuses on utilizing the spatial channel correlation used in multi-antenna communication for the same purpose[10–12].Recently,as deep learning technologies have received attention for their superior ability to extract the principal components of data, there have been many efforts to apply them to CSI compression [13–18] and estimation [19–23].In particular, in this field, autoencoders are commonly employed: A receiver compresses CSI data with the encoder of an autoencoder, and the transmitter reconstructs the original CSI using the decoder.A novel CSI compression scheme,CSINet[13],is built on a convolutional autoencoder,where the authors regard the CSI compression problem as a typical 2D image compression task.Along this line,numerous variants of CSINet have been developed for various purposes[14–17].

    Although the aforementioned approaches show that deep learning can be used as a very effective tool for CSI compression, there still remain several critical issues to be solved in terms of how the transceivers can practically share the models.Neural network-based CSI compression schemes are basically premised on sharing a model between a transmitter and a receiver,which means that some transmissions for this model sharing,and their accompanying cost,are unavoidable.In this paper,we refer to this as model sharing overhead.Unfortunately,this overhead has not been thoroughly taken into account in most existing studies, and in many cases, it is assumed that the transceivers already share a model or that model sharing will rarely happen.However,as we will see later,model sharing can occur quite often in practice,since the model cannot guarantee a high degree of generalization to wireless channel data.If the model cannot properly cope with CSI values that it has not experienced during training, then the compression and recovery will fail, which leads to an inevitable process of model re-training and sharing.Of course,in some channel environments where a clear pattern is found,as shown in Fig.1a,the overhead problem may not be so serious.However,this situation cannot be always guaranteed; the actual channels may look more like the one in Fig.1b.In addition, due to the strong randomness of the change in the wireless channel state, simply increasing the amount of training data does not result in noticeable generalization enhancement.Rather,it is more important to use a proper set of training data that reflects the pattern and tendency of the current channel status well,and for this,an appropriate channel status prediction can be of great help.

    Figure 1:Examples of channel coefficient changes over time for two scenarios.Only the real parts of the complex channel coefficients of the first path between the first transmitting antenna and three receiving antennas are displayed.(a)Stable channel case(b)Dynamic channel case

    As mentioned earlier,most of the previous approaches focus only on making the model work at higher compression ratios,and thus they prefer deep and wide networks in their designs.For example,CSINet [13] uses five convolutional layers and two fully connected layers, and it successfully compresses CSI with a high compression ratio of up to 1/64,which obviously outperforms conventional compressed sensing-based approaches[24,25].However,from the model sharing perspective,the model is still too big to share;roughly calculated,for an 8×2 MIMO channel with 8 paths,it needs more than 1,000 decoder parameters in total,which actually makes it larger than the original CSI(i.e.,256=8×8×2×2,where the last number denotes the real and imaginary parts of complex channel coefficients).In this case,model sharing is practically not available since it consequently cancels out the benefits of compression.

    In order to overcome these limitations, in this paper, we propose a lightweight CSI feedback scheme (LCF).Similar to recent approaches, LCF exploits deep neural networks to achieve better CSI compression, but we focus more on ensuring that the model does not impose a substantial burden on the network when being shared.LCF mainly consists of two parts:CSI prediction and CSI compression.First,for CSI prediction,LCF employs a long short-term memory(LSTM)structure to infer future CSI values,which are in turn used to train the CSI compression model.In particular,to generate multiple future CSI predictions effectively,we apply an autoregressive approach to our model.The actual channel compression and reconstruction is conducted by a convolutional autoencoder,where 3D convolution layers are adopted to capture channel correlations in the three dimensions of the transmitting antenna,receiving antenna,and path delays.The resulting compression model appears to be simple compared to recent proposals;however,we will show that this lightweight network structure is still sufficient for achieving high CSI compression performance,with a much cheaper model sharing overhead.

    The proposed CSI feedback scheme is developed and evaluated in a TensorFlow[26]deep learning framework.In order to investigate the performance of LCF for various channel environments, we simulate channel coefficients by applying the WINNER II channel model[27,28].We provide microbenchmarks that evaluate the performance of each CSI prediction and compression process,and we compare the overall performance of LCF with those of AFC [3] and CSINet [13] in terms of CSI recovery accuracy and model sharing overhead.Through extensive experiments, we show that LCF obtains more stable and better CSI compression performance in terms of the mean squared error(MSE),with as low as 10%of the model sharing overhead of the existing approaches.We summarize the main contributions of this paper as follows:

    1.We propose a novel deep learning-based CSI feedback scheme,LCF,which effectively reduces the CSI feedback overhead by using CSI prediction based on autoregressive LSTM and CSI compression with a convolutional autoencoder.We propose the use of CSI predictions to train the autoencoder,so that the compression model can be valid even in highly dynamic wireless channels.

    2.We design a CSI feedback algorithm to make the transmitter and the receiver effectively share the compression model,which has not been investigated well in previous studies.The proposed algorithm can be applied to the existing deep learning-based CSI compression approaches as well.

    3.The performance of LCF is evaluated for various wireless channel scenarios using the WINNER II model,and it is also compared with those of other approaches through extensive experiments.LCF shows more stable and better CSI compression performance, using only about 10%of the overhead of existing approaches.

    The rest of this paper is organized as follows.In Section 2,we review the previous works related to this paper.In Section 3, we provide the preliminaries of this work, and Section 4 describes LCF in detail.Section 5 evaluates the performance of the proposed scheme,and we conclude this paper in Section 6.

    2 Related Work

    Numerous sc hemes have been proposed to address the CSI feedback overhead problem using diverse types of channel correlations.The channel coherence time, during which the channel state remains highly correlated, has been used as a key metric for eliminating unnecessary CSI feedback transmissions in many studies [3,8–10,12].Huang et al.[8] analyze the effect of time-domain compression, based on a theoretical model of channel correlation over time.Sun et al.[9] simulate the 802.11n single-user MIMO(SU-MIMO)performance in time-varying and frequency-selective channel conditions.AFC [3] computes the expected SINR by comparing the previous and the current CSI values and then utilizes it to determine whether to skip a CSI feedback transmission or not.

    Similar ideas can be applied to compressing the frequency domain CSI values.Since in OFDM systems, the channel estimation should be performed on each subcarrier, appropriate subcarrier grouping can reduce the feedback size significantly.In MIMO systems,spatial correlation could also be used for CSI compression.Gao et al.[10]design a channel estimation scheme for an MIMO-OFDM channel using both temporal and spatial correlations.Ozdemir et al.[11] analyze the parameters affecting spatial correlation and its effect on MIMO systems, and Karabulut et al.[12] investigate the spatial and temporal channel characteristics of 5G channel models, considering various user mobility scenarios.These schemes can be further improved with proper quantization schemes that encode the original CSI data with a smaller number of bits.AFC[3]employs an adaptive quantization scheme on top of the integrated time and frequency domain compression,and CQNET[14]is designed for optimizing codeword quantization using deep learning for massive MIMO wireless transceivers.Among other things, it is actually being used for codebook-based CSI reporting in current cellular and Wi-Fi systems[1,2].

    Recently,as deep learning technologies have received attention for their powerful performance in extracting the principal components of data,there have been many efforts to use this capability for CSI compression[13–18]and estimation[19–23].The autoencoder model is widely used in this field since it best fits the problem context.A novel CSI compression scheme,CSINet[13],uses a convolutional autoencoder to solve the CSI compression problem by turning it into a typical 2D image compression problem.Along this line,numerous variants of CSINet have been developed so far[14–18].RecCsiNet[15] and CSINET-LSTM [16] incorporate LSTM structures into the existing autoencoder model to benefit from the temporal and frequency correlations of wireless channels.The authors of PRVNet[17]employ a variational autoencoder to create generative models for wireless channels.In DUalNet[18],the channel reciprocity is utilized for CSI feedback in FDD scenarios.Most of these approaches validate the feasibility of deep learning as an effective tool for CSI compression and feedback;however,there remain several practical open issues related to model sharing and generalization,which will be discussed in the following section.

    3 Preliminaries

    In this section,we describe the channel model and propagation scenarios used in this paper,and we explain the model sharing overhead problem,which motivates this work,in greater detail.

    3.1 Channel Model and Propagation Scenarios

    We consider an SU-MIMO communication scenario in which a receiver equipped withNrantennas feeds the estimated CSI back to its transmitter,which is equipped withNtantennas.Uniform Linear Array(ULA)antennas with 2 cm spacing are assumed for both the transmitter and the receiver.For simplicity,moving network scenarios are not considered.In order to simulate channel coefficients for diverse channel environments,we adopt the WINNER II channel model[27–31],which has been widely used in wireless communication research activities; it was recommended as a baseline for measuring radio communication performance in ITU-R (International Telecommunication Union-Radio communication sector)[29,30].According to this model,the channel coefficients are generated based on a Clustered Delay Line(CDL)model(Fig.2),where the propagation channel is described as being composed of a number of separate clusters with different rays,and each cluster has a number of multipath components that have the same delay values but differ in the Angle-of-Departure(AoD)and Angle-of-Arrival(AoA).

    In this paper, we consider two different propagation scenarios, namely “stable”and “dynamic”scenarios, which model indoor office environments and bad city macro cells, respectively.Here, the former,as the name suggests,has a smaller channel status variation than the latter.Tab.1 shows the basic statistics of the two channels.We use MATLAB to gather channel coefficient data for each case,and CSI data is sampled every 2 ms at the center frequency of 5.25 GHz.Note that MATLAB provides a toolbox for working with the WINNER model [27], which allows us to freely customize network configuration such as sampling rate, center frequency, number of base stations and mobile stations,and their geometry and location information.In particular, since the main channel parameters for various radio propagation scenarios defined by the WINNER model are already configured, the corresponding channel coefficient values can be easily obtained through this.For each scenario,channel coefficient data is expressed as a four-dimensional normalized complex matrix whose shape is (Ns×Nd×Nt×Nr), whereNsis the length of the sampled data andNdis the number of path delays.Throughout this paper,we use the terms“CSI”and“channel coefficients”interchangeably.

    Figure 2: Concept of the Clustered Delay Line (CDL) model [28].Each cluster has a number of multipath components that have the same delay values but differ in the angle-of-departure and angleof-arrival

    Table 1: Statistics of the two channel models

    To show the difference between the two scenarios,we plot the changes over time of the channel coefficients for each scenario in Fig.1.These channel coefficients are the values corresponding to the first path of receiving antennas 1–3 and transmitting antenna 1,and only the real parts of these values are displayed in the plot.From the figure,we can observe the spatial and temporal correlation of the channels for both scenarios,even though they differ in degree.In the case of the stable channel scenario(Fig.1a),similar signal patterns are repeated quite clearly over time and also among the three receiving antennas.In the case of the dynamic channel scenario(Fig.1b),it is difficult to find a clear pattern like that in the previous case,yet we can still see correlations in the two domains.

    We can see the difference between the two channels in terms of correlation more clearly in Fig.3.In this figure,we measure the correlation coefficient of any two CSI instances separated byT,using the following formula[3,32]:

    whereLis the total length of CSI instances andh(t)is the CSI instance at timet.Note that the above equation can be also applied to computing the correlation in the spatial domain (Fig.3b) by changing the definition of the separation.As expected,the stable scenario has overall higher temporal correlations than the dynamic channel,as shown in Fig.3a;their coherence times1Channel coherence time is defined as the point when the correlation value drops to 0.5[33].are 420 and 40 ms,respectively.Compared to the temporal correlation result,higher spatial channel correlations among receiving antennas are observed in both cases(Fig.3b),though the degree of correlation in the stable channel is still higher than that in the dynamic channel.

    Figure 3: Temporal and spatial correlation of the two channels used in this paper.(a) Temporal correlation(b)Spatial correlation

    3.2 Model Sharing Overhead

    As we saw earlier,wireless channels are basically diverse,and their characteristics are thus hard to generalize; some channels remain highly correlated over time for long periods of time, e.g., the stable channel,while others may experience large channel fluctuations,e.g.,the dynamic channel.This aspect, unfortunately, has not been fully taken into account in most of the previous deep learningbased CSI compression schemes, although this leads to a substantial model sharing overhead that eventually limits the gains of deep learning.To identify the model sharing overhead in more detail,we revisit the CSI compression performance of CSINet[13]for three different channel environments,including the two channels described in the previous subsection.As a baseline,we additionally consider a purely random channel, where the channel coefficients are sampled from the normal distribution with zero mean and a variance of 0.1.Note that the stable channel used in this paper belongs to an ideal case in which we can readily predict how the channel changes in the future, while the random channel can be viewed as being at the other end of the spectrum.The mean squared error(MSE)in Section 5 is used as a performance metric,and we train the model using Adam optimization[34]with 10,000 CSI datasets and a maximum of 1,000 epochs.

    First,Fig.4a shows the compression performance according to varying compression ratios.What we pay attention to here is the performance in the dynamic channel:It deteriorates rapidly with the compression ratio,and its MSE value reaches 0.1 even at the low compression ratio of 1/8,which is too large to be used practically.Throughout this paper,the compression failure criterion,denoted asδthr,is set to the MSE of 0.1,and considering that for any wireless channels,the performance of CSINet will be somewhere between the two curves of the stable(i.e.,blue curve)and the random(i.e.,black dotted curve)channels,we can conclude that it is practically available only with the compression ratio of 1/8.When compression fails, the model has to be retrained and shared between the transceivers again,which eventually causes additional transmissions,i.e.,model sharing overhead.Unfortunately,most of the previous approaches focus only on making the model work at higher compression ratios,and thus they prefer deep and wide networks in their designs,hoping that model sharing will not occur frequently.However,as can be seen,it could occur very frequently.In the case of the dynamic channel,at compression ratios greater than 1/16, for every CSI sample the model will always produce a high recovery error above the threshold;in this case,its heavy network structure will accelerate the feedback overhead increase.

    Figure 4:Performance of an existing deep learning-based CSI compression scheme for various channel environments.(a)MSE vs.Compression ratio(b)MSE vs.Time(c)MSE vs.Amount of training data

    Therefore,we have to consider proper model sharing when designing a deep learning-based CSI compression scheme.Fig.4 shows the result when 50 consecutive CSI values are compressed with a fixed model.If the channel is quite reliable, such as in the stable channel case, we might keep the high compression gains of deep learning, but this is not always guaranteed; as can be seen in the dynamic channel case, high recovery errors could continue over time, thus leading to additional feedback transmissions.One might think that this problem can be solved through training the model with more data and thus strengthening the model generalization.Unfortunately, this approach may not be very effective when dealing with wireless channel data,which generally have large irregularities over time.As shown in Fig.4,even if we increase the number of CSI training data samples,there are only slight performance gains.In this case, the amount of training data may not be that important;rather,it is more important to use a proper set of training data that reflects the pattern and tendency of the current channel status well,and for this,appropriate channel status predictions could be very helpful.

    We describe the proposed CSI feedback scheme in detail in the following section.

    4 LCF

    4.1 Overview

    In this section, we provide an overview of the proposed scheme.As mentioned before, LCF is composed of two main processes, as shown in Fig.5, for which different deep learning models are used:1)autoregressive LSTM is used for CSI prediction,and 2)a convolutional autoencoder is used for CSI compression and reconstruction.In the first process,as the name suggests,a receiver generates predictions for future channel states using accumulated CSI values,which in turn will be used as the training dataset for autoencoder optimization.This process is described in detail in Section 4.2.Next,in the second step,the actual CSI compression and recovery is performed.Using the encoder of the autoencoder,a receiver compresses the estimated CSI into anM-byte codeword and then sends it back to the transmitter.Upon receiving the compressed CSI,the transmitter reconstructs the original CSI with the decoder of the autoencoder,which has been already shared by the receiver.More details on the compression model will be described in Section 4.3.Ideally,the feedback process of LCF requires only anM-byte data transmission,and therefore the feedback overhead can be significantly reduced.However, as mentioned earlier, such a gain is not always achievable; in some channel environments,the compression model could quickly become less effective and invalid.To tackle this issue,in LCF,a receiver dynamically updates and shares the compression model depending on the expected CSI recovery error obtained during the CSI compression.We explain this in more detail in Section 4.4.

    Figure 5: Overall structure of LCF.It consists of an LSTM-based CSI prediction process and an autoencoder-based CSI compression and reconstruction process.Note that the modules in the dotted box are executed as needed

    4.2 CSI Prediction Using Autoregressive LSTM

    Figure 6:Autoregressive LSTM for CSI prediction.CSI predictions are made for each combination of the path delay,transmitting antenna,and receiving antenna

    The proposed prediction model has two layers, an LSTM layer and a fully connected layer, as shown in Fig.6.CSI predictions are made for each combination of the path delay, transmitting antenna, and receiving antenna.Additionally, we handle the real and imaginary parts of complex channel coefficients separately,since complex numbers are inherently not comparable in size and thus cannot be directly used in optimization.That is, we use 2Nd Nt Nrmodels in total, and each model is used to generate CSI predictions for the corresponding combination.The input data shape for the models is(batch×Ni× 1),where the last number indicates the number of units(features)in the fully connected layer.One distinct feature of CSI data is that input data samples keep arriving sequentially to the model, and relatively old data samples can quickly become less effective.In this case, we can use online learning;instead of always training on the entire data set,i.e.,in most cases,training is performed only on a small data set containing new data samples,i.e.,In particular,the weights obtained in the previous step can be reused for performance.We basically use the MSE as the objective for optimization and employ Adam optimization[34].

    For better understanding, we illustrate an example of CSI prediction in Fig.7.The channel coefficients shown in the figure are sampled from the channel of the first path between transmitting antenna 1 and receiving antenna 1 in the stable channel case.We depict two curves for both the real and imaginary parts of the channel coefficients in the figure.This example corresponds to the case ofNi=No= 20,which means the channel coefficients from index 10 to index 29 are used to generate the next 20 CSI values from index 30 to index 49.As expected,the prediction performance is basically dependent on the previous prediction results;errors are continuously accumulated as predictions are made,and thus the prediction accuracy gradually drops as time advances.Starting from the MSE of 0.0001,the gap between the actual data and the prediction becomes apparent continuously,and at the last position,it grows to up to 0.0032.However,we note that this level of error is acceptable for the training data,as we will see later.

    Figure 7:CSI prediction example for the channel of the first path between transmitting antenna 1 and receiving antenna 1

    4.3 Convolutional Autoencoder-Based CSI Compression and Reconstruction

    The proposed CSI compression model has the typical structure of an autoencoder, as shown in Fig.8.It consists of two parts: An encoder and a decoder.Letf encandf decbe the encoder and the decoder,respectively.The encoder takes the current CSI(i.e.,Hi)as input,which is a four-dimensional channel coefficient matrix whose shape is (Nd×Nt×Nr× 2), where the last element denotes the real and the imaginary parts.The first layer in the encoder is a 3D convolution layer, where threedimensional filters are used to capture the channel correlation in both the spatial (for both the transmitting and receiving antennas)and delay domains.By default,we use 16(3×3×3)filters,and the LeakyReLU activation function with a parameter of 0.3 is applied.Stripping is not used.The feature maps acquired from this layer are then transferred to a fully connected layer withMunits through average downsampling with a shape of(2×2×2)and flattening,and thus theM-byte compressed CSI, denoted asHM, can be obtained as a result, i.e.,HM=f enc(Hi;θenc), whereθencis a set of the encoder parameters.

    The decoder is basically the mirror of the encoder.It first passes the encoded data(i.e.,HM)to a fully connected layer withNfunits,whereNfis the size of the output of the convolution layer in the encoder, and then transfers the outcome to a convolution layer through the upsampling layer.Like the convolution layer in the encoder,the convolution layer in the decoder also takes three-dimensional filters;however,we employ a transposed convolution layer in the decoder to match the input shape and the output shape,i.e.,(Nc×Nt×Nr×2).The same LeakyReLU activation function is applied,and L2 regularizers with a parameter of 0.001 are applied to all layers in the model.As a result,the decoder reconstructs the original CSI data from the compressed data,HM;that is, ?H=f dec(f enc(Hi;θenc);θdec)=f dec(HM;θdec),whereθdecis the decoder parameters.In the following subsection,we will explain how the model parameters(θencandθdec)are trained and shared between transceivers.

    Figure 8:The proposed CSI compression and reconstruction model.3D convolution layers are adopted in both the encoder and decoder

    4.4 Model Training and Sharing

    In LCF, the whole parameters of the autoencoder (bothθencandθdec) are trained by a receiver,but only the decoder parameters(θdec)are sent to the transmitter if needed.At the very beginning,the receiver trains the autoencoder with(No+1)CSI values,including the current CSI,i.e.,Hi,and the newly generatedNoCSI predictions,i.e.,nd through this step,it obtains the trained encoder and decoder parameters,respectively.Since the decoder model is updated,its parameters need to be sent to the transmitter by the receiver.Now,the process of compressing the target CSI is conducted by retraining the model on it.Note that in this step,training is performed with the decoder parameters fixed,since the decoder parameters are already shared with the transmitter.In this process,θencparameters are still trainable,so they may have different values before and after training.However,since they are not shared with the transmitter and are used only by the receiver,they do not have a significant impact on the system as a whole.

    Every time the receiver compresses CSI,as a result of training,it obtains an optimization error,denoted asδ,which corresponds to the expected CSI recovery error at the transmitter.Depending on this value, it makes a decision about whether to send the decoder parameters to the transmitter or not.If aδvalue is less than a predefined threshold, i.e.,δthr, the receiver sends only the compressed CSI,i.e.,HM,as this implies that the decoder parameters are still valid enough for the current target CSI thanks to the CSI prediction.Otherwise,the receiver obsoletes the previous decoder parameters;then, it re-trains the entire model and sends the newly trained decoder parameters (i.e.,θdec) to the transmitter with the compressed CSI.

    Algorithm 1:Receiverimages/BZ_1364_265_2386_1765_2868.png

    Algorithm 2:Transmitterimages/BZ_1365_264_444_1764_689.png

    To summarize, we provide the entire proposed CSI feedback algorithm in Algorithms 1 and 2.Basically, the proposed method is designed for two communication entities, but it can be extended to multi-user scenarios as well.However,in this case,the transmitter may have to maintain different models for different users, which can cause additional operational burdens such as increased model sharing overhead.Therefore,we should consider mixing the proposed deep learning-based approach with traditional approaches.We leave this issue for our future work.

    5 Performance Evaluation

    5.1 Settings

    In this section, the performance of LCF is evaluated.We use TensorFlow 2 [26] to develop the proposed deep learning models of LCF and conduct extensive experiments on an Intel-i7 machine with 16 GB RAM and an NVIDIA RTX 3080 GPU.Using the MATLAB WINNER II model[27],we generate CSI datasets for both scenarios,which are sampled every 2 ms.When training the models,we use 70%of the total dataset for training,and the other 20%and 10%are used for validating and testing,respectively.For model optimization,we use the MSE as the objective,and employ Adam optimization[34]with a maximum of 1,000 epochs and a learning rate of 0.001.The default parameters used in the experiments are described in Tab.2.

    Table 2: Default parameters

    To compare the performance of LCF with those of other approaches,we additionally implement AFC[3]and CSINet[13].Unfortunately,since all these schemes have different features and feedback policies,we have to make some modifications to them to ensure a fair comparison.The main changes are as follows:

    · AFC: As AFC is not a machine learning-based approach, it does not require a training step and determines the degree of compression by calculating the expected compression error each time it receives CSI.The adaptive bit-quantization scheme is excluded since it can be applied to the other schemes as well.In the original AFC,the compression ratio can also be dynamically adjusted depending on the channel status,which is different from the other two schemes,which use a fixed compression ratio.In this study,for simplicity,we apply a fixed compression ratio to AFC.

    · CSINet: CSINet considers only single-antenna receiver cases.In order to extend it to multiantenna receiver cases,we repeatedly apply it to the channel of each receiving antenna.We use the same training configuration(both dataset and optimizer)for both CSINet and LCF.

    · Both: All these schemes can skip a CSI transmission or model (i.e., decoder) parameter transmission if the expected CSI recovery error is less thanδthr.Note that even if this condition is satisfied,LCF and CSINet should still send the compressed CSI to the transmitter.

    Compression ratioαis defined as the ratio of the compressed data size to the original CSI data size(i.e.,),and as a key performance metric,we measure the MSE,defined in the following equation [13]: MSE =.In addition to the MSE, we use the cosine similarity between the original CSI and the reconstructed CSI to determine the value ofδthr.The imperfect CSI due to the compression causes changes in the resulting beam steering vectors,which can be measured as the cosine similarity between the two CSI values[3,13].Fig.9 shows the cosine similarity values as a function of MSE forNt=Nr= 4 andNd= 16.To draw this plot,for each MSE value,we generate two sets of CSI matrices:One is randomly generated from the standard normal distribution,and the other is generated by adding random noise of the given MSE value to the previous matrix.After that,we compute the cosine similarity between the two matrices for each MSE value.The result is quite predictable;the cosine similarity decreases with the value of MSE.Based on this result,we setδthras 0.1,where the cosine similarity is around 0.95.

    In the following subsections, we first investigate the performance of each model used in LCF through micro-benchmarks,and then we compare the overall performance of LCF with those of the other approaches.

    5.2 Micro-Benchmarks

    5.2.1 CSI Prediction Model

    We investigate the impact of LSTM parameters on the prediction performance, according to varying numbers of LSTM units and differentNiandNocombinations.Fig.10 illustrates the plots ofNo=Nicases for each scenario.It can be seen that the prediction accuracy decreases with the value ofNo,except for the case whereNo= 5 and 256 units are used in the dynamic channel.This result is consistent with the previous observation shown in Fig.7,where,as the model predicts CSI values for times farther in the future,the prediction error becomes larger.

    Figure 9:The same cosine similarity defined in CSINet[13]is used.We set the threshold for retraining as the point where the cosine similarity is around 0.95

    Figure 10:CSI prediction performance for two channel cases,according to different numbers of LSTM units and Ni and No value combinations.In most cases,the prediction accuracy becomes better with more LSTM units and smaller Ni and No values.(a)Stable channel(b)Dynamic channel

    The CSI prediction performance is mainly affected by theNovalue,but the number of LSTM units also has an effect.For all cases,the prediction becomes more accurate as the number of LSTM units increases,except for theNo=5 case for the dynamic channel.This exception is due to the overfitting problem;using more LSTM units increases the model capacity too much,making it difficult to handle unobserved CSI data.Recall that in this evaluation,the dynamic channel has a relatively high and rapid channel variation compared to the stable one, and thus it is more likely to suffer from overfitting.Overall, better CSI prediction results are observed in the stable channel case than in the dynamic channel case;the worst MSE for the stable case is around 0.02,while for the dynamic case,it reaches almost 0.9.However,by taking smallNiandNovalues,we can improve CSI predictions in the dynamic channel as well;whenNois 5 and the number of LSTM units is 32,the prediction error is at its lowest value of 0.02.

    5.2.2 CSI Compression and Recovery Model

    In this subsection, we evaluate the performance of the CSI compression model.To do this, for each scenario we first generate 20 CSI predictions withNi=No= 20 and use them to train the autoencoder.After that,we compress and restore the corresponding CSI label data with the trained model,measuring the difference between the two data.Recall that the decoder parameters are fixed once they are trained.We repeat this experiment 100 times and take the average value.

    Fig.11 shows the results in terms of the MSE and the number of decoder parameters,according to various numbers of encoder filters and compression ratios.First, from Figs.11a and 11b, we can observe that the number of filters affects the compression performance greatly.For the same compression ratio,using more filters causes lower recovery errors.Unfortunately,in return for this high performance,the use of more filters causes the model to be larger,resulting in a high model sharing overhead, as shown in Fig.11c.The number of filters is not the only factor to have an impact on the compression performance;the compression ratio affects the performance as well,since it directly affects the number of units in the fully connected layers of the autoencoder.As expected,the decoder size decreases with the compression ratio, at the expense of a high compression error.We also find that the recovery errors in the stable channel are overall lower than those in the dynamic channel,even though this difference is very subtle.

    5.3 Macro-Benchmarks

    In this evaluation,we compare the performance of LCF with those of AFC[3]and CSINet[13].We run each scheme for 20 consecutive CSI values and measure the average MSE and feedback size at different compression ratios from 1/8 to 1/128.Here,the feedback size is defined as the combined sizes of the model and the compressed CSI that are sent to the transmitter.We repeat this evaluation for both the stable and dynamic channel scenarios,and Fig.12 shows the results.

    From the results,we can see that AFC takes advantage of a small feedback size;for both scenarios,the maximum overhead is only 32,which is a much smaller number than those of other deep learningbased schemes.However, it suffers from high CSI recovery errors.As shown in Figs.12a and 12d,its MSE values are all larger than 1, which is practically implausible.However, we note that the actual AFC can perform better than this modified AFC because of its adaptive compression ratio and quantization,which have been excluded in this evaluation.Compared to AFC,CSINet and LCF both have lower MSE results.When the compression ratio is 1/8, CSINet obtains minimum MSE values of 0.07 and 0.24 for both cases, respectively.Although these values are better than those of AFC,they still seem to be somewhat unstable.In particular,when using a higher compression ratio such as 1/128,or in a highly dynamic wireless channel scenario,the CSI recovery error increases significantly.Its MSE values are at around 0.5 in the dynamic case, as shown in Fig.12d, which verifies our hypotheses that CSI compression would quickly become less effective without proper model updates.To make matters worse,this result eventually incurs a substantial model sharing overhead,as shown in Figs.12b and 12e.The reason the two curves of the two scenarios show different patterns,i.e.,one is going up while the other is going down, is because the major factor affecting the feedback size is different.In the stable channel case,model sharing rarely occurs at low compression ratios,and thus the feedback overhead decreases with the compression ratio;from Fig.12c,we can see that only 1–2 model sharing transmissions happen when the compression ratio is lower than 1/32.However,using a higher compression ratio results in more frequent model sharing,causing the decoder size to take up most of the feedback overhead;as a result,when the compression ratio is 1/128,model feedback occurs for every CSI sample(Fig.12c).Conversely,for the dynamic channel case,model sharing occurs for all CSI,as shown in Fig.12f,so the number of model transmissions no longer has a significant impact on the results.In this case,the higher the compression ratio,the smaller the model size,which at the same time reduces the feedback overhead.

    Figure 11:Impact of model parameters on the CSI compression model performance.For(a)and(b),‘FN’denotes the number of filters.In this evaluation,(3×2×3)filters are used,since the number of receiving antennas is two.(a)Recovery error(stable)(b)Recovery error(dynamic)(c)Decoder size

    Overall,LCF outperforms the other approaches in terms of MSE.Even at the highest compression ratio of 1/128,it achieves an MSE value of 0.05,which is much lower than those of the other schemes.More surprisingly, LCF obtains this result with lower feedback overhead; the average feedback overhead values are only around 40 and 120,respectively,for both cases.From Fig.12b,we can see that LCF has a higher feedback overhead than CSINet in the stable channel case when compression ratios are low (e.g., 1/8 and 1/16).This is due to the fact that LCF directly takes three-dimensional channel data as input,and thus the number of units in the fully connected layers is inherently larger than that of CSINet for the same compression ratio.As shown here, the gains of LCF may not be noticeable in these special situations,where model updates are not required much.However,in most cases, compared to the existing CSI feedback approaches, LCF obtains more stable and higher CSI compression performance,with only 10%of the model sharing overhead of the other approaches.

    Figure 12:LCF outperforms AFC and CSINet in terms of MSE and feedback overhead.Even at the highest compression ratio of 1/128,it obtains much lower MSE values with lower feedback overhead.(a)Recovery error(stable)(b)Feedback size(stable)(c)#Feedback transmissions(stable)(d)Recovery error(dynamic)(e)Feedback size(dynamic)(f)#Feedback transmissions(dynamic)

    6 Conclusion and Future Work

    In this paper, we propose LCF, which addresses the issues of conventional autoencoder-based CSI feedback schemes, specifically that CSI compression quickly becomes less effective and incurs an excessive model sharing overhead over time.Employing an autoregressive LSTM model, LCF generates CSI predictions and then exploits them to train the autoencoder, so that the compression model will be valid even for highly dynamic wireless channels.In order to fully capture the channel correlations to achieve higher CSI compression, three-dimensional convolutional layers are directly applied to the autoencoder.As a result, compared to the existing CSI feedback approaches, LCF obtains more stable and better CSI compression performance in terms of MSE,with only 10%of the model sharing overhead of the other approaches.

    The LSTM model in LCF performs properly for forecasting time-series CSI data, but unfortunately it has the well-known drawback of a long training time.Several approaches can be considered to remedy this issue.First, instead of training the prediction model on all of the data, we can use an ensemble learning strategy that would update the model with the new data, and combine it with the existing model [35,36].To overcome this limitation of LCF, we could also consider using a different type of network.Gated Recurrent Units (GRU) [37] could be one good alternative since it can take advantage of low computation with a smaller number of parameters compared to LSTM.Generally, Convolutional Neural Networks (CNNs) are computationally cheaper than the models in the Recurrent Neural Network (RNN) family, and thus they could be used for this task as well.In this case, it is easier to combine the two models of LCF, which are currently separated, into one model,resulting in higher efficiency.These schemes should be carefully considered not only with the two channel models currently used in this paper, but also with more realistic and diverse channel environments.We leave these issues for our future work.

    Funding Statement:This work was supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(No.2021R1F1A1049778).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    久久久久久久午夜电影| 亚洲av免费高清在线观看| 午夜福利在线在线| 亚洲电影在线观看av| 人人妻人人看人人澡| 天堂网av新在线| 国产视频内射| 成人av在线播放网站| а√天堂www在线а√下载| 亚洲国产欧洲综合997久久,| 国产一区二区三区av在线 | 黄片无遮挡物在线观看| 小说图片视频综合网站| 少妇熟女aⅴ在线视频| 欧美潮喷喷水| 直男gayav资源| 国产69精品久久久久777片| 久久久色成人| 性色avwww在线观看| 岛国在线免费视频观看| 91久久精品国产一区二区三区| 日韩三级伦理在线观看| 久久精品国产自在天天线| 少妇丰满av| 日产精品乱码卡一卡2卡三| 不卡一级毛片| h日本视频在线播放| 国产伦在线观看视频一区| 99热精品在线国产| 在线播放国产精品三级| 欧美成人精品欧美一级黄| 成人鲁丝片一二三区免费| 国产淫片久久久久久久久| 免费观看的影片在线观看| 岛国在线免费视频观看| 国产黄色小视频在线观看| 91久久精品国产一区二区三区| 亚洲欧洲日产国产| 亚洲内射少妇av| 九色成人免费人妻av| 黑人高潮一二区| 男女下面进入的视频免费午夜| 日本与韩国留学比较| 村上凉子中文字幕在线| 国产亚洲5aaaaa淫片| 91午夜精品亚洲一区二区三区| 久久热精品热| 日韩欧美在线乱码| 成人一区二区视频在线观看| 欧洲精品卡2卡3卡4卡5卡区| 久久久国产成人精品二区| 免费在线观看成人毛片| 乱系列少妇在线播放| 国产片特级美女逼逼视频| 精品欧美国产一区二区三| 好男人在线观看高清免费视频| 男插女下体视频免费在线播放| 亚洲五月天丁香| 91久久精品国产一区二区成人| 亚洲aⅴ乱码一区二区在线播放| av在线天堂中文字幕| 我要看日韩黄色一级片| 久久精品综合一区二区三区| 两个人视频免费观看高清| 精品久久久久久成人av| 亚洲欧美日韩卡通动漫| 亚洲美女视频黄频| 三级国产精品欧美在线观看| 国产精品不卡视频一区二区| 久久草成人影院| 在线观看午夜福利视频| 深夜精品福利| 国产亚洲5aaaaa淫片| 夫妻性生交免费视频一级片| 亚洲人成网站在线观看播放| 亚洲国产精品久久男人天堂| 99热这里只有是精品在线观看| 久久午夜亚洲精品久久| 深爱激情五月婷婷| 自拍偷自拍亚洲精品老妇| 你懂的网址亚洲精品在线观看 | 久久久国产成人免费| 偷拍熟女少妇极品色| 五月玫瑰六月丁香| www日本黄色视频网| 久久久久久九九精品二区国产| 天堂中文最新版在线下载 | 插阴视频在线观看视频| 噜噜噜噜噜久久久久久91| 久久久久久伊人网av| 亚洲欧美精品自产自拍| 色综合色国产| 日韩在线高清观看一区二区三区| 亚洲图色成人| 非洲黑人性xxxx精品又粗又长| 两个人视频免费观看高清| 国产成人福利小说| 国内揄拍国产精品人妻在线| 日本五十路高清| h日本视频在线播放| 欧美变态另类bdsm刘玥| 欧美成人a在线观看| 精品久久久久久久久亚洲| 欧美+亚洲+日韩+国产| 黄色一级大片看看| 精品国内亚洲2022精品成人| 女的被弄到高潮叫床怎么办| 麻豆成人午夜福利视频| 两个人视频免费观看高清| АⅤ资源中文在线天堂| 精品久久久久久久久亚洲| 成人毛片a级毛片在线播放| 伦理电影大哥的女人| 国产激情偷乱视频一区二区| 国内久久婷婷六月综合欲色啪| 日韩一区二区三区影片| 久久人人爽人人片av| 亚洲婷婷狠狠爱综合网| 乱人视频在线观看| 日日撸夜夜添| 亚洲精品乱码久久久久久按摩| 国产精品1区2区在线观看.| 中国美白少妇内射xxxbb| 啦啦啦观看免费观看视频高清| 别揉我奶头 嗯啊视频| 麻豆乱淫一区二区| 国产麻豆成人av免费视频| 男女视频在线观看网站免费| 狠狠狠狠99中文字幕| 一个人免费在线观看电影| 久久久久久伊人网av| 午夜a级毛片| 91午夜精品亚洲一区二区三区| 日本-黄色视频高清免费观看| 国产精品.久久久| 欧美高清成人免费视频www| 欧美一级a爱片免费观看看| 又爽又黄a免费视频| 欧洲精品卡2卡3卡4卡5卡区| 最近的中文字幕免费完整| 精品一区二区三区视频在线| 成人无遮挡网站| 久久6这里有精品| 成人漫画全彩无遮挡| 长腿黑丝高跟| 久久久久久伊人网av| av免费观看日本| av天堂在线播放| 成年女人看的毛片在线观看| 寂寞人妻少妇视频99o| 99久久久亚洲精品蜜臀av| 一进一出抽搐gif免费好疼| 人体艺术视频欧美日本| 内射极品少妇av片p| 波多野结衣巨乳人妻| 日本一本二区三区精品| 内地一区二区视频在线| 99热这里只有是精品在线观看| 亚洲欧美中文字幕日韩二区| 久久久久九九精品影院| 在线观看美女被高潮喷水网站| 国产私拍福利视频在线观看| 国产极品天堂在线| 国内少妇人妻偷人精品xxx网站| 在线观看免费视频日本深夜| 久久精品国产亚洲av天美| 乱系列少妇在线播放| 国产一区二区激情短视频| 变态另类丝袜制服| 久久久久久久久久黄片| 热99在线观看视频| 久久国产乱子免费精品| 国产三级中文精品| 在线观看免费视频日本深夜| 国语自产精品视频在线第100页| 欧美日韩综合久久久久久| 亚洲国产日韩欧美精品在线观看| 中文字幕av在线有码专区| 免费观看精品视频网站| 亚洲国产日韩欧美精品在线观看| 三级经典国产精品| 美女被艹到高潮喷水动态| 精品人妻一区二区三区麻豆| 国产精品永久免费网站| 美女 人体艺术 gogo| 日本撒尿小便嘘嘘汇集6| 简卡轻食公司| 一本久久中文字幕| 一级毛片我不卡| 午夜福利在线观看吧| 一个人观看的视频www高清免费观看| 综合色丁香网| 黄色视频,在线免费观看| 国产一区二区三区在线臀色熟女| 国内久久婷婷六月综合欲色啪| 99热全是精品| 嘟嘟电影网在线观看| 成人二区视频| 老师上课跳d突然被开到最大视频| 精品国内亚洲2022精品成人| 国产精品乱码一区二三区的特点| 91狼人影院| 日本熟妇午夜| 久久精品国产亚洲av涩爱 | 久久精品国产亚洲av香蕉五月| 美女国产视频在线观看| 日韩欧美三级三区| 亚洲内射少妇av| 黄色视频,在线免费观看| 两个人视频免费观看高清| 亚洲av熟女| 老司机影院成人| 青春草国产在线视频 | 亚洲欧美日韩高清在线视频| 麻豆一二三区av精品| 国产在线男女| 国产精品乱码一区二三区的特点| 在线天堂最新版资源| 我的老师免费观看完整版| 久久热精品热| 人人妻人人澡欧美一区二区| 精华霜和精华液先用哪个| 国产精品,欧美在线| 国产黄a三级三级三级人| 国产精品日韩av在线免费观看| 在线观看美女被高潮喷水网站| 欧美zozozo另类| 亚洲最大成人手机在线| 国产老妇伦熟女老妇高清| 韩国av在线不卡| 九九爱精品视频在线观看| 国内精品久久久久精免费| 内射极品少妇av片p| 亚洲欧美日韩东京热| 亚洲最大成人中文| 国产精品1区2区在线观看.| 国产老妇女一区| 亚洲第一区二区三区不卡| 亚洲中文字幕一区二区三区有码在线看| 永久网站在线| 国产精品综合久久久久久久免费| 此物有八面人人有两片| 国产午夜精品久久久久久一区二区三区| 热99在线观看视频| 色噜噜av男人的天堂激情| 免费不卡的大黄色大毛片视频在线观看 | 国产真实伦视频高清在线观看| 99国产极品粉嫩在线观看| 亚洲自拍偷在线| 亚洲欧美精品专区久久| 日韩中字成人| 深夜精品福利| 欧美一区二区亚洲| 插阴视频在线观看视频| 国产精品美女特级片免费视频播放器| 日本撒尿小便嘘嘘汇集6| 国产精品永久免费网站| 亚洲人成网站在线观看播放| 人体艺术视频欧美日本| 久久国产乱子免费精品| 色综合亚洲欧美另类图片| 哪里可以看免费的av片| 99国产精品一区二区蜜桃av| 男插女下体视频免费在线播放| 国产视频内射| 高清在线视频一区二区三区 | 免费看美女性在线毛片视频| 久久综合国产亚洲精品| 人妻少妇偷人精品九色| 国产视频内射| 亚洲天堂国产精品一区在线| 亚洲,欧美,日韩| 日本免费一区二区三区高清不卡| 亚洲激情五月婷婷啪啪| 久久亚洲精品不卡| 国产黄a三级三级三级人| 我要看日韩黄色一级片| videossex国产| 国产精品一区www在线观看| 九九爱精品视频在线观看| 日韩在线高清观看一区二区三区| 国产成人freesex在线| 国产老妇女一区| 自拍偷自拍亚洲精品老妇| 国产私拍福利视频在线观看| av在线播放精品| 成年免费大片在线观看| 国产色婷婷99| 看免费成人av毛片| 久久精品91蜜桃| 久久欧美精品欧美久久欧美| 精华霜和精华液先用哪个| 免费不卡的大黄色大毛片视频在线观看 | 黄色视频,在线免费观看| 国产精品一区二区三区四区久久| 午夜福利在线观看免费完整高清在 | 成人国产麻豆网| 精品久久久久久久久av| 青青草视频在线视频观看| 天天躁日日操中文字幕| 熟女人妻精品中文字幕| 精品久久久久久成人av| 免费观看精品视频网站| 色5月婷婷丁香| 亚洲中文字幕日韩| 国产三级在线视频| 亚洲综合色惰| 亚洲最大成人av| 少妇猛男粗大的猛烈进出视频 | 国产美女午夜福利| 91精品国产九色| 国产精品电影一区二区三区| 成人特级黄色片久久久久久久| avwww免费| 久久精品综合一区二区三区| 亚洲av免费高清在线观看| 久久精品国产亚洲网站| 别揉我奶头 嗯啊视频| av在线亚洲专区| 成人毛片a级毛片在线播放| 色综合亚洲欧美另类图片| 99国产精品一区二区蜜桃av| 国产真实乱freesex| 内地一区二区视频在线| 可以在线观看毛片的网站| 国产成人一区二区在线| 国产高清不卡午夜福利| 亚洲真实伦在线观看| 亚洲一级一片aⅴ在线观看| 欧美日韩乱码在线| 欧美xxxx黑人xx丫x性爽| 久久久久性生活片| 亚洲天堂国产精品一区在线| 久久久久久久久久成人| 麻豆久久精品国产亚洲av| 看非洲黑人一级黄片| 亚洲精品国产成人久久av| 毛片一级片免费看久久久久| 最近2019中文字幕mv第一页| 尤物成人国产欧美一区二区三区| 国产伦在线观看视频一区| 成人毛片60女人毛片免费| 男女做爰动态图高潮gif福利片| 亚洲精品国产成人久久av| 美女cb高潮喷水在线观看| 女的被弄到高潮叫床怎么办| 少妇的逼水好多| 久久精品91蜜桃| 午夜a级毛片| 嫩草影院新地址| 久久午夜亚洲精品久久| 亚洲人成网站高清观看| 亚洲精品自拍成人| av视频在线观看入口| 免费看av在线观看网站| av在线亚洲专区| 日日啪夜夜撸| 国产精品三级大全| 99久久无色码亚洲精品果冻| 国产视频内射| 亚洲欧美日韩高清在线视频| 欧美不卡视频在线免费观看| av在线观看视频网站免费| 久久久久久国产a免费观看| 插阴视频在线观看视频| 国产一级毛片在线| 欧美变态另类bdsm刘玥| 亚洲欧美精品自产自拍| 婷婷六月久久综合丁香| 成人午夜高清在线视频| 女的被弄到高潮叫床怎么办| 麻豆成人av视频| 亚洲欧美精品专区久久| 亚洲精品国产成人久久av| 国产淫片久久久久久久久| 中文在线观看免费www的网站| 久久鲁丝午夜福利片| 精品人妻视频免费看| 日本免费a在线| 麻豆av噜噜一区二区三区| 中文字幕av在线有码专区| 最好的美女福利视频网| 在线观看美女被高潮喷水网站| 成人亚洲精品av一区二区| 久久鲁丝午夜福利片| 国产一区二区三区av在线 | 在线播放国产精品三级| 国产探花极品一区二区| 深夜a级毛片| 日本av手机在线免费观看| 日本黄色片子视频| 麻豆成人av视频| 国产真实伦视频高清在线观看| or卡值多少钱| 美女 人体艺术 gogo| 午夜免费男女啪啪视频观看| ponron亚洲| 国产av不卡久久| 精品欧美国产一区二区三| 国产伦精品一区二区三区四那| 欧美一区二区精品小视频在线| 久久精品综合一区二区三区| 欧美色视频一区免费| 久久久精品大字幕| 九九在线视频观看精品| 国产成人91sexporn| 午夜福利在线观看吧| 蜜桃久久精品国产亚洲av| 黄色配什么色好看| 18禁黄网站禁片免费观看直播| 成人特级av手机在线观看| 熟女人妻精品中文字幕| 亚洲三级黄色毛片| 亚洲色图av天堂| 久久国产乱子免费精品| 一个人免费在线观看电影| 干丝袜人妻中文字幕| 尾随美女入室| 只有这里有精品99| 日本免费a在线| 如何舔出高潮| 一个人看视频在线观看www免费| 国产高清激情床上av| 卡戴珊不雅视频在线播放| 国产亚洲av片在线观看秒播厂 | 久久人人精品亚洲av| 免费看美女性在线毛片视频| 成人鲁丝片一二三区免费| 特级一级黄色大片| 91午夜精品亚洲一区二区三区| 激情 狠狠 欧美| 中文字幕精品亚洲无线码一区| 国产精品不卡视频一区二区| 黄片无遮挡物在线观看| 嫩草影院新地址| 亚洲国产精品久久男人天堂| 九九久久精品国产亚洲av麻豆| 国产一区二区亚洲精品在线观看| 国产黄色视频一区二区在线观看 | 国产av一区在线观看免费| 夜夜看夜夜爽夜夜摸| 久久久久久伊人网av| 国产亚洲av嫩草精品影院| 国产在线精品亚洲第一网站| 99久久久亚洲精品蜜臀av| 麻豆国产97在线/欧美| 成人鲁丝片一二三区免费| av福利片在线观看| 国产精品久久久久久精品电影| 最近手机中文字幕大全| 又黄又爽又刺激的免费视频.| 亚洲国产高清在线一区二区三| 又粗又硬又长又爽又黄的视频 | 十八禁国产超污无遮挡网站| ponron亚洲| 欧美性猛交╳xxx乱大交人| 免费看av在线观看网站| 国产毛片a区久久久久| 少妇丰满av| 一本久久精品| 国产探花在线观看一区二区| 91午夜精品亚洲一区二区三区| 黄色视频,在线免费观看| 久久精品久久久久久久性| 人妻制服诱惑在线中文字幕| 草草在线视频免费看| 最后的刺客免费高清国语| 国产大屁股一区二区在线视频| 熟女电影av网| 亚洲精品粉嫩美女一区| 99热6这里只有精品| 国产成人freesex在线| 婷婷六月久久综合丁香| 给我免费播放毛片高清在线观看| 美女被艹到高潮喷水动态| 久久99精品国语久久久| 岛国在线免费视频观看| 久久精品久久久久久久性| 天天一区二区日本电影三级| 免费av毛片视频| 一本久久中文字幕| 国产伦精品一区二区三区四那| 亚洲成人中文字幕在线播放| 久久久久九九精品影院| 小蜜桃在线观看免费完整版高清| 哪个播放器可以免费观看大片| 亚洲精品国产av成人精品| 欧美高清成人免费视频www| 久久精品久久久久久噜噜老黄 | 美女内射精品一级片tv| 日本一本二区三区精品| 日韩av不卡免费在线播放| 亚洲七黄色美女视频| 国产精品久久电影中文字幕| 欧美三级亚洲精品| 乱人视频在线观看| 国产日韩欧美在线精品| 国产伦精品一区二区三区视频9| 麻豆av噜噜一区二区三区| 悠悠久久av| av在线播放精品| 亚洲国产精品久久男人天堂| 国产午夜福利久久久久久| 国产视频首页在线观看| 国产精品久久久久久久久免| 精品不卡国产一区二区三区| 97在线视频观看| 亚洲一区高清亚洲精品| 精品不卡国产一区二区三区| 国产精品三级大全| 一级av片app| 国产极品精品免费视频能看的| 乱人视频在线观看| 亚洲第一电影网av| 男女啪啪激烈高潮av片| 日日摸夜夜添夜夜爱| 午夜精品一区二区三区免费看| 欧美日本亚洲视频在线播放| 91在线精品国自产拍蜜月| 国产久久久一区二区三区| 人妻夜夜爽99麻豆av| 精品免费久久久久久久清纯| 国产高清三级在线| 只有这里有精品99| 18禁裸乳无遮挡免费网站照片| 成人永久免费在线观看视频| 国产蜜桃级精品一区二区三区| 国产淫片久久久久久久久| 国产午夜精品论理片| 免费不卡的大黄色大毛片视频在线观看 | 亚洲av电影不卡..在线观看| 熟女电影av网| 国产av不卡久久| 精品久久久久久成人av| 亚洲电影在线观看av| 听说在线观看完整版免费高清| 老女人水多毛片| 亚洲精品成人久久久久久| 成人性生交大片免费视频hd| h日本视频在线播放| 久久久精品欧美日韩精品| 真实男女啪啪啪动态图| 18禁黄网站禁片免费观看直播| 国内精品宾馆在线| 亚洲人成网站在线观看播放| 国产精品久久久久久精品电影| 久久精品久久久久久久性| 亚洲一区二区三区色噜噜| 免费搜索国产男女视频| 小蜜桃在线观看免费完整版高清| 精品一区二区免费观看| 韩国av在线不卡| 色哟哟哟哟哟哟| 99久久无色码亚洲精品果冻| 18+在线观看网站| 亚洲av男天堂| 给我免费播放毛片高清在线观看| 国产伦精品一区二区三区四那| 哪个播放器可以免费观看大片| 能在线免费看毛片的网站| 丝袜喷水一区| 三级国产精品欧美在线观看| 国产黄色小视频在线观看| 嘟嘟电影网在线观看| av.在线天堂| 丝袜美腿在线中文| 人人妻人人看人人澡| 亚洲av.av天堂| 婷婷色av中文字幕| 国产爱豆传媒在线观看| 久久午夜福利片| 欧美日韩国产亚洲二区| 日韩三级伦理在线观看| 九九爱精品视频在线观看| 亚洲真实伦在线观看| 日韩亚洲欧美综合| 成人亚洲欧美一区二区av| av视频在线观看入口| 91精品一卡2卡3卡4卡| 国产av不卡久久| 中文字幕人妻熟人妻熟丝袜美| 最好的美女福利视频网| 在线免费十八禁| 亚洲人成网站在线播放欧美日韩| 最近中文字幕高清免费大全6| 久久久久久久亚洲中文字幕| 91av网一区二区| 色视频www国产| 久久热精品热| 午夜免费激情av| 成年版毛片免费区| 日韩成人av中文字幕在线观看| 日韩欧美在线乱码| 国产一区二区在线av高清观看| 青春草国产在线视频 | 亚洲成人中文字幕在线播放| 国产中年淑女户外野战色| ponron亚洲| 久久九九热精品免费| 女同久久另类99精品国产91| 国产精品久久久久久av不卡| www.av在线官网国产| 一个人免费在线观看电影| 日日摸夜夜添夜夜添av毛片| 国产精品爽爽va在线观看网站| 成人欧美大片| 99在线人妻在线中文字幕| 午夜精品一区二区三区免费看| 99久久中文字幕三级久久日本| 国产精品永久免费网站| 少妇被粗大猛烈的视频| 91精品国产九色| 精品人妻视频免费看| 啦啦啦啦在线视频资源| 国产极品天堂在线|