• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Deep Learning-based Wireless Signal Classification in the IoT Environment

    2022-08-23 02:21:04HyejiRohSheungminOhHajunSongJinseoHanandSangsoonLim
    Computers Materials&Continua 2022年6期

    Hyeji Roh,Sheungmin Oh,Hajun Song,Jinseo Han and Sangsoon Lim

    Department of Computer Engineering,Sungkyul University,Anyang,430742,Korea

    Abstract: With the development of the Internet of Things (IoT), diverse wireless devices are increasing rapidly.Those devices have different wireless interfaces that generate incompatible wireless signals.Each signal has its own physical characteristics with signal modulation and demodulation scheme.When there exist different wireless devices,they can suffer from severe Cross-Technology Interferences(CTI).To reduce the communication overhead due to the CTI in the real IoT environment, a central coordinator can be able to detect and identify wireless signals existing in the same communication areas.This paper investigates how to classify various radio signals using Convolutional Neural Networks(CNN),Long Short-Term Memory(LSTM)and attention mechanism.CNN can reduce the amount of computation by reducing weights by using convolution,and LSTM belonging to RNN models can alleviate the long-term dependence problem.Furthermore, attention mechanism can reduce the short-term memory problem of RNNs by reexamining the data output from the decoder and the entire data entered into the encoder at every point in time.To accurately classify radio signals according to their weights, we design a model based on CNN, LSTM, and attention mechanism.As a result, we propose a model CLARINet that can classify original data by minimizing the loss and detects changes in sequences.In a case of the real IoT environment with Wi-Fi, Bluetooth and ZigBee devices, we can normally obtain wireless signals from 10 to 20 dB.The accuracy of CLARINet’s radio signal classification with CNN-LSTM and attention mechanism can be seen that signal-to-noise ratio (SNR) exhibits high accuracy at 16 dB to about 92.03%.

    Keywords:Attention mechanism;wireless signal;CNN-LSTM;classification;deep-learning

    1 Introduction

    In the Internet of Things(IoT)environment,wireless signals are different between various wireless devices, and wireless signals are complexly mixed to form crowded signals [1].When there exist different wireless networks such as Wi-Fi, Bluetooth and ZigBee in the 2.4 GHz spectrum band,they can suffer from severe Cross-Technology Interferences (CTI).To reduce the communication overhead due to the CTI in the real IoT environment,a central coordinator can be able to detect and identify wireless signals existing in the same communication areas.Therefore, it is very challenging task to classify and obtain individual wireless signals in spaces where there are many obstacles,such as people or walls, or where there are many wireless signals, and to use them reliably.Based on the result of wireless signal classification,each IoT device adjusts its own communication channel to avoid congestions.In environments where wireless signals are diverse and heavily intertwined using a variety of wireless devices,it is difficult to solve these problems using traditional wireless signal classification methods.Therefore,deep learning models are being studied to solve complex wireless signal problems wisely,enabling models and systems to be created that show good performance compared to previously proposed wireless signal classification models.

    Convolutional Neural Network (CNN), which belongs to deep neural network among deep learning techniques,use convolution to reduce the number of weights required for image processing,thereby reducing computation and aiming for effective image processing.CNN consists of convolution layers and pooling layers, and solves the vanishing gradient problem using Rectified Linear Unit(ReLU)as activation function.In addition,CNN is characterized by deriving output values of a given size as a result through input values of a given size.Through this,most studies utilize CNN to predict time series data.

    Recurrent Neural Network (RNN), one of the artificial neural networks, is a sequence model of deep learning, which uses input and output data split into sequence units for natural language processing.And since it has a circular structure, it processes sequence-type inputs through internal memory.The Sequence-to-Sequence (Seq2Seq) model and the Long Short-Term Memory (LSTM)cell are representative of the RNN.The Seq2Seq model consists of two architectures,an encoder and a decoder, which processes every word in the entered sentence sequentially, eventually compressing the information of every word into a context vector.This context vector transmits the compressed information to the decoder architecture, processes it to the desired conditions, and outputs them sequentially.Each cell of the encoder and decoder of the Seq2Seq model consists of an LSTM cell or a Gated Recurrent Units(GRU)cell.However,interpreting sentences using the Seq2Seq model has the problem of losing information or vanishing gradient,which results in some information disappearing when the input sentence is long, resulting in reduced accuracy.LSTM is an RNN with the simplest form, which can alleviate the long-term dependence problem that rely on previous computational result to lose memory.It also calculates weights so that important inputs can be recognized by passing through a total of three gates:forget gate,input gate,and output gate.LSTM is highly utilized for longterm signals such as long-term sentences and time series predictions, as it can be used to remember and store important parts from past data,preserve them,and extract necessary parts by iterating the task of applying them to previous and current data[2,3].However,the problem with LSTM is that it is likely to be interrupted by an infinite increase in memory,and that the computation speed is quite slow[4].To address these problems,there are cases where we apply peephole connections to LSTM or use GRU that simplify computation to update hidden states[5].

    Furthermore,attention mechanism selectively learns only the parts that have a significant impact at every point in time,thus reducing the short-term memory problem of RNN.Attention mechanism transfers all the outputs that went through the encoder to the decoder and computes the sum of weights for the outputs of all the encoders through the decoder’s memory cells to determine the important words.This process allows the decoder to focus on and process words that are considered more important than other words.

    CLARINet,the model proposed in this paper,is designed to classify wireless signals similar to the original by applying attention mechanism based on CNN-LSTM,which has recently been utilized in time series prediction to show outstanding prediction performance.CLARINet is passes data through Conv1D twice, then through LSTM layer twice, applies attention mechanism, and applies softmax function,reducing distortion and loss of original data of the wireless signal.We propose a final signal classification model that minimizes distortion by passing signal data to CNN and three gates inherent in LSTM cells to classify importance,obtaining attention score and attention value,connecting hidden states at point in time,and iterating output layer computation process to weight important signals.And we describe experiments to verify accuracy and their result.

    Our main contributions are summarized as follows.

    · We propose a model that can be classified among various radio signals by minimizing distortion and loss for each wireless signal.

    ·We devise a method to classify wireless signals wisely based on CNN-LSTM by applying the techniques used in natural language processing.

    · We propose a model that most accurately classifies wireless signals by applying attention mechanism for various existing wireless signal analysis methods.

    ·We help to obtain the individual’s wireless signal in a crowded space,and propose a reliable model for the accuracy of wireless signal classification.

    The rest of this paper is organized as follows.Section 2 describes studies using radio machine learning dataset and related studies on wireless signal classification using traditional wireless signal classification techniques, deep learning.Section 3 describes the structure and operating principles of the CLARINet model applied with our proposed technique, and presents experimental results in Section 4.Finally,we conclude this paper in Section 5.

    2 Related Work

    This chapter describes the study of classifying radio signals by reducing interference in complex radio signal environments and the study of solutions to address deep learning-based radio signal classification.

    In Section 2.1,we describe a study dealing with a novel algorithm for identifying wireless signal modulation or a wireless signal classification technique that proposes improved directions.

    In Section 2.2,we describe the study of techniques for classifying modulation of radio signals by applying various techniques in deep learning.

    2.1 Radio Signal Classification Study

    Methods for identifying modulation for wireless signals have long been studied.The different devices that make up the Internet of Things communicate using different wireless signals.However,if many IoT devices are used in one space,interference occurs with different wireless signals.Therefore,it is difficult for wireless devices to seamlessly obtain individual wireless signals in large spaces.Since wireless signals are modulated without maintaining the original signal during the communication process,accurately classifying signals between complex signal interferences takes considerable time and requires long training.Recently,to address this problem,we have attempted radio signal classification using artificial intelligence technology and continue our research to show near 90% performance[6].Furthermore, we analyzed a study that judged accuracy on the results of implementing a fast,real-time wireless signal classification network to accurately classify the modulation of wireless signals [7].Furthermore, with a study where a novel algorithm extracts key features to identify modulation of radio signals, we designed a model that accurately classifies signals in complex radio signal environments [8].We designed CLARINet to secure individual radio signals because these studies have similar exact classifications for wireless signal modulation in many spaces.Furthermore,we seek to solve the problem of low accuracy of radio signal classification through deep learning-based models in a way to detect conventional radio signal modulation.

    2.2 Wireless Signal Classification Solution based on Deep Learning

    Recently, research on designing deep learning-based models has been rapidly evolving, and research has been underway to increase the performance of wireless signal classification by applying various models of deep learning to one or more existing designed wireless signal classification methods.For example,we aim to solve the problem of modulation of wireless signals by designing an extended framework based on CNN to increase the accuracy of radio signal classification [9] or to learn amplitude and phase information of training data through a model based on one of the RNN models[10].To increase efficiency such as accuracy or performance of the designed model,we have attempted to classify wireless signals by applying deep learning[11]or leveraging high-order cumulants(HOC)and machine learning [12].We seek to address the problem of failure to go beyond a certain range of accuracy by grafting an attention mechanism that can solve the information loss problem caused by encoding all information with fixed-length vectors, as the point of using the LSTM model when classifying wireless signals is similar.

    There is also an example of analyzing a model using radio machine learning dataset to evaluate the accuracy and performance of deep learning-based signal classification models based on signal-to-noise ratio(SNR).The radio machine learning dataset contains data on 11 signals,including 8PSK,AMDSB,AM-SSB,BPSK,CPFSK,GFSK,PAM4,QAM16,QAM64,QPSK,and WBFM.Our goal is to use this dataset as training data for deep learning models to solve problems that do not increase the accuracy of the previously proposed models.One study presented a plan to improve the performance of the model as automatic modulation classification(AMC)works progress[13],and there is a study that improves speed and accuracy by designing models with higher accuracy than conventional models[14].There is a study that proposed an algorithm that can increase the dataset so that the deep learning model can learn enough to improve the problem of lack of datasets[15].In addition,there is a study in which deep learning-based models have been trained by utilizing radio machine learning dataset to solve problems vulnerable to adversarial attacks.Furthermore,a classifier utilizing deep learning has a case of using radio machine learning dataset to demonstrate that even highly dependent and short radio signals can be misclassified when classifying radio signals [16].And one study used a CNNbased model to extract features learned using CNN to cluster wireless signal modulation types even for training data that are not labeled[17].

    As such, radio machine learning dataset is similar in that it is used to build and validate multiple wireless signal classification techniques using deep learning to analyze accuracy and check performance.Therefore,our designed CLARINet model similarly clarifies the criteria for determining accuracy by leveraging radio machine learning dataset as learning data to validate the performance of the model and the training.

    3 Design of Model

    We design a deep learning-based wireless signal classification model CLARINet, which incorporates attention mechanism into the results through two Conv1D layers and two LSTM layers.Model CLARINet allows complex and diverse wireless signal input data to extract data even in the context of transformations or distortions of attributes via CNN,and when extracted data is entered via LSTM’s encoder,it processes each gate’s characteristics via LSTM’s three gates,and applies the output results through LSTM cells to attention mechanism.Attention mechanism analyzes more intensively on data considered as necessary data through LSTM cells to produce more accurate classification results compared to results using only LSTM.Finally,the final result obtained by attention mechanism applies softmax regression using a cross entropy function as a cost function,classifying the first entered complex radio signals into a total of 11 radio signals (8PSK, AM-DSB, AM-SSB, BPSK, CPFSK,PAM4,QAM16,QAM64,and WPSK).

    3.1 Data

    The collected signals for data classification have a complex number of forms for flexibility and simplicity for mathematical operations,expressed as I=Acos(φ)and Q=Asin(φ).A andφrefer to the instantaneous amplitude and phase of the collected signal.RadioML2016.10a dataset follows a data representation using I,Q,and we use RadioML2016.10a dataset for training and performance evaluation of CLARINet model proposed in this paper.RadioML2016.10a dataset is a synthetic dataset with modulation methods currently in commercial use using GNU radio,which implements similar real-world noise environments such as multipath fading and white noise.This data set contains 128 sample data of 4 samples/symbol.It also consists of Python dict data stored in the form of Python pickle files, and consists of keys and values.Each key consists of 11 modulation methods and -20 to 18 dB of SNR tuple,and the value has a numpy array of(1000,2,128)corresponding to 220 key values.It consists of 1000 sample windows with two values,I and Q of 128 samples.

    In this work, we judged that low SNR data adversely affected learning performance, so we conducted learning using SNR from-10 to 18 dB,and simply because learning using I and Q values did not perform well, we changed I and Q values to phase and amplitude values.Furthermore, we compress from 128 samples of data to 64 samples by replacing two close values with average values for better learning performance.As a result,we have achieved approximately six times the performance improvement in CPU environments,with no significant variations in the shape of the data and little impact on accuracy.This allows us to complete the learning in a reasonable amount of time without using GPUs.In the learning process for maintenance,it is believed that it will be able to save a lot of money when learning using cloud servers.

    3.2 CNN

    CNN has a structure in which input data is configured to go through the convolution layer and the pooling layer,with a fully connected layer at the end.The entire structure of CNN is the same as Fig.1.

    The convolution layer is responsible for maintaining the shape of the input data,creating a feature map for each filter,and the pooling layer receives the output data from the convolution layer as input,reducing the size of the output data or emphasizing specific data.This allows data to be extracted from the data where features are modified or distorted.This allows data to be extracted from the data where features are modified or distorted.In addition,this layer can automatically extract properties.The convolution layer is a necessary layer,and the pooling layer is an optional layer.

    Figure 1:CNN architecture

    The convolution layer uses filters (two-dimensional matrices in the form of N×M) to extract features of the image.In two-dimensional data consisting of height and width, N×M-sized filters are traversed at a specified interval,multiplying the overlapping data by the values of the elements in the kernel,and then adding all the multiplied values.The traveling interval is called the“stride”,and if the “stride”is specified as 1, it moves one column at a time and make a convolution.The output from the convolution is called feature map,and the application of the active function to feature map is called activation map.After the convolution process,the output data is smaller than the input data,which goes through a process called padding to prevent the output data from decreasing.Padding is the process of filling the edge of the input data with a specific value by a specified size,usually zeroed.

    Pooling layers include max pooling,min pooling,and average pooling.Likewise,the concept of filter and stride is applied to the pooling operation,and usually the filter and stride are identical so that all elements can be processed once.For max pooling, we extract the maximum value from the region where the filter and the data overlap,and similarly,average pooling is the method of extracting the mean.The convolution and pooling operations look similar in that the filter and stride concepts are used,but the pooling operations differ in that no weights exist.

    We apply CNN layer to CLARINet to reduce the loss to the original data and make sure that we do not lose the association of each radio signal.

    3.3 LSTM

    LSTM is used to alleviate the long-term dependence problem of RNN mentioned in Section 1,and the overall structure consists of cells and three gates,such as Fig.2.Cells have values for arbitrary time intervals,and three gates are responsible for removing unnecessary information or leaving only necessary information that is considered important.

    Cell state of LSTM is the horizontal line at the top of Fig.2.Cell state(Ct)represents the state of the cell and is used to obtain the state of the next cell through the state of the previous cell,and serves to convey information generated from the previous node.This cell state allows the operation to be repeated again because it moves information considered important information in the previous step.LSTM performs the process of adding or erasing information from cell states through three gates.tanhstands for hyperbolic tangent function andσfor sigmoid function.LSTM also has three gates,each consisting of a get gate,an input gate,and an output gate.These gates are used to obtain hidden and cell state values.For each gate,a sigmoid function is applied to determine whether the previous data will affect the following data based on the derived values between 0 and 1.

    Figure 2:LSTM architecture

    The operating process of LSTM,such as Fig.3,first enters the wireless signal via the forget gate into the input information.Forget gate is a gate that determines what information is reflected in the current information through the operation of Eq.(1)and is determined via sigmoid function.

    Figure 3:LSTM gate architecture

    t represents the time point,Wfrepresents the weight,andbfrepresents the bias.

    Theftof Eq.(1)has a value of[0,1]through the sigmoid function.If theftis close to 1,the previous information is reflected a lot.If theftis close to 0,the value is reflected less.

    The input gate is configured as Fig.3 and is responsible for remembering the information to store for new information.To remember new information, perform the operations in Eqs.(2) and (3).t represents the time point,andWiandWCdenotes the weight.biandbCdenote bias.

    Theitof Eq.(2) shall have the value [0, 1] obtained by the sigmoid function and shall be determined to reflect the present information.When the value ofitis 1,[-1,1]values obtained from the hyperbolic tangent function of the Eq.(3)determine the candidate vector to be added to the cell state.These two values select information and determine how much to remember.

    Then,to update the contents of the forget gate and input gate,we go through Fig.3.Update the information by applying Eq.(4)toCt-1crossed from the previous node.

    Finally,output gate is the final step in determining which data to output,corresponding to the final step in Fig.3.The input is entered into the sigmoid function and the value[0,1]is output,which determines whether to export part of the cell state to output.This output is then passed through the hyperbolic tangent function to the input of the next state.

    We leverage CLARINet to automatically do well filtering on given data via CNN layer.Furthermore,we put CNN layers at the forefront of CLARINet to learn important characteristics for each radio signal when classifying radio signals,and to preserve the association of the data,leveraging them to remember the features of radio signals.

    3.4 Attention Mechanism

    Attention references and applies the entire input sentence at each point in predicting the output data,and concentrates the data associated with the data to be predicted and forwards it to the decoder.This mechanism allows us to deliver more data than was previously delivered.The method in which attention mechanism is used is the same as Fig.4.

    Figure 4:Attention mechanism architecture

    The attention score must be obtained to apply the attention mechanism to the layer consisting of LSTM.Attention score is a score that determines the similarity between the hidden state of the encoder and the hidden state stof the encoder at the present time.The softmax function is applied to obtain an attention distribution in which the sum of all values is equal to 1,and each value is an attention weight.This value and the hidden state provide an attention value,αt.The hidden state of the decoder is connected to this value,and the operation of Eq.(7)is performed to create and useas the input of the output layer.Wcstands for weight matrix,and bcstands for bias.

    We apply an attention mechanism to CLARINet to design an accurate classification by weighting the factors that have a significant impact when distinguishing the features of each radio signal.

    3.5 Architecture of CLARINet

    Finally, our designed network of CLARINet is designed to reduce conversion, distortion, and loss to the original data through two Conv1D layers,extract the data,and delete the rest of the data except the critical data through two LSTM layers.We then design an attention mechanism to weight important data so that it can be classified focusing on important data when analyzing radio signals.The structure of CLARINet performing this process is the same as Fig.5.

    Figure 5:CLARINet model architecture

    CLARINet receives the original radio signal for the frequency band signal as input data.We set the number of filters, the underlying property of CLARINet, to 64.Filter determines the filter size,determines the stride,and recognizes the data according to the length of the stride,and generates the activation map as an output.Therefore,CLARINet is designed to generate 64 activation maps over 64 filters.

    We designed that data passed through CNN layers remember the features each wireless signal has in order to classify wireless signal over LSTM layer,considering then as important information.It is also designed to remove hidden unit with a 60%chance by leaving the dropout of CLARINet at 0.6.In the LSTM layer,dropout is a type of regularization that solves overfitting and makes it not dependent on any single data through dropout.Through this,the CLARINet model is designed to be overfitted and non-dependent.

    Output data through CNN layer and LSTM layer are applied to attention mechanism to solve the gradient loss problem, increasing accuracy, and classified radio signals have an organic relationship with each other.Eventually,data passed to attention mechanisms are classified into 11 radio signals(8PSK,AM-DSB,AM-SSB,BPSK,CPFSK,GFSK,PAM4,QAM16,QAM64,QPSK,WBFM)via softmax regression.

    4 Result of Experiment

    To compare the performance of CLARINet on accuracy,we conducted experiments by selecting a model that combines two CNN layers and two LSTM layers,two CNN layers,one LSTM layer,two LSTM layers,two CNN layers and attention mechanism,as a comparison group.The results of our comparison of the accuracy according to SNR are as shown in Fig.6.

    Figure 6:Classification accuracy comparison of CLARINet

    If SNR is in the range of-20 to-10 dB,CLARINet and all the seven comparator models selected show less than 10% accuracy, with a narrow rise.Therefore, we analyze the range of SNR based on-10 dB or more to clearly classify the accuracy for CLARINet.Our design of CLARINet shows an average accuracy of 77.40% for SNRs above -10 dB, highest accuracy from 16 dB to 92.03%, and approximately 7.34%higher than that of seven comparators at 16 dB.

    We analyze on seven comparator models selected to classify radio signals, and we find that the classification accuracy of models with CNN layers is higher than that of models without CNN layers,and that the accuracy of models with two CNN layers in the range 0 to 18 dB shows an average accuracy of 80%.Furthermore,we compare models based on CNN layers,showing that the two-applied LSTM layers have approximately 2.98% higher accuracy than the single attention mechanism.This shows that CNN layer-based models exhibit high values in the range of 0 to 18 dB in determining accuracy for radio signal classification,and that additional LSTM layer or attention mechanism can be applied to increase accuracy.

    The results of CLARINet experiments on-8,0,16 and 18 dB on SNR basis to verify accuracy in classifying radio signals in complex radio signal environments were shown as Figs.7–10.

    Figure 7:Confusion matrix at SNR-8 dB

    SNR below -10 dB has little significance in data results because of its low accuracy, and SNR of -8 dB can be found to be mostly low in accuracy, such as Fig.7.However, out of a total of 11 modulation techniques, we can confirm that AM-DSB, AM-SSB, PAM4 and QAM64 exceed 50%accuracy,which means that the modulation techniques can be classified to some extent even if noise is severe.

    According to Fig.8,CLARINet shows nearly 90%accuracy on average at SNR 0 dB,and from 4 dB,it can be seen that the accuracy is over 90%on average.In particular,we show that AM-SSB is more classified with an accuracy of over 90%.While the SNR of 0 dB is mostly over 90%accuracy,8PSK,QAM16,QAM64,and WBFM of the 11 modulation techniques do not exceed 90%accuracy.QAM16 is a subset of QAM64,which often misjudges QAM16 as QAM64 because only the bits that can be sent from one signal are different.WBFM has the lowest accuracy among the 11 modulation techniques, and the signal from WBFM is misclassified as AM-DSB due to the absence of a signal because it was modulated in a real audio stream.At 18 dB SNR,most modulation techniques,such as Fig.9,show an average accuracy of over 90%,and the accuracy for QAM16 is about 10%higher than the accuracy classified in 0 dB.Furthermore,WBFM was also shown to be 10%higher than the accuracy classified at 0 dB,but relatively lower than other signals.

    Figure 8:Confusion matrix at SNR 0 dB

    Figure 9:Confusion matrix at SNR 18 dB

    Figure 10:Confusion matrix at SNR 16 dB

    We retained the basic properties of CLARINet to verify the experimental results according to the properties of CLARINet, and proceeded with the experiment by changing the number of filters.CLARINet has 64 filters, and the number of filters in the comparison model we selected for comparison is 32, 64, 128, and 256.The experimental results comparing accuracy according to the number of filters are as shown in Fig.11.

    Figure 11:Classification accuracy results by number of filters

    When the number of filters is 64, they represent the highest accuracy at all 0 to 18 dB, and the highest accuracy at 16 dB to 92%.While 128 filters and 256 filters represent similar accuracy overall,it can be seen that 128 filters represent slightly higher accuracy.The number of filters increases mainly as the layers are placed behind, with CNN layers located in front of the CLARINet model, with a relatively small number of 64 filters having higher accuracy than 128 filters and 256 filters.

    We further analyzed for models with 64 filter and 128 filter indicating high accuracy in Fig.11.The results of analyzing the two models based on total steps, initial loss, final loss, and runtime are the same as Tab.1.

    Table 1: 64 filter and 128 filter analysis results

    Comparing epochs according to the number of filters, 64 filter shows that Total Steps was 20 more times than 128 filter, which led to more learning.We used categorical_crossentropy as a loss function of CLARINet, and after checking the loss cost, we found that both 64 filter and 128 filter show approximately 1–2% loss.We can see that the final runtime of our designed CLARINet takes about two hours less than 128 filters,depending on the number of filters.

    In addition, experimental results comparing accuracy by changing the value of dropout, a regulatory technique to prevent overfitting,were shown as Fig.12.P,the hyperparameter of dropout,means probability.The probability of dropout temporarily changes depending on the value of this p.Our designed CLARINet is designed by selecting the p value of dropout as 0.6.

    Figure 12:Classification accuracy results by dropout

    According to Fig.12,when dropout is 0.6,it is shown that the highest accuracy is from 0 to 18 dB,and when dropout is 0.4 it is the next highest accuracy.If dropout is 0.8 then low performance indicates that strong regulation indicates low accuracy,and the most common accuracy when left at 20%to 60%.It showed an accuracy of 92%at 16 dB.Through this,we confirm that designing to dropout with a 60%chance at each training step can yield the best performance to improve the accuracy of CLARINet.

    Based on the experimental results,we design CLARINet as a structure of two CNN layers,two LSTM layers,and an attention mechanism,with 64 filters on the CNN layer and 0.6 dropout on the LSTM layer,selecting the model with the highest accuracy.

    5 Conclusion

    We propose a novel model CLARINet that integrates CNN layer with LSTM layer and attention mechanism as a deep learning-based solution for classifying wireless signals in the IoT environment.Many previous studies have attempted radio signal classification based on original signals with less distortion or loss of radio signals,and have proposed efforts to improve performance on radio signal classification by incorporating various techniques.However, the exact classification of each radio signal has yet to be completely resolved, as radio signals are not separated and propagated, but are complexly propagated in crowded spaces.Previous studies have mainly improved accuracy problems by implementing LSTM-based models.Therefore, we design a model CLARINet with LSTM layer and attention mechanism applied to CNN-based models to accurately classify complex radio signals for each feature.

    We show that CLARINet,which is designed to allow wireless signals to obtain individual radio signals in congested spaces,shows approximately 60%accuracy for environments with SNR of-20 to 18 dB,with approximately 92.03%accuracy at 16 dB.Analysis of CNN layer and LSTM layer used in CLARINet structure shows that CNN-based models have an average accuracy of about 40%higher than LSTM-based models.Through this,we have shown that classifying complex radio signals through CNN-based models exhibits higher accuracy than those that do not.Furthermore,we use attention mechanism to weight features on radio signals,remember only important features of radio signals and classify them,identify the possibility of minimizing distortion and loss,and finally confirm that they can be classified into 11 radio signals via softmax regression.

    In the future,we plan to improve accuracy by changing the attributes of CLARINet models or by adding layers,and explore ways to improve on misclassifying QAM16 as QAM64 and misclassifying WBFM as AM-DSB.Furthermore,we plan to utilize CLARINet to conduct experiments on image classification to improve the problem by applying it to problems that suffer from data loss or distortion,and simplify our model to optimize the overall performance.

    Funding Statement:This work was supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(No.2021R1F1A1063319).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲av日韩精品久久久久久密| 亚洲午夜理论影院| 午夜福利视频1000在线观看| 欧美成人a在线观看| 国产高清有码在线观看视频| 十八禁人妻一区二区| 久99久视频精品免费| 一本精品99久久精品77| 久久香蕉精品热| avwww免费| 超碰av人人做人人爽久久| 简卡轻食公司| 看免费av毛片| 国产爱豆传媒在线观看| 美女黄网站色视频| 桃红色精品国产亚洲av| 尤物成人国产欧美一区二区三区| 小说图片视频综合网站| 午夜a级毛片| 观看美女的网站| 国产日本99.免费观看| 国产伦精品一区二区三区四那| 亚洲精品在线美女| 欧美黑人欧美精品刺激| 久久午夜亚洲精品久久| 天堂网av新在线| 动漫黄色视频在线观看| 亚洲激情在线av| 性插视频无遮挡在线免费观看| 特大巨黑吊av在线直播| 欧美精品啪啪一区二区三区| 草草在线视频免费看| 99国产精品一区二区蜜桃av| 级片在线观看| 国产视频一区二区在线看| av欧美777| av视频在线观看入口| 91麻豆av在线| 成人国产综合亚洲| avwww免费| 69av精品久久久久久| 极品教师在线视频| 亚洲最大成人中文| 给我免费播放毛片高清在线观看| 国产在线精品亚洲第一网站| 51国产日韩欧美| av天堂在线播放| 久久热精品热| 国语自产精品视频在线第100页| 国产成+人综合+亚洲专区| 国产伦精品一区二区三区四那| 国产熟女xx| 欧美精品啪啪一区二区三区| 成年女人看的毛片在线观看| 欧美日本亚洲视频在线播放| www日本黄色视频网| 国产私拍福利视频在线观看| 欧美一区二区国产精品久久精品| 亚洲av日韩精品久久久久久密| 我要看日韩黄色一级片| 最近最新免费中文字幕在线| 国产精华一区二区三区| 国内精品久久久久久久电影| 美女高潮的动态| 精品一区二区三区av网在线观看| 国产综合懂色| 成年女人看的毛片在线观看| 精品熟女少妇八av免费久了| 欧美一级a爱片免费观看看| 身体一侧抽搐| 听说在线观看完整版免费高清| 国产日本99.免费观看| 中文资源天堂在线| 丁香欧美五月| 国产成人影院久久av| 免费在线观看亚洲国产| 亚洲欧美日韩卡通动漫| 欧美潮喷喷水| 中文字幕免费在线视频6| 国产精品伦人一区二区| 99久久成人亚洲精品观看| 国产在线精品亚洲第一网站| 欧美高清成人免费视频www| 老女人水多毛片| 中文字幕久久专区| 欧美最黄视频在线播放免费| 欧美日韩瑟瑟在线播放| 一区二区三区激情视频| 欧美+亚洲+日韩+国产| av国产免费在线观看| 国产成年人精品一区二区| 听说在线观看完整版免费高清| 欧美xxxx黑人xx丫x性爽| 日韩有码中文字幕| 男人狂女人下面高潮的视频| 久久国产精品人妻蜜桃| 亚洲国产日韩欧美精品在线观看| 国语自产精品视频在线第100页| 日本黄大片高清| 88av欧美| 免费在线观看成人毛片| 伦理电影大哥的女人| 久久久成人免费电影| 国模一区二区三区四区视频| av在线老鸭窝| 国产精品一区二区性色av| 国产黄a三级三级三级人| 无人区码免费观看不卡| 欧美+日韩+精品| 国产一区二区在线观看日韩| 午夜精品在线福利| 久久久久免费精品人妻一区二区| 久久这里只有精品中国| 国产欧美日韩一区二区精品| 一本综合久久免费| 国产精品av视频在线免费观看| 国模一区二区三区四区视频| 中文字幕高清在线视频| 91av网一区二区| 夜夜爽天天搞| 亚洲在线观看片| 国内少妇人妻偷人精品xxx网站| a级毛片a级免费在线| 夜夜夜夜夜久久久久| 日韩国内少妇激情av| 久久精品久久久久久噜噜老黄 | 琪琪午夜伦伦电影理论片6080| 热99re8久久精品国产| 精品久久久久久久人妻蜜臀av| 免费看a级黄色片| 欧美成人性av电影在线观看| 亚洲在线自拍视频| 最好的美女福利视频网| 国产精品久久久久久人妻精品电影| 国产av一区在线观看免费| 亚洲中文字幕一区二区三区有码在线看| 国产国拍精品亚洲av在线观看| 在线看三级毛片| 在线十欧美十亚洲十日本专区| 99久久精品热视频| 日韩有码中文字幕| 少妇高潮的动态图| 哪里可以看免费的av片| 极品教师在线视频| 1000部很黄的大片| 亚洲av一区综合| 国产精品一区二区性色av| 精品久久国产蜜桃| 欧美黑人欧美精品刺激| 亚州av有码| 国产伦精品一区二区三区四那| 天堂影院成人在线观看| 啦啦啦韩国在线观看视频| 国产在线精品亚洲第一网站| 国产激情偷乱视频一区二区| 欧美日本视频| 久久精品国产亚洲av涩爱 | 国产成人福利小说| 黄色女人牲交| 亚洲,欧美精品.| www.999成人在线观看| 最新中文字幕久久久久| 哪里可以看免费的av片| 欧美黑人巨大hd| 老司机午夜福利在线观看视频| 免费一级毛片在线播放高清视频| 少妇熟女aⅴ在线视频| 欧美日本亚洲视频在线播放| 88av欧美| 免费搜索国产男女视频| 国产亚洲欧美98| 国产激情偷乱视频一区二区| 夜夜看夜夜爽夜夜摸| 中文字幕人成人乱码亚洲影| av在线天堂中文字幕| 欧美zozozo另类| 亚洲欧美精品综合久久99| 99视频精品全部免费 在线| 九九热线精品视视频播放| 欧美日韩亚洲国产一区二区在线观看| 又粗又爽又猛毛片免费看| 成人鲁丝片一二三区免费| 看免费av毛片| 黄色女人牲交| 久久久国产成人精品二区| 亚洲精品一卡2卡三卡4卡5卡| 天堂√8在线中文| 日本黄色片子视频| 一区二区三区激情视频| 在线观看美女被高潮喷水网站 | 2021天堂中文幕一二区在线观| 天天躁日日操中文字幕| 欧美高清成人免费视频www| 草草在线视频免费看| 熟女人妻精品中文字幕| 757午夜福利合集在线观看| 88av欧美| 日本三级黄在线观看| 午夜福利高清视频| www.www免费av| 国产综合懂色| 欧美在线黄色| 好男人在线观看高清免费视频| 国产午夜精品论理片| 精品一区二区免费观看| 成人鲁丝片一二三区免费| 一区二区三区激情视频| 欧美在线一区亚洲| 中文资源天堂在线| 99在线人妻在线中文字幕| 婷婷丁香在线五月| 91午夜精品亚洲一区二区三区 | 男人狂女人下面高潮的视频| 一个人看视频在线观看www免费| 久99久视频精品免费| 日本黄色片子视频| 亚洲三级黄色毛片| 激情在线观看视频在线高清| 国产乱人视频| 黄片小视频在线播放| 嫁个100分男人电影在线观看| bbb黄色大片| 在线看三级毛片| 看黄色毛片网站| 久久久久久九九精品二区国产| 午夜亚洲福利在线播放| 亚洲中文日韩欧美视频| 97超视频在线观看视频| 一卡2卡三卡四卡精品乱码亚洲| 欧美一区二区精品小视频在线| 亚洲av熟女| 精品国内亚洲2022精品成人| 丰满的人妻完整版| 成年女人毛片免费观看观看9| 亚洲成人久久性| 国产精品久久久久久人妻精品电影| 看片在线看免费视频| 国产高清激情床上av| 国内精品久久久久久久电影| 日韩欧美 国产精品| 啪啪无遮挡十八禁网站| 色5月婷婷丁香| 在线免费观看的www视频| 亚洲一区高清亚洲精品| 日本免费一区二区三区高清不卡| 久久九九热精品免费| 天天一区二区日本电影三级| 国产精品爽爽va在线观看网站| 国产综合懂色| 久久久成人免费电影| 美女高潮的动态| 悠悠久久av| 亚洲av熟女| 久久久久久久久中文| 男人舔女人下体高潮全视频| netflix在线观看网站| 国产欧美日韩精品亚洲av| 久久6这里有精品| 国产精品久久电影中文字幕| 赤兔流量卡办理| 黄色日韩在线| 99热这里只有是精品在线观看 | 成年版毛片免费区| 亚洲真实伦在线观看| 热99在线观看视频| 亚洲国产精品成人综合色| 久久草成人影院| 欧美绝顶高潮抽搐喷水| 国内揄拍国产精品人妻在线| 嫩草影院入口| 2021天堂中文幕一二区在线观| 国产精品综合久久久久久久免费| 18禁黄网站禁片免费观看直播| 欧美精品国产亚洲| 成年人黄色毛片网站| 免费观看精品视频网站| 两个人的视频大全免费| 成人特级av手机在线观看| 国产麻豆成人av免费视频| 国产欧美日韩精品一区二区| 国产精品综合久久久久久久免费| a级毛片免费高清观看在线播放| 国产真实乱freesex| 真人做人爱边吃奶动态| 嫩草影视91久久| 亚洲最大成人av| 一级黄片播放器| 色5月婷婷丁香| 国产毛片a区久久久久| 无遮挡黄片免费观看| 亚洲av美国av| 波多野结衣高清无吗| 哪里可以看免费的av片| 色av中文字幕| 婷婷精品国产亚洲av| 人人妻,人人澡人人爽秒播| 亚洲片人在线观看| 51午夜福利影视在线观看| 精品久久久久久久末码| 欧美性感艳星| 亚洲人与动物交配视频| 国产又黄又爽又无遮挡在线| 日韩免费av在线播放| 午夜福利免费观看在线| 啪啪无遮挡十八禁网站| 草草在线视频免费看| 免费看美女性在线毛片视频| 欧美成人一区二区免费高清观看| 欧美丝袜亚洲另类 | 亚洲人成网站高清观看| 欧美成人a在线观看| 夜夜夜夜夜久久久久| 亚洲成人久久性| 男人的好看免费观看在线视频| 国产白丝娇喘喷水9色精品| 一个人免费在线观看电影| 看黄色毛片网站| 九色成人免费人妻av| 观看免费一级毛片| 欧美黄色片欧美黄色片| 国产成人a区在线观看| 亚洲综合色惰| 99久久成人亚洲精品观看| 最新中文字幕久久久久| 国产成人啪精品午夜网站| 亚洲欧美日韩无卡精品| 国产精品伦人一区二区| av国产免费在线观看| 女人被狂操c到高潮| 午夜福利在线在线| 桃色一区二区三区在线观看| 亚洲精品在线观看二区| 99热这里只有是精品在线观看 | 精品午夜福利视频在线观看一区| 2021天堂中文幕一二区在线观| 午夜激情欧美在线| 久久欧美精品欧美久久欧美| 欧美色欧美亚洲另类二区| 精品人妻偷拍中文字幕| 午夜激情福利司机影院| 亚洲真实伦在线观看| 一级黄片播放器| 91午夜精品亚洲一区二区三区 | 真人做人爱边吃奶动态| 女生性感内裤真人,穿戴方法视频| 国内精品久久久久精免费| 亚洲av美国av| 免费在线观看日本一区| АⅤ资源中文在线天堂| 免费一级毛片在线播放高清视频| 午夜福利免费观看在线| 中文资源天堂在线| 搡老妇女老女人老熟妇| 人妻制服诱惑在线中文字幕| 丝袜美腿在线中文| 日韩精品中文字幕看吧| 丝袜美腿在线中文| 变态另类成人亚洲欧美熟女| 久久久久国内视频| 亚洲中文字幕一区二区三区有码在线看| 嫁个100分男人电影在线观看| 99视频精品全部免费 在线| 久久久久国内视频| 午夜福利在线观看免费完整高清在 | 男插女下体视频免费在线播放| 变态另类成人亚洲欧美熟女| 桃红色精品国产亚洲av| 全区人妻精品视频| 在现免费观看毛片| 精品人妻偷拍中文字幕| 97超级碰碰碰精品色视频在线观看| 波多野结衣高清无吗| 精品无人区乱码1区二区| 国产精品久久久久久亚洲av鲁大| 亚洲一区高清亚洲精品| 中文字幕人成人乱码亚洲影| 日本在线视频免费播放| 高清日韩中文字幕在线| 十八禁国产超污无遮挡网站| 久久人人爽人人爽人人片va | 亚洲综合色惰| 精品久久久久久久久久免费视频| 九色成人免费人妻av| 99久久精品一区二区三区| 国产精品三级大全| 韩国av一区二区三区四区| 欧美一区二区亚洲| 亚洲欧美日韩东京热| 99国产精品一区二区蜜桃av| 小说图片视频综合网站| 日韩亚洲欧美综合| 国产精品久久久久久亚洲av鲁大| 久久久国产成人精品二区| 美女免费视频网站| 亚洲久久久久久中文字幕| 搡女人真爽免费视频火全软件 | 国产亚洲av嫩草精品影院| 亚洲第一区二区三区不卡| 亚洲成av人片免费观看| 亚洲 国产 在线| 国产日本99.免费观看| 无遮挡黄片免费观看| 婷婷丁香在线五月| 丰满的人妻完整版| 国产色婷婷99| a级毛片a级免费在线| 在线观看舔阴道视频| 欧美性感艳星| 白带黄色成豆腐渣| 国产亚洲av嫩草精品影院| 国产精品一区二区性色av| 国产人妻一区二区三区在| 波多野结衣高清无吗| 很黄的视频免费| av中文乱码字幕在线| 亚洲精品456在线播放app | 亚洲精华国产精华精| 99视频精品全部免费 在线| 午夜精品在线福利| 在线观看av片永久免费下载| 女人被狂操c到高潮| 午夜福利欧美成人| 尤物成人国产欧美一区二区三区| 国产欧美日韩精品亚洲av| 午夜视频国产福利| 精品免费久久久久久久清纯| 久久国产精品影院| 麻豆国产av国片精品| 在线观看av片永久免费下载| 性欧美人与动物交配| 最新中文字幕久久久久| 久久久久久久亚洲中文字幕 | 久久精品国产亚洲av涩爱 | 亚洲精品亚洲一区二区| 亚洲第一欧美日韩一区二区三区| www.色视频.com| 久久久精品大字幕| 黄色视频,在线免费观看| 亚洲欧美日韩东京热| 国产精品久久视频播放| 99久久久亚洲精品蜜臀av| 99热6这里只有精品| 亚洲国产精品999在线| 亚洲美女搞黄在线观看 | 婷婷六月久久综合丁香| 怎么达到女性高潮| 琪琪午夜伦伦电影理论片6080| 窝窝影院91人妻| 99热只有精品国产| or卡值多少钱| 亚洲欧美日韩卡通动漫| 9191精品国产免费久久| 1000部很黄的大片| 桃红色精品国产亚洲av| 国产精品1区2区在线观看.| 69人妻影院| 欧美日本视频| 成年女人永久免费观看视频| 国内精品久久久久精免费| 宅男免费午夜| 色尼玛亚洲综合影院| 少妇丰满av| 女人被狂操c到高潮| 国产av在哪里看| 午夜激情欧美在线| 亚洲精品在线观看二区| 亚洲久久久久久中文字幕| 深夜精品福利| 搡老熟女国产l中国老女人| 免费人成视频x8x8入口观看| 亚洲精品在线美女| xxxwww97欧美| 国产蜜桃级精品一区二区三区| 免费在线观看日本一区| 国产伦人伦偷精品视频| 全区人妻精品视频| 日本撒尿小便嘘嘘汇集6| 少妇丰满av| 夜夜躁狠狠躁天天躁| 舔av片在线| 99久久九九国产精品国产免费| 久久九九热精品免费| av天堂在线播放| 好男人电影高清在线观看| 欧美丝袜亚洲另类 | 国产三级黄色录像| 日韩精品中文字幕看吧| 亚洲最大成人手机在线| 国产野战对白在线观看| 尤物成人国产欧美一区二区三区| 成人特级av手机在线观看| 久久久久国产精品人妻aⅴ院| 午夜福利视频1000在线观看| 免费看a级黄色片| 久久伊人香网站| 十八禁网站免费在线| 国产伦一二天堂av在线观看| 十八禁人妻一区二区| 国产精品久久久久久亚洲av鲁大| 欧美高清性xxxxhd video| 99在线人妻在线中文字幕| 国产探花极品一区二区| 国产亚洲av嫩草精品影院| 日本五十路高清| 直男gayav资源| 亚洲内射少妇av| 日韩国内少妇激情av| 午夜福利视频1000在线观看| 亚洲乱码一区二区免费版| 在线播放国产精品三级| 国产一区二区激情短视频| 欧美精品啪啪一区二区三区| 好男人在线观看高清免费视频| 精品久久久久久,| 日本在线视频免费播放| 90打野战视频偷拍视频| 免费无遮挡裸体视频| 精品熟女少妇八av免费久了| 一a级毛片在线观看| 亚洲在线观看片| 欧美日韩乱码在线| 不卡一级毛片| 国产成人欧美在线观看| 男女床上黄色一级片免费看| 久久精品国产亚洲av涩爱 | 无遮挡黄片免费观看| 观看免费一级毛片| 免费大片18禁| 国产一级毛片七仙女欲春2| 两人在一起打扑克的视频| 国产蜜桃级精品一区二区三区| 波多野结衣巨乳人妻| 色精品久久人妻99蜜桃| 欧美又色又爽又黄视频| 国产精品永久免费网站| 久久久久久久午夜电影| 在线观看66精品国产| 午夜激情欧美在线| 久久久久久久久大av| 国产熟女xx| www.www免费av| 日本与韩国留学比较| 最近最新中文字幕大全电影3| 18禁在线播放成人免费| 亚洲 欧美 日韩 在线 免费| 成年女人毛片免费观看观看9| 十八禁网站免费在线| 又黄又爽又免费观看的视频| 国产精品美女特级片免费视频播放器| 草草在线视频免费看| 午夜日韩欧美国产| 日韩欧美精品免费久久 | 级片在线观看| 免费看美女性在线毛片视频| 国产一区二区三区视频了| 女人被狂操c到高潮| 亚洲熟妇中文字幕五十中出| 国产精品一区二区三区四区免费观看 | 88av欧美| 国内精品一区二区在线观看| 国产精品美女特级片免费视频播放器| 久久婷婷人人爽人人干人人爱| 欧美高清成人免费视频www| 99riav亚洲国产免费| 精品久久久久久,| 看免费av毛片| 亚洲精品影视一区二区三区av| 99国产极品粉嫩在线观看| 十八禁国产超污无遮挡网站| 日日干狠狠操夜夜爽| 亚洲专区中文字幕在线| 国产在视频线在精品| 少妇裸体淫交视频免费看高清| 麻豆久久精品国产亚洲av| 校园春色视频在线观看| 九九热线精品视视频播放| 熟女电影av网| 男人狂女人下面高潮的视频| 亚洲精品粉嫩美女一区| 一级黄色大片毛片| 欧美3d第一页| 精品久久国产蜜桃| 欧美区成人在线视频| 久久草成人影院| 国产亚洲欧美在线一区二区| 国产老妇女一区| 男人狂女人下面高潮的视频| 有码 亚洲区| 日本成人三级电影网站| 欧美黄色淫秽网站| 波多野结衣高清作品| 国产乱人伦免费视频| 97超级碰碰碰精品色视频在线观看| 日本精品一区二区三区蜜桃| 又黄又爽又免费观看的视频| x7x7x7水蜜桃| 波多野结衣高清作品| 一个人看的www免费观看视频| 久久99热6这里只有精品| 国产av不卡久久| 亚洲av二区三区四区| 日本撒尿小便嘘嘘汇集6| 欧美精品国产亚洲| av福利片在线观看| 国产成人欧美在线观看| 国产亚洲欧美在线一区二区| 国产精品久久久久久久电影| 色av中文字幕| 黄色视频,在线免费观看| 悠悠久久av| 中文字幕人妻熟人妻熟丝袜美| 天美传媒精品一区二区| 91午夜精品亚洲一区二区三区 | 麻豆久久精品国产亚洲av| 丰满人妻熟妇乱又伦精品不卡|