• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Dense Spatial-Temporal Graph Convolutional Network Based on Lightweight OpenPose for Detecting Falls

    2023-12-12 15:49:22XiaoruiZhangQijianXieWeiSunYongjunRenandMithunMukherjee
    Computers Materials&Continua 2023年10期

    Xiaorui Zhang,Qijian Xie,Wei Sun,Yongjun Ren,2,3 and Mithun Mukherjee

    1School of Computer Science,Nanjing University of Information Science&Technology,Nanjing,210044,China

    2Wuxi Research Institute,Nanjing University of Information Science&Technology,Wuxi,214100,China

    3Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology(CICAEET),Nanjing University of Information Science&Technology,Nanjing,210044,China

    4School of Automation,Nanjing University of Information Science&Technology,Nanjing,210044,China

    5School of Artificial Intelligence,Nanjing University of Information Science&Technology,Nanjing,210044,China

    ABSTRACT Fall behavior is closely related to high mortality in the elderly,so fall detection becomes an important and urgent research area.However,the existing fall detection methods are difficult to be applied in daily life due to a large amount of calculation and poor detection accuracy.To solve the above problems,this paper proposes a dense spatial-temporal graph convolutional network based on lightweight OpenPose.Lightweight OpenPose uses MobileNet as a feature extraction network,and the prediction layer uses bottleneck-asymmetric structure,thus reducing the amount of the network.The bottleneck-asymmetrical structure compresses the number of input channels of feature maps by 1×1 convolution and replaces the 7×7 convolution structure with the asymmetric structure of 1×7 convolution,7×1 convolution,and 7×7 convolution in parallel.The spatial-temporal graph convolutional network divides the multi-layer convolution into dense blocks,and the convolutional layers in each dense block are connected,thus improving the feature transitivity,enhancing the network’s ability to extract features,thus improving the detection accuracy.Two representative datasets,Multiple Cameras Fall dataset(MCF),and Nanyang Technological University Red Green Blue+Depth Action Recognition dataset(NTU RGB+D),are selected for our experiments,among which NTU RGB+D has two evaluation benchmarks.The results show that the proposed model is superior to the current fall detection models.The accuracy of this network on the MCF dataset is 96.3%,and the accuracies on the two evaluation benchmarks of the NTU RGB+D dataset are 85.6%and 93.5%,respectively.

    KEYWORDS Fall detection;lightweight OpenPose;spatial-temporal graph convolutional network;dense blocks

    1 Introduction

    Falling,as the number one killer of elderly injury and death,has great harm to the health of the elderly [1].According to statistics,the incidence rate of falls among the elderly is 31.8%,and 32.5%of the falls require medical treatment and even endanger life[2].Falls are most common in the home environment[3],which threatens the elderly who live alone more severely than others.Given that the elderly living alone stay on the floor longer when they fall without the help of others,it is more prone to result in health problems such as dehydration,internal bleeding,and even death.Therefore,there is an urgent need for an effective automatic fall detection system that can help the elderly timely after a fall.

    Researchers have carried out a lot of research in this area and made a significant advance.Currently,many solutions for fall detection have been proposed.These solutions can be broadly classified into three categories:wearable non-visual sensors-based approaches,environmental sensorsbased approaches,and computer vision-based approaches.

    The method based on wearable usually collects human motion data using portable sensors such as embedded bracelets and belts,and then implements fall detection through Support Vector Machine(SVM)or set threshold[4–6].However,the method is likely to fail to collect reliable data,e.g.,people may forget to wear the device,especially for the elderly or people with dementia[7].The method based on environmental sensors is to detect whether a human falls by placing sensors in the detection area,in which commonly used sensors such as pressure sensors,infrared sensors,and sound sensors are included.The method does not require people to wear any equipment and has good comfort.However,this method has high equipment cost,and long-term exposure to infrared rays will lead to many adverse effects such as premature skin aging,pigment disorder,and eye damage.At the same time,the frequency of infrared rays will be affected by sunlight,which will affect the accuracy of detection.The method based on computer vision is to use some cameras to collect video/image information of the human body,extract human body features by image processing technology,and then analyze the motion state of the human body[8].

    Falls can generally be detected by spatial-temporal information[9],optical flow information[10],temporal features[11],and sequence of human skeletons[12].Among them,the sequences of human skeletons-based methods usually transmit important information that can be used to detect human action[13,14].At present,OpenPose is mostly used for the extraction of skeletal sequences.This model can realize the tracking of a human face,torso and limbs,and even fingers.It is not only suitable for single people but also for multiple people.However,OpenPose uses Visual Geometry Group(VGG-19) to extract features and uses multiple 7×7 convolutions in the prediction stage,resulting in a large number of parameters and calculations.It is difficult to meet the requirement for low delay and fast response in daily home scenes.Therefore,it is necessary to design a lightweight OpenPose.And the early methods of fall detection using skeletons are to use coordinates of skeletal joints to form feature vectors at each time step,and then analyze the temporal features.But these methods cannot explicitly exploit the spatial relationships between skeletal points.The spatial-temporal graph convolutional network(ST-GCN)[15],which is based on the sequence of skeletal graphs,can focus on the edges of skeletal joints with consistent connectivity and the connected edges of the same skeletal joints in a continuous time step,automatically capturing the spatial information and time dynamic information of the skeletal joints.Although the model has higher resolution,and contains more location information and detail information,it is prone to fall into gradient disappearance due to ignoring shallow features,which affects the detection results.

    Therefore,we design a dense spatial-temporal graph convolutional network (DST-GCN) based on lightweight OpenPose for fall detection.The lightweight OpenPose uses MobileNet to replace VGG-19 as the feature extraction network,and adopts a bottleneck-asymmetrical structure for the prediction layer.The bottleneck-asymmetrical design compresses the quantity of input channels of feature maps by 1×1 convolution,and replaces the 7×7 convolution structure with the asymmetric structure of 1×7 convolution,7×1 convolution,and 7×7 convolution in parallel.In this way,the amount of calculation of the OpenPose is reduced.And to improve the accuracy of model detection,the dense spatial-temporal graph convolution network divides the multi-layer spatial-temporal graph convolution layer in the model into two dense blocks.In each dense block,all spatial-temporal convolutional layer is connected with the previous spatial-temporal convolution layers among the dense block.In this way,the transitivity of features is strengthened,the network’s ability to extract features is strengthened,and the accuracy of human fall detection is improved.Also,a transition layer is added after each dense block,which can relieve the increased calculation problem resulting from the dense connection.

    In short,our contributions are summarized as follows:

    1.We design a lightweight OpenPose,which first extracts feature through the MobileNet,and then feeds the features into the prediction layer containing the bottleneck-asymmetrical structure to predict the key points of the skeleton.Our method reduces the calculation amount of OpenPose.

    2.We propose a dense graph convolutional model for extracting spatial-temporal features.Spatial-temporal graph convolutional layers are densely connected in each dense block to fully extract temporal and spatial features,improving the accuracy of human fall detection.

    3.Our method is verified on multiple datasets,experimental results show the proposed network has less computation on the premise of meeting the fall detection accuracy.

    The rest of the article will be organized as follows.Section 2 introduces the related work of this paper.Section 3 presents the model proposed in this paper in detail.Section 4 introduces the datasets and hardware used in the experiments,and presents the experimental results and analysis.Section 5 summarizes the content of this paper and gives directions for further research.

    2 Related Work

    In recent years,due to the elderly population increasing and the falls frequent,research on the fall detection system for the elderly has attracted extensive attention.This paper aims at improving the real-time and accuracy of fall detection by improvement of OpenPose and ST-GCN.The following will introduce our related work through OpenPose and ST-GCN.

    2.1 OpenPose

    The OpenPose model [16] proposed by Cao et al.is different from the traditional detection methods.It mainly uses Part Affinity Fields(PAF)[17]to perform a bottom-up method for human pose estimation.The structure of OpenPose can be divided into two parts.The role of the first part is to extract features through VGG-19.And the second part is the prediction layer,which is used to obtain the position of skeletal points through the feature.The prediction layer uses multiple stages to extract information on the position of the skeletal points,and the internal structure of the prediction layer uses multiple 7×7 convolution kernels in all stages except the first stage.

    The detection speed of OpenPose is significantly improved in comparison with the traditional networks,the network’s detection accuracy is high.The detected number of people in an image does not significantly affect the detection accuracy and speed of the network.Many researchers have done further research on OpenPose.Reference [18] combined OpenPose with a bidirectional long short-term memory (LSTM) to effectively improve the accuracy of human pose recognition in complex environments.Reference[19]extracted and marked the key points of the human body through OpenPose,and then defected falls through MobileNetV2.

    Among the above methods,most of the performance indicators of OpenPose in detection capability are good.However,OpenPose extracts feature through VGG-19,which has a large amount of calculation[20].Therefore,OpenPose is difficult to apply in practical scenarios.

    At present,many researchers have improved OpenPose.Reference [21] proposed a lightweight real-time human pose detection network that used ResNet 18 to extract features.Reference[22]used MobileNet to replace 12 convolution modules in VGG 19,and proposed a real-time detection of the human skeletal points network.

    At the same time,the prediction layer uses multiple 7×7 convolution kernels.Although a large convolution kernel can obtain a larger receptive field,it also causes a large amount of calculation burden.Therefore,this paper designs a bottleneck-asymmetric structure for the prediction layer based on using Mobilenet to replace VGG 19 to extract features.The bottleneck-asymmetrical structure compresses the number of input channels of feature maps by 1×1 convolution,and replaces the 7×7 convolution structure with the asymmetric structure of 1×7 convolution,7×1 convolution,and 7×7 convolution in a parallel manner,which reduces the amount of calculation of the OpenPose.See Section 3.2 of this article for details.

    2.2 ST-GCN

    Previous deep learning models mostly adopted traditional Convolutional Neural Network(CNN)and Long Short-Term Memory (LSTM) to learn spatial and temporal features,respectively [23,24].However,the traditional CNN cannot operate on topological maps,and LSTM is computationally intensive,making it difficult to train.To solve the above problems,ST-GCN first proposed a general graph-based framework to model human skeletons,jointly extracting spatial-temporal features from both temporal and spatial dimensions.ST-GCN contains 9 spatial-temporal convolutional layers,and each spatial-temporal convolutional layer contains a spatial convolutional layer,a temporal convolutional layer,and a residual structure.The skeleton sequence graph first obtains the spatial features through the spatial convolution layer,and then obtains the temporal features through temporal convolution.Finally,the residual mechanism is used to fuse the original input and spatial features and temporal features to form the output features of the spatial-temporal graph convolution layer.

    At present,many researchers have improved ST-GCN.Reference[25]divided the skeletal sequence into multiple parts (such as head and limbs) according to the body structure,input them into the spatial-temporal graph convolutional layers respectively,and finally merged them into the overall result.Reference [26] took the joint points and skeletons in the graph of skeletal sequence as independent inputs,and constructed a dual-stream spatial-temporal graph convolution network based on different inputs.Reference [27] used a lightweight displacement map operation instead of a convolution operation to obtain spatial-temporal features through spatial displacement maps and time displacement maps.Reference[28]proposed a multimodal feature fusion learning strategy,which uses spatial-temporal graph convolutional networks and one-dimensional (1D) convolution to generate two sets of spatial-temporal kinematic gait features from skeleton sequences.Reference[29]proposed an adaptive multi-level graph convolutional network that uses spatial convolutions to extract spatial features and multi-scale temporal convolutions to capture temporal features.The above algorithms based on ST-GCN have improved the accuracy of human action recognition to a certain extent.However,in the above methods,the correlation between each spatial-temporal graph convolutional layer and the feature transitivity are poor,which limits the capabilities of the spatial-temporal graph convolutional network.

    To solve the above problems,we design a dense spatial-temporal graph convolutional neural network.All spatial-temporal graph convolution layers are divided into two dense spatial-temporal graph convolution blocks,and the spatial-temporal graph convolution layer in each dense block is connected by a dense connection mechanism[30],which improves the communication between layers and strengthens the transitivity of features,making more efficient use of features.And in order to solve the problem of increased calculation caused by dense connection,a transition layer is added after each dense block.See Section 3.3 of this article for details.

    3 Proposed Method

    In this section,we design a Dense Spatial Temporal Graph Convolution Network (DST-GCN)based on lightweight OpenPose for fall detection.Next,we will introduce Architecture,Lightweight OpenPose,DST-GCN,and Classifier in detail.

    3.1 Architecture

    The network framework of our method is shown in Fig.1.Our method mainly includes lightweight OpenPose,DST-GCN,and a classifier.In our current research,we first use Lightweight OpenPose to obtain skeletal data from ordinary video data,and use the data to construct a spatial temporal skeleton sequence map as input of the dense spatial-temporal graph convolutional network.The multi-layer spatial-temporal graph convolutional network is used for convolution,and a higher-level feature map is gradually generated on the graph.Finally,the Softmax classifier is used to detect whether there is a fall behavior in the video.

    Figure 1:Model for detecting elderly falls

    First,the video is sent to lightweight OpenPose,and the skeletal sequence is obtained through MobileNet and prediction layers inside lightweight OpenPose.Then input the skeletal sequence into DST-GCN to get spatial-temporal features.The classifier detects whether is a fall behavior based on the obtained features.Next,we introduce Lightweight OpenPose,DST-GCN,and classifier one by one.

    3.2 Lightweight OpenPose

    As shown in Fig.2,the Lightweight OpenPose structure can be roughly divided into two parts.The first part is a feature extraction layer,and the second part is a dual-branch multi-stage prediction layer.

    Figure 2:Lightweight OpenPose structure

    We use MobileNet to extract features.To save spatial resolution and reuse backbone weights,we use dilation convolution and set the dilation parameter value to 2.And in order to improve the effect of the model,we only use a part of the layers of MobileNet,including from the first layer to the conv5_5 layer.

    OpenPose uses a large number of 7×7 convolution kernels in the prediction layer.The 7×7 convolution kernel can obtain a large receptive field,but it will also cause a large amount of calculation.To alleviate the problem,this paper adopts the bottleneck-asymmetric structure for each 7×7 convolution of the prediction layer,as shown in Fig.3.

    Figure 3:Bottleneck-asymmetric architecture

    First,the number of channels of the input features of the asymmetric convolution layer is compressed through 1×1 convolution;the compressed features are input into the asymmetric convolution for calculation.After the calculation and fusion are completed,the channel number of the output feature of the convolutional layer is restored by 1×1 convolution.

    Meanwhile,a single 7×7 convolution structure is replaced with a parallel structure of a 7×7 convolution,a 7×1 convolution,and a 1×7 convolution in the asymmetric convolution,which improves model performance without increasing computation.The parallel structure is mainly operated by Batch Normalization (BN) operation and branch fusion.BN is performed after each branch of the parallel structure,and then the outputs of the three branches are fused to get final output.

    When the neural network transmits information,there will be a problem of informational loss,we use the residual structure to add the original input feature to the output feature of the above-mentioned asymmetric-bottleneck structure.The residual structure solves the problem of informational loss by directly transmitting the input feature to the output.

    The aforementioned lightweight OpenPose is used to obtain the spatial-temporal skeletal sequence from the video,and then the spatial-temporal skeletal sequence is sent to DST-GCN to extract the spatial-temporal features.Finally,the features are sent to the classifier to detect whether there is falling behavior.

    3.3 DST-GCN

    To strengthen the transitivity of features and the ability of the model to extract features,we adopt a dense connection mechanism in the multi-layer spatial-temporal graph convolutional layer.The ninelayer spatial-temporal graph convolutional structure includes two dense blocks.The first five layers are one dense block,and the last four layers are one dense block.In each dense block,every layer of spatial-temporal graph convolution is connected with all layers ahead of itself to strengthen the transitivity of features and improve the reliability of human fall detection.

    The dense connection mechanism means that the input of thetlayer is not only related to the output of thet-1 layer,but also related to the output of every layer in the model.The output of thetlayer is denoted as:

    where [X0,X1,...,Xt-1] represents to stack the feature maps ofX0,X1,...,Xt-1in the channel dimension.Hrepresents nonlinear transformation.

    A dense connection is realized by cross-layer feature channel splicing.The splicing is to expand the depth of the channel,and increase the number of channels,resulting in an increased model calculation amount.To solve the above problems,a transition layer is designed behind each the dense block to control the model calculation amount.Reduce the number of channels by 1×1 convolution layer,and use an average pooling layer with two strides to halve the width and height of the feature map.Fig.4 shows the structure of dense-spatial temporal graph convolution layer,and Fig.5 shows the internal structure of each layer of spatial-temporal graph convolution.

    Figure 4:(Continued)

    Figure 4:The structure of dense spatial-temporal graph convolution layer

    Figure 5:The internal structure of each layer

    DST-GCN is used to extract the spatiotemporal features of the bone sequence,and the features are sent to the classifier for classification to detect whether there is a drop in the video.

    3.4 Classifier

    We use the Softmax classifier to output the result of fall detection.There are two fully connected layers in the classifier.The second fully connected layer does not use dropout and reduces the dimension to the number of categories,and finally outputs the classification result of falls.

    This paper uses the binary cross-entropy loss function,and adds the L2 regularization term to further avoid over-fitting of the model.We use stochastic gradient descent (SGD) to constrain each parameter update of the loss function to an appropriate size.The objective loss function including L2 regularization is:

    whereLrepresents the target the objective loss function,arepresents the sample index,mrepresents the total number of samples,?yarepresents the sample label,where the negative class recorded as 0,the positive class recorded as 1,yarepresents the predicted positive probability,andλ‖θ‖2represents a regular term of L2.

    4 Experiment

    Below we evaluate the method proposed in this article.First,we will verify our proposed method through ablation experiments and then compare our method with currently existing methods.We will introduce the datasets,evaluating indicators,implementation details,and experimental results from four aspects.

    4.1 Datasets

    We select two widely used fall datasets to complete the fall detection task,the two datasets are MCF[31]and NTU RGB+D[32].

    The MCF dataset contains 24 sets of video data,each captured from 8 ordinary cameras positioned at different angles.Among them,the first 22 sets of data include one or multiple instances of falling behavior,as well as daily activities such as walking,lying down,and some interfering actions like squatting and lying on a sofa.The last two sets of data(23rd and 24th)do not contain any falling behavior but consist of common daily activities.This dataset covers a variety of falling postures,including but not limited to frontal falls,lateral falls,and consecutive falls,enabling an effective evaluation of fall detection algorithms’performance.In this paper,we conducted separate training for the data from the 8 different camera angles.The data was split into training and testing sets at an approximate ratio of 7:3,where sets 8 to 24 were used as the training set,and sets 1 to 7 were used as the testing set.Fig.6 contains the key frames of the MCF dataset with different‘Fall’and‘No Fall’poses.

    Figure 6:Different frames of multi camera fall dataset(a)fall(b)no fall

    The NTU RGB+D is currently the largest indoor action recognition dataset with 60 action categories.In our research,we utilized a subset of samples from the NTU RGB+D dataset,specifically including A8 sitting down,A43 falling down,A80 squatting,A108 flipping down,and A111 walking categories.These behavior categories have ambiguous boundaries,and they exhibit some similarities in motion patterns and postures,making them prone to classification errors and therefore more demanding models.Among these,we selected all samples from the A43 falling down category and also sampled a portion of data from the remaining four categories.By combining these selected samples with the A43 falling down samples,we obtained a comprehensive dataset consisting of 948 falling samples and 1052 non-falling samples.The authors of this dataset proposed two evaluation benchmarks:CS(cross subject)benchmark and CV(cross view)benchmark.

    4.2 Evaluating Indicators

    We use floating point operations (FLOPs) and accuracy (Acc) to comprehensively evaluate our method.FLOPs are floating-point numbers,which are usually understood as the amount of calculation and can be used to measure the complexity of models.Acc represents the probability of correctly detecting a fall.The specific calculation formula is as follows:

    whereHis the height of the input feature map,Wis the width of the input feature map,Cinis the number of input channels,Kis the size of the convolution kernel,andCoutis the number of output channels.The experiment selected the fall action as a positive sample,and the rest of the action as a negative sample.TP(True Positive)in the formula represents a fall sample predicted by the model as fall;TN(True Negative)represents a non-falling sample predicted by the model as non-falling;PandNrepresent positive and negative samples,respectively.In the application of fall detection,the modelAccshould be as high as possible,because there may be serious consequences if it is not detected when the fall occurs.

    4.3 Implementation Details

    The equipment we used included a computer with an Intel Core i7-10870 CPU and 16 GB RAM,and a remote server consisting of two NVIDIA GeForce RTX 3090 GPUs,and we use Pytorch 1.11 to build our network model.

    In this paper,the batch size is set to 32 and trained by SD optimizer.At the same time,we set the momentum size to 0.9,set the weight decay size and the initial learning rate to 0.001,and after every 30 epochs of training,the learning rate becomes nine-tenths of the previous rate.The study was conducted for a total of 90 epochs in training.

    4.4 Experimental Results

    We divide the experiment into two parts:the ablation experiment and the comparative experiment.The results of these two experiments will be used to illustrate the effectiveness of our method respectively.

    4.4.1 Ablation Experiment

    In order to analyze the importance of each component to the entire model,we designed ablation experiments.Table 1 shows the comparison between the lightweight OpenPose model in this paper and the original OpenPose model in terms of the calculation amount,and Table 2 shows the comparison of the DST-GCN in this paper and the ST-GCN model.The detection accuracy and the amount of calculation have changed to different degrees,which proves the effectiveness of our proposed method.

    Table 1:Ablation experiment results of lightweight models

    Table 2:Ablation experiment results of dense connection mechanism

    The calculation amount of the lightweight OpenPose in this paper is 20457 million,and the accuracy of this network on the MCF dataset is 96.3%,and the accuracy on the two evaluation benchmarks of the NTU RGB+D dataset are 85.6%and 93.5%,respectively.First of all,we remove the lightweight processing of lightweight OpenPose,and use the original OpenPose without processing the feature extraction layer and prediction layer to process the output data.The calculation amount of the OpenPose model is 29124 million,which greatly increases the amount of calculation compared with the lightweight structure.

    Secondly,we remove the dense connection mechanism of spatial-temporal graph convolution layers,and use the original ST-GCN to detect whether there is falling behavior.The accuracy rates of this network on the MCF dataset and the NTU RGB+D dataset are 93.4%,82.1%,and 89.3%,all of which have different degrees of decline.In summary,the ablation experiment results show that our method can reduce the computational complexity of the model and improve the detection accuracy.

    4.4.2 Comparative Experiments

    Our method is characterized by less computation and higher accuracy for fall detection.Table 3 shows the amount of calculation of our method and the other two methods.

    Table 3:The comparison between our method and other methods

    The experimental results show that for the amount of calculation,our method outperforms the other two methods,and the amount of calculation of our method is 20457 million.This is because we use the MobileNet to replace VGG-19 as the feature extraction network,the MobileNet is a lightweight network with less amount of calculation.At the same time,we design a bottleneck-asymmetrical structure for the prediction layer,we use two 1×1 convolutions in the bottleneck-asymmetric structure to reduce the amount of calculation by reducing the number of channels in the feature map.

    We adopt ST-GCN [15],Deep progressive reinforcement learning for skeleton-based action recognition (DPRL+GCNN) [34],spatial reasoning and temporal stack learning (SRN-TSL) [35]as benchmarks to compare our method,and the results are shown in Table 4.To see the advantages of our method more clearly and intuitively,Fig.7 is made based on the data in Table 4.It can be found from the experimental results that the detection accuracy of our model is higher than that of the comparative methods.This is because we design a multi-layer spatial-temporal graph convolution structure that includes two dense blocks,every spatial-temporal graph convolution layer in each dense block is connected with all the spatial-temporal graph convolutions ahead of itself.It strengthens the transitivity of features and the model’s ability to extract features,and improves the accuracy of human fall detection.

    Table 4:The comparison between our method and other methods

    Figure 7:The comparison between our method and other methods

    5 Conclusion

    In this paper,we propose a dense spatial-temporal graph convolutional network based on lightweight OpenPose.This paper improves the OpenPose model,replaces the feature extraction network,and improves the structure of the prediction layer to reduce the calculation of the model.And this paper adopts a dense connection mechanism for a spatial-temporal graph convolutional network,which strengthens the transitivity of features and the network’s ability to extract features,and the accuracy of human fall detection is improved.

    In the future,we will further study from the following directions: We will include the model’s robustness,fault tolerance,and flexibility in our considerations.Robustness refers to the model’s resistance to perturbations and disturbances in the input data.A highly robust model can maintain good performance even when facing noise,outliers,or incomplete data,and it is less affected by such disturbances.A model with high robustness demonstrates better generalization ability in the real world,allowing it to handle diverse data situations.Fault tolerance refers to the model’s tolerance towards errors.A model with high fault tolerance can continue to function properly and provide reasonable output results when facing errors,defects,or poor-quality data.Fault tolerance indicates the model’s ability to adapt to exceptional situations and ensures system stability and reliability.Flexibility refers to the model’s adaptability to different types of data and various problems.A highly flexible model can accommodate diverse data distributions and problems,allowing it to fit different types of samples flexibly.In the design and selection of the model,it is essential to consider these properties.An ideal model should possess robustness and fault tolerance,maintaining stable performance when facing disturbances,errors,and variations.Simultaneously,it should demonstrate a certain level of flexibility,enabling it to adapt flexibly to diverse data and problem scenarios.Such a model can better handle uncertainties and changes in practical applications,exhibiting superior adaptability and reliability.

    Acknowledgement:We are grateful to Nanjing University of Information Science and Technology for providing a research environment and computing equipment.

    Funding Statement:This study was supported,in part,by the National Nature Science Foundation of China under Grant Numbers 62272236,62 376128;in part,by the Natural Science Foundation of Jiangsu Province under Grant Numbers BK20201136,BK20191401.

    Author Contributions:Study conception and design:X.Z.,Q.X.;data collection:Q.X.;analysis and interpretation of results:X.Z.,Q.X.;draft manuscript preparation:X.Z.,Q.X.,W.S.and Y.R.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:All datasets and materials are publicly available.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产精品影院久久| 国产成人av教育| 女人精品久久久久毛片| 丁香六月欧美| 一区二区三区高清视频在线| 国产又爽黄色视频| 国产高清有码在线观看视频 | 此物有八面人人有两片| 久久久水蜜桃国产精品网| 一夜夜www| 波多野结衣一区麻豆| 自拍欧美九色日韩亚洲蝌蚪91| 可以在线观看毛片的网站| 亚洲色图av天堂| 亚洲成人精品中文字幕电影| 极品教师在线免费播放| 国内精品久久久久久久电影| 人妻丰满熟妇av一区二区三区| 亚洲熟女毛片儿| 午夜福利一区二区在线看| 亚洲av熟女| 十分钟在线观看高清视频www| 免费一级毛片在线播放高清视频 | 国产精华一区二区三区| 国产aⅴ精品一区二区三区波| 不卡av一区二区三区| 国产欧美日韩一区二区三区在线| av有码第一页| 香蕉国产在线看| 正在播放国产对白刺激| 亚洲欧美激情在线| 国产一卡二卡三卡精品| 亚洲精品久久国产高清桃花| 伦理电影免费视频| 日本黄色视频三级网站网址| 亚洲第一青青草原| 亚洲色图 男人天堂 中文字幕| 色尼玛亚洲综合影院| av电影中文网址| 欧美一区二区精品小视频在线| 久久精品国产亚洲av香蕉五月| 搡老妇女老女人老熟妇| 国产1区2区3区精品| 女同久久另类99精品国产91| 久久中文看片网| 国产精品久久视频播放| 久久久久精品国产欧美久久久| www日本在线高清视频| 久久国产乱子伦精品免费另类| 亚洲无线在线观看| 亚洲免费av在线视频| 国产精品电影一区二区三区| 国产精品久久久久久亚洲av鲁大| 91av网站免费观看| 可以免费在线观看a视频的电影网站| 国产精品一区二区免费欧美| 999久久久精品免费观看国产| 亚洲第一青青草原| 在线观看日韩欧美| 亚洲专区字幕在线| 久久狼人影院| 日韩av在线大香蕉| 久久国产精品人妻蜜桃| 黄色 视频免费看| 亚洲久久久国产精品| 亚洲五月婷婷丁香| 久久人人爽av亚洲精品天堂| 亚洲精品中文字幕在线视频| 精品久久久久久久毛片微露脸| 精品一区二区三区四区五区乱码| 欧美 亚洲 国产 日韩一| 国产男靠女视频免费网站| 午夜福利免费观看在线| www国产在线视频色| 无限看片的www在线观看| 久久国产精品男人的天堂亚洲| 日日干狠狠操夜夜爽| www国产在线视频色| 欧美绝顶高潮抽搐喷水| 国产在线观看jvid| 色播亚洲综合网| 一卡2卡三卡四卡精品乱码亚洲| 日韩精品中文字幕看吧| 91麻豆精品激情在线观看国产| 久久精品91无色码中文字幕| 人妻丰满熟妇av一区二区三区| 女人被躁到高潮嗷嗷叫费观| 欧美日韩中文字幕国产精品一区二区三区 | 欧美日韩一级在线毛片| 麻豆一二三区av精品| 91字幕亚洲| 久久精品人人爽人人爽视色| www国产在线视频色| 欧美 亚洲 国产 日韩一| 国产不卡一卡二| 中文字幕另类日韩欧美亚洲嫩草| 亚洲avbb在线观看| 在线观看66精品国产| avwww免费| 日本 欧美在线| 国产片内射在线| 麻豆成人av在线观看| 丰满的人妻完整版| 91国产中文字幕| 一级毛片女人18水好多| 成人国产一区最新在线观看| 中文字幕另类日韩欧美亚洲嫩草| 91大片在线观看| 国产亚洲精品久久久久久毛片| 亚洲av美国av| 中文字幕人妻熟女乱码| 国产伦一二天堂av在线观看| 黄色女人牲交| 如日韩欧美国产精品一区二区三区| 啦啦啦韩国在线观看视频| 少妇粗大呻吟视频| 男女之事视频高清在线观看| 欧美中文综合在线视频| 亚洲成人免费电影在线观看| 久久久水蜜桃国产精品网| 国产麻豆成人av免费视频| 久久精品国产综合久久久| 一级a爱片免费观看的视频| 黑丝袜美女国产一区| 日本vs欧美在线观看视频| 国产伦一二天堂av在线观看| 他把我摸到了高潮在线观看| 女警被强在线播放| 亚洲国产精品合色在线| 国产欧美日韩一区二区三| 午夜免费激情av| 国内精品久久久久精免费| 国产欧美日韩一区二区三区在线| 国产亚洲精品久久久久5区| www国产在线视频色| 亚洲av五月六月丁香网| 99精品欧美一区二区三区四区| 日本一区二区免费在线视频| 国产精品乱码一区二三区的特点 | 丁香六月欧美| 久久国产乱子伦精品免费另类| 久久亚洲真实| 中文字幕久久专区| 啦啦啦免费观看视频1| 一级片免费观看大全| 色尼玛亚洲综合影院| 黄网站色视频无遮挡免费观看| 99国产精品免费福利视频| 18禁黄网站禁片午夜丰满| 一级毛片精品| 欧美av亚洲av综合av国产av| 给我免费播放毛片高清在线观看| 波多野结衣巨乳人妻| 久久人人精品亚洲av| 免费人成视频x8x8入口观看| 9色porny在线观看| 涩涩av久久男人的天堂| 一级,二级,三级黄色视频| xxx96com| av天堂在线播放| 亚洲av第一区精品v没综合| 成人三级做爰电影| 不卡一级毛片| 久久中文字幕人妻熟女| 亚洲天堂国产精品一区在线| 国产精品影院久久| 老鸭窝网址在线观看| 50天的宝宝边吃奶边哭怎么回事| 亚洲欧美日韩无卡精品| 在线观看日韩欧美| 在线观看66精品国产| 亚洲中文av在线| 久久久久国产一级毛片高清牌| 精品不卡国产一区二区三区| 久久久国产精品麻豆| 成人亚洲精品av一区二区| 亚洲精品在线美女| 日本a在线网址| 久久人妻av系列| 亚洲av成人不卡在线观看播放网| 18美女黄网站色大片免费观看| 国内精品久久久久精免费| 国产主播在线观看一区二区| 嫁个100分男人电影在线观看| 99国产精品99久久久久| 精品乱码久久久久久99久播| 亚洲免费av在线视频| 一级a爱视频在线免费观看| 黄色视频不卡| 极品人妻少妇av视频| 亚洲一区二区三区不卡视频| 国产精品二区激情视频| 99久久综合精品五月天人人| 免费在线观看影片大全网站| 91av网站免费观看| 看片在线看免费视频| 亚洲欧美日韩无卡精品| 91国产中文字幕| 午夜日韩欧美国产| 国产精品二区激情视频| 色尼玛亚洲综合影院| 亚洲 国产 在线| 欧美日韩亚洲综合一区二区三区_| 级片在线观看| 亚洲欧美日韩无卡精品| 日韩高清综合在线| 不卡av一区二区三区| 午夜福利,免费看| 99久久精品国产亚洲精品| 国产亚洲av高清不卡| 色综合站精品国产| 亚洲天堂国产精品一区在线| 国产精品久久久av美女十八| 如日韩欧美国产精品一区二区三区| 免费在线观看视频国产中文字幕亚洲| 他把我摸到了高潮在线观看| 中出人妻视频一区二区| 国产一级毛片七仙女欲春2 | 国产精品 欧美亚洲| 欧美一区二区精品小视频在线| 久久久久久久精品吃奶| 人妻丰满熟妇av一区二区三区| 色在线成人网| 日韩有码中文字幕| 日韩中文字幕欧美一区二区| 亚洲黑人精品在线| 91在线观看av| 亚洲精品粉嫩美女一区| 天堂影院成人在线观看| 美女大奶头视频| 欧美色视频一区免费| 99国产精品一区二区三区| 老汉色av国产亚洲站长工具| 色尼玛亚洲综合影院| 亚洲少妇的诱惑av| 久9热在线精品视频| 老司机深夜福利视频在线观看| 深夜精品福利| 欧美中文综合在线视频| 成人三级黄色视频| 三级毛片av免费| 国产亚洲av高清不卡| 国产精品精品国产色婷婷| 亚洲情色 制服丝袜| 91成人精品电影| 国产色视频综合| 欧美性长视频在线观看| 国产精品自产拍在线观看55亚洲| 亚洲情色 制服丝袜| 亚洲国产毛片av蜜桃av| 在线观看日韩欧美| 在线观看www视频免费| 男女之事视频高清在线观看| 丝袜美足系列| 久久香蕉国产精品| √禁漫天堂资源中文www| 亚洲精品在线观看二区| 麻豆国产av国片精品| 老熟妇乱子伦视频在线观看| 亚洲人成77777在线视频| xxx96com| 欧美黑人精品巨大| 老司机福利观看| 99久久久亚洲精品蜜臀av| 老鸭窝网址在线观看| 国产亚洲精品av在线| 国产精品,欧美在线| 精品久久蜜臀av无| 日本五十路高清| 又大又爽又粗| 一边摸一边抽搐一进一出视频| 女性生殖器流出的白浆| 精品一区二区三区av网在线观看| 久久亚洲精品不卡| 精品久久蜜臀av无| 久久人人97超碰香蕉20202| 又大又爽又粗| 国产精品一区二区精品视频观看| 最近最新中文字幕大全免费视频| 国产亚洲精品一区二区www| 亚洲熟妇中文字幕五十中出| 久久影院123| 国产xxxxx性猛交| 日本一区二区免费在线视频| 97人妻天天添夜夜摸| 两个人免费观看高清视频| 99国产精品一区二区蜜桃av| www.精华液| 精品一区二区三区视频在线观看免费| 久久精品影院6| 精品一区二区三区四区五区乱码| 老熟妇乱子伦视频在线观看| 国产精品,欧美在线| 精品熟女少妇八av免费久了| 波多野结衣巨乳人妻| 国产精品影院久久| 午夜亚洲福利在线播放| svipshipincom国产片| 亚洲中文字幕一区二区三区有码在线看 | 久久狼人影院| 午夜福利在线观看吧| 国产精品秋霞免费鲁丝片| 好看av亚洲va欧美ⅴa在| 悠悠久久av| 深夜精品福利| 三级毛片av免费| 亚洲va日本ⅴa欧美va伊人久久| 一区二区日韩欧美中文字幕| 别揉我奶头~嗯~啊~动态视频| 韩国av一区二区三区四区| 婷婷六月久久综合丁香| tocl精华| 久热爱精品视频在线9| 精品免费久久久久久久清纯| 亚洲一区二区三区不卡视频| 精品国产乱码久久久久久男人| 可以在线观看毛片的网站| 亚洲第一青青草原| 久久影院123| 午夜a级毛片| 欧美黄色淫秽网站| 亚洲国产精品sss在线观看| 日韩有码中文字幕| 黄色 视频免费看| 青草久久国产| 精品国产超薄肉色丝袜足j| 视频在线观看一区二区三区| 国产精品99久久99久久久不卡| 亚洲精品中文字幕一二三四区| 亚洲国产欧美网| 日韩三级视频一区二区三区| 免费观看精品视频网站| 亚洲成人国产一区在线观看| 久久热在线av| 精品乱码久久久久久99久播| 91老司机精品| 免费高清视频大片| 亚洲在线自拍视频| 天堂√8在线中文| 久久影院123| 两个人免费观看高清视频| 熟女少妇亚洲综合色aaa.| 国产三级在线视频| 性欧美人与动物交配| 久久久久亚洲av毛片大全| 日韩有码中文字幕| 亚洲专区中文字幕在线| 国产一区二区激情短视频| 欧美丝袜亚洲另类 | 窝窝影院91人妻| 巨乳人妻的诱惑在线观看| 亚洲aⅴ乱码一区二区在线播放 | 久久精品影院6| 亚洲一区高清亚洲精品| 精品久久久精品久久久| 日本 av在线| 很黄的视频免费| 777久久人妻少妇嫩草av网站| 美女大奶头视频| 国产熟女午夜一区二区三区| 欧美国产精品va在线观看不卡| 99热只有精品国产| e午夜精品久久久久久久| 又黄又爽又免费观看的视频| 91麻豆精品激情在线观看国产| 欧美人与性动交α欧美精品济南到| 午夜福利影视在线免费观看| 国产又爽黄色视频| av网站免费在线观看视频| 成人亚洲精品av一区二区| 黑人巨大精品欧美一区二区mp4| 国产一区二区在线av高清观看| 色综合欧美亚洲国产小说| 亚洲成人免费电影在线观看| 日韩av在线大香蕉| 精品久久久久久久人妻蜜臀av | 18禁美女被吸乳视频| 国产人伦9x9x在线观看| 女人高潮潮喷娇喘18禁视频| 久久久久国内视频| 国产亚洲精品av在线| 久久天堂一区二区三区四区| 精品一区二区三区四区五区乱码| 在线观看66精品国产| 亚洲性夜色夜夜综合| 国内久久婷婷六月综合欲色啪| 国产野战对白在线观看| 久久久久久亚洲精品国产蜜桃av| 丁香欧美五月| 欧洲精品卡2卡3卡4卡5卡区| 久久精品国产亚洲av香蕉五月| www.熟女人妻精品国产| 欧美激情 高清一区二区三区| 国产在线精品亚洲第一网站| 国产精品二区激情视频| 日韩大尺度精品在线看网址 | 无限看片的www在线观看| 大陆偷拍与自拍| 亚洲午夜理论影院| 国产亚洲精品久久久久5区| 国产免费男女视频| 亚洲黑人精品在线| 国产精品亚洲一级av第二区| 麻豆国产av国片精品| 国产高清视频在线播放一区| 非洲黑人性xxxx精品又粗又长| 91字幕亚洲| 极品人妻少妇av视频| 人人妻人人澡欧美一区二区 | 午夜免费观看网址| 50天的宝宝边吃奶边哭怎么回事| 99在线人妻在线中文字幕| 国产激情欧美一区二区| 国产精品爽爽va在线观看网站 | 国产精品98久久久久久宅男小说| a在线观看视频网站| 激情视频va一区二区三区| 精品欧美国产一区二区三| 亚洲中文字幕一区二区三区有码在线看 | 日韩精品中文字幕看吧| 一级,二级,三级黄色视频| 国产亚洲av高清不卡| 后天国语完整版免费观看| 午夜视频精品福利| 老司机午夜十八禁免费视频| 精品久久久久久,| 亚洲精品国产一区二区精华液| 久热这里只有精品99| 亚洲精品国产色婷婷电影| 亚洲第一青青草原| 一级毛片高清免费大全| 中文字幕另类日韩欧美亚洲嫩草| 亚洲欧美精品综合一区二区三区| 给我免费播放毛片高清在线观看| 日韩欧美国产一区二区入口| 可以免费在线观看a视频的电影网站| 精品一品国产午夜福利视频| 久久国产精品人妻蜜桃| 亚洲无线在线观看| 欧美性长视频在线观看| 18禁裸乳无遮挡免费网站照片 | 久久久久亚洲av毛片大全| 最新在线观看一区二区三区| 欧美黑人精品巨大| 中文字幕久久专区| 成人三级黄色视频| 国产单亲对白刺激| 国产av一区二区精品久久| av天堂久久9| 天堂√8在线中文| 一区二区三区国产精品乱码| 亚洲精品中文字幕一二三四区| 午夜激情av网站| 给我免费播放毛片高清在线观看| 自线自在国产av| 精品久久蜜臀av无| 亚洲激情在线av| 中文字幕久久专区| 国产av精品麻豆| 成年版毛片免费区| 香蕉国产在线看| 乱人伦中国视频| 老司机午夜福利在线观看视频| 精品人妻在线不人妻| 制服人妻中文乱码| 国产精品爽爽va在线观看网站 | 黄网站色视频无遮挡免费观看| 久久久国产成人免费| 久久久久亚洲av毛片大全| 成人三级做爰电影| 欧美黄色片欧美黄色片| 丝袜美足系列| 高清黄色对白视频在线免费看| 亚洲中文字幕日韩| 欧美日韩福利视频一区二区| 亚洲自拍偷在线| 欧美国产日韩亚洲一区| 99在线人妻在线中文字幕| av电影中文网址| 夜夜爽天天搞| 成人av一区二区三区在线看| 国产亚洲欧美98| 亚洲片人在线观看| 国内精品久久久久精免费| 国产精品精品国产色婷婷| 亚洲午夜精品一区,二区,三区| 免费看十八禁软件| 天天躁夜夜躁狠狠躁躁| 欧美日韩亚洲国产一区二区在线观看| 免费在线观看完整版高清| 久久 成人 亚洲| 两个人视频免费观看高清| 在线观看免费视频日本深夜| 久久久精品国产亚洲av高清涩受| av在线播放免费不卡| 精品久久久久久久久久免费视频| 精品久久久久久久毛片微露脸| 成年版毛片免费区| 69精品国产乱码久久久| 天天躁狠狠躁夜夜躁狠狠躁| 国产精品一区二区在线不卡| 91成年电影在线观看| 757午夜福利合集在线观看| 久久九九热精品免费| 国产精品影院久久| 亚洲精品国产色婷婷电影| 国产精品九九99| 日韩精品中文字幕看吧| 精品欧美国产一区二区三| 波多野结衣av一区二区av| 伦理电影免费视频| 天天添夜夜摸| 九色亚洲精品在线播放| 三级毛片av免费| avwww免费| 国产激情欧美一区二区| 最近最新中文字幕大全免费视频| 身体一侧抽搐| 国产亚洲精品第一综合不卡| 中文字幕最新亚洲高清| 天天添夜夜摸| 亚洲三区欧美一区| 两人在一起打扑克的视频| 国产蜜桃级精品一区二区三区| 可以在线观看毛片的网站| 亚洲一码二码三码区别大吗| 一边摸一边做爽爽视频免费| 亚洲成a人片在线一区二区| 午夜福利一区二区在线看| 精品高清国产在线一区| 亚洲aⅴ乱码一区二区在线播放 | 女同久久另类99精品国产91| 国产精品秋霞免费鲁丝片| av视频免费观看在线观看| 久久国产亚洲av麻豆专区| 久久久久久久久久久久大奶| 纯流量卡能插随身wifi吗| 国产精品久久电影中文字幕| 亚洲精品粉嫩美女一区| 欧美精品亚洲一区二区| 美女 人体艺术 gogo| av在线播放免费不卡| 国产成人免费无遮挡视频| cao死你这个sao货| 自线自在国产av| 午夜成年电影在线免费观看| 久久精品人人爽人人爽视色| 又紧又爽又黄一区二区| 很黄的视频免费| 国产91精品成人一区二区三区| 久久青草综合色| 亚洲七黄色美女视频| 久久久久久免费高清国产稀缺| 可以免费在线观看a视频的电影网站| 亚洲久久久国产精品| 免费在线观看视频国产中文字幕亚洲| 国产欧美日韩一区二区精品| 99re在线观看精品视频| 真人做人爱边吃奶动态| 在线观看午夜福利视频| 人人妻,人人澡人人爽秒播| 制服人妻中文乱码| 免费看美女性在线毛片视频| 欧美日韩黄片免| 成熟少妇高潮喷水视频| 久久香蕉国产精品| 一级黄色大片毛片| 桃色一区二区三区在线观看| 日韩欧美三级三区| 欧美日韩瑟瑟在线播放| 99精品久久久久人妻精品| 久热这里只有精品99| 丝袜美足系列| 禁无遮挡网站| 国产不卡一卡二| 久久 成人 亚洲| 亚洲精品久久国产高清桃花| 日韩成人在线观看一区二区三区| 50天的宝宝边吃奶边哭怎么回事| 日韩av在线大香蕉| 日韩精品青青久久久久久| 午夜久久久久精精品| 黄频高清免费视频| 亚洲精品中文字幕在线视频| www.精华液| 免费高清在线观看日韩| 国产99白浆流出| 国产亚洲欧美精品永久| 视频区欧美日本亚洲| 黄网站色视频无遮挡免费观看| 亚洲欧美精品综合久久99| 欧美激情极品国产一区二区三区| 国语自产精品视频在线第100页| 国产成人av教育| 日韩欧美三级三区| 美女国产高潮福利片在线看| 三级毛片av免费| 人人妻人人爽人人添夜夜欢视频| 亚洲午夜精品一区,二区,三区| 99久久国产精品久久久| 欧美黄色片欧美黄色片| 欧美日韩瑟瑟在线播放| 国产成人一区二区三区免费视频网站| 亚洲,欧美精品.| 欧美性长视频在线观看| 热99re8久久精品国产| 宅男免费午夜| 美女国产高潮福利片在线看| 日韩一卡2卡3卡4卡2021年| 宅男免费午夜| 激情在线观看视频在线高清| 国内精品久久久久久久电影| 美女高潮喷水抽搐中文字幕| 777久久人妻少妇嫩草av网站| 97超级碰碰碰精品色视频在线观看|