• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An Efficient Approach to Escalate the Speed of Training Convolution Neural Networks

    2024-03-11 06:29:00PabithaAnushaJayasimhan
    China Communications 2024年2期

    P Pabitha,Anusha Jayasimhan

    Department of Computer Technology,Anna University,Madras Institute of Technology Campus,Chennai 600044,India

    Abstract: Deep neural networks excel at image identification and computer vision applications such as visual product search,facial recognition,medical image analysis,object detection,semantic segmentation,instance segmentation,and many others.In image and video recognition applications,convolutional neural networks (CNNs) are widely employed.These networks provide better performance but at a higher cost of computation.With the advent of big data,the growing scale of datasets has made processing and model training a time-consuming operation,resulting in longer training times.Moreover,these large scale datasets contain redundant data points that have minimum impact on the final outcome of the model.To address these issues,an accelerated CNN system is proposed for speeding up training by eliminating the noncritical data points during training alongwith a model compression method.Furthermore,the identification of the critical input data is performed by aggregating the data points at two levels of granularity which are used for evaluating the impact on the model output.Extensive experiments are conducted using the proposed method on CIFAR-10 dataset on ResNet models giving a 40%reduction in number of FLOPs with a degradation of just 0.11%accuracy.

    Keywords: CNN;deep learning;image classification;model compression

    I.INTRODUCTION

    In the last decade,the massive generation of data through the internet and smart devices has led to the ever increasing need of data analysis.Digital data is also expanding exponentially and is widely available;however,using conventional software tools and technology,it is difficult,if not impossible,to manage and analyze this data.Not only the amount of data being processed each day ranges upto several petabytes but within the next few years the total data on Earth will approximately be in trillions of gigabytes [1-3].Deep neural networks have achieved state-of-the-art performance in a number of fields such as computer vision when compared to earlier techniques based on human created visual qualities.Large-scale datasets,cutting edge contemporary Graphical Processing Units(GPUs),and novel network topologies enable the creation of previously unimaginable large CNN models.Various neural network models,such as AlexNet,VGGNet,and ResNets,have grown from 8 layers to more than 100 layers.

    While this massive amount of data has the potential to revolutionize every part of our civilization,extracting useful knowledge from it is not an easy undertaking.The massive and rapidly growing body of data concealed in vast quantities of non-traditional data,calls for the creation of new technologies as well as the close cooperation of interdisciplinary teams.While deep learning has proven to be effective in analyzing large or heavy datasets,they come with a high processing cost and memory requirements,which makes them a major problem for devices with limited resources.An embedded device,an IoT device,or a mobile phone all have restricted resources,making it challenging to install a standard deep model despite its broad popularity.For instance,the VGG-16 model uses more than 500 MB of storage and has 138.34 million parameters,requiring 30.94 billion float point operations to identify a single image.A model of this size can easily surpass the computational capacity of small devices.

    The use of neural networks like CNNs in realworld applications is hampered by several factors.1) Model size: CNNs’ strong representation capacity stems from their millions of learnable parameters.Those parameters,as well as network structure data,must be saved on disk and read into memory during inference.For instance,storing a conventional CNN trained on ImageNet requires more than 300MB of capacity,placing a heavy burden on the capabilities of embedded systems.2)Run-time memory: CNN intermediary processes may consume more memory during inference time than storing the model parameters,even with batch size 1.This isn’t an issue for highend GPUs,but it’s prohibitively expensive for many low-power applications.3)Number of computing operations: On high-resolution photos,convolution procedures are computationally demanding.On a mobile device,a big CNN could take several minutes to process a single image,making it unsuitable for realworld applications.

    As a result,a model training approach that addresses the scalability of vast datasets and model parameters while also resolving the critical issues when distributing larger models on low-resource devices is required.The following are the primary contributions of this work.

    ? Critical dataset identification is a pre-training step that removes redundant data points from a large dataset to reduce the processing time.During the iterative training phase,the significance and criticality of certain data instances are identified,with the goal of retaining only critical data and eliminating redundant data points.A granularity based approach is used to group the redundant data points and each group is represented by a single aggregated data point.The unimportant aggregated points are eliminated based on their effect value.

    ? As discussed before,model compression has become the need of the hour in order to train and deploy very deep CNN models on end devices.After the removal of redundant data elements,the efficiency of training CNN models is improvised further by model compression.The technique used to perform filter pruning maintains the model capacity and speeds up model training and inference with negligible performance impact.

    ? The effectiveness and efficiency of the suggested two-fold acceleration and compression approach will be demonstrated through experiments on CIFAR-10 dataset.

    The rest of the paper is organized as follows.In Section II a brief look at the related work in the field of neural network acceleration and model compression strategies has been presented.The proposed methodology is described in depth in Section III.In Section IV,the datasets used in this study as well as the specifics of the implementation are discussed.Section V discusses the results and analysis.Section VI summarises the conclusion and other discussions.

    II.RELATED WORK

    2.1 Removal of Redundant Data Points

    In the past,many researchers have devised techniques for the removal of redundant or unimportant data points from a large dataset for speeding up the convergence of a model.[4]prove that there exists atleast a 10%redundancy in ImageNet and CIFAR-10 datasets.By applying a hierarchical clustering technique,the redundant samples are removed with no major impact on the test accuracy.[5] and [6] conducted experiments to prove that different training examples play a crucial role in the convergence of training parameters at different phases of the training process.To speed up the convergence of the model,[7] dynamically filter out the training data from a given mini-batch.By using reinforcement learning,the algorithm learns to achieve the maximum validation accuracy in the least possible time with just half the number of original training examples.It learns a policy function to select the best training examples from a given mini-batch.[8]implement an importance sampling scheme which identifies only those samples that cause the maximum amount of change in the update of parameters.By reducing the variance on the gradient-norm,the method achieves faster convergence and lesser loss for image classification as well as sequence modeling problems.

    2.2 Model Compression

    The attention-based pruning strategy suggested by Haceneet al.[9]turns a convolution layer into a shift layer.For every channel,just one weight remains at the end which is identified as the most important connection.This results in improvements in memory usage and computation time while preserving high accuracy at the expense of needing more parameters for training but fewer for inference.H.Tessieret al.invented Selective Weight Decay,which carries out effective and continuous pruning during the training [10].It takes advantage of weight decay in order to attain network sparsity.It eliminates the need for fine-tuning once the network has been trimmed.It can be coupled with any pruning criteria or pruning structure,providing a framework for many different permutations.P.Jorgeet al.[11]implement pruning at initiation,a technique that avoids fine-tuning and precludes any changes to architecture during or after training.A novel approach was devised to calculate connection sensitivity postpruning,alongside the development of two approximation techniques for gradually enhancing connection sensitivity.These methods enable the pruning of networks at initialization,facilitating the transformation of initially unimportant parameters into crucial ones as the pruning process progresses.When pruning at high sparsity levels,this technique performs better,but when pruning at moderate sparsity levels,it performs worse.

    N.Leeet al.[12]execute a new method called SNIP(Single Shot Network Pruning based on Connection Sensitivity),which is based on connection sensitivity.It prunes irrelevant connections for a specific task in a single step prior to training and works with a wide range of neural network models.They prove that the connection sensitivity measure is particularly beneficial in diagnosing significant connections in the network from a totally untrained network because this results in incredibly sparse models.Rendaet al.[13]presented learning rate rewinding as a means of retraining after pruning instead of fine tuning.The network is retrained from scratch after pruning,unlike fine-tuning,which is done at the slowest learning rate,this involves training the unpruned weights using the same learning-rate pattern as training.This retraining has been demonstrated to produce greater results than fine-tuning,albeit at a much higher cost.ThiNet,a framework described by Luoet al.[14],focuses on developing more compact and compressed CNN models,resulting in acceleration.It is a filter pruning method,where the output channels of layer i+1 decide and guide the removal of filters of layer i.AutoPrune is an autonomous network pruning model created by Xiaoet al.[15]to eliminate network redundancies for easier deployment.Rather than using original weights,it prunes the system by optimizing a set of trainable supplementary parameters.The auxiliary parameters are less susceptible to hyperparameters and more resistant to noise throughout training,and the gradient update algorithms for these supplementary parameters are designed to maintain them consistent with pruning tasks,removing network redundancy and improving recoverability automatically.

    AutoCompress is an automatic structural weight pruning system presented by N.Liuet al.[16],which compresses models to speed up inference.It uses an automated technique for determining hyperparameters,such as the per-layer weight pruning rate and structured weight pruning schemes.The goal is to reduce the number of parameters or FLOPs as much as possible while maintaining precision.H.Liet al.[17]implement a CNN compression strategy in which they trim the filters from the CNN that have a minimal impact on the output accuracy.They prune the filters by determining the absolute weight total of each filter’s importance and then retraining the trimmed network to regain its accuracy.Deep compression approach for compressing deep learning models was introduced by S.Hanet al.[18].The compression is accomplished by a three-stage pipeline that includes network pruning,learned quantization of weights and weight sharing,followed by Huffman coding,resulting in a model with minimal accuracy loss.This is done to decrease the amount of data stored in neural networks as well as the amount of energy required to complete inference stages on large networks.To learn more compact CNNs,Liuet al.[19] presented a network slimming technique.With sparsity regularization,it adds a scaling factor for each channel.By applying an L1 regularization on the scaling factors during the batch normalization step,some scaling factors are suppressed to zero,which consequently zeroes out the corresponding weaker channels of a layer.Heet al.[20]proposed a new channel level pruning strategy for very deep CNN that can be used to speed them.They propose a two-step iterative approach based on LASSO regression and least square reconstruction to successfully trim each layer.The algorithm can be used in multilayer and multi-branch situations.Simplified CNNs are inference effective networks that achieve improved speed-ups while preserving comparable accuracy.

    P.Singhet al.[21] use adaptive filter pruning to accelerate deep convolutional neural networks.It decreases both the total number of parameters and the total computation time.Two modules execute alternatively-one which maximizes the number of pruned filters and the other module minimizes the drop in accuracy.A.G.Howardet al.[22] created a new class of efficient models known as MobileNets,which are utilized for mobile and smaller embedded vision applications that are resource constrained.To produce a lightweight deep neural network,they developed a streamlined architecture that leverages depth wise separable convolution operations.Based on the limitations of the problem,two global hyperparameters were used to pick the proper sized model for the application,and these parameters helped in a reasonable level of accuracy tradeoff to reduce the size and latency.Hubaraet al.[23] introduced Quantized Neural Network (QNN) as an enhanced technique for quantifying neurons and weights throughout inference and training.All multiply-accumulate operators (MACs) are substituted with XNOR operators in this network.QNNs reduce memory capacity and accesses during the forward pass and substitute most mathematical operation with bitwise operations,lowering power consumption and enhancing calculation speeds.A Binarized Neural Network is created when QNNs are employed with incredibly low precision,such as when just one bit is being used for weight and activation.Table 1 summarizes all the related work in the field of network compression.In the two-fold accelerated CNN system,the network compression technique i.e soft pruning is combined with reduction of large datasets that further speeds up the training process.

    Table 1.A comparison of related work in acceleration of CNN.

    III.PROPOSED TWO-FOLD ACCELERATED CNN SYSTEM

    Considering the scalability of enormous datasets and model parameters,during the iterative training process,the criticality of data instances is calculated using the effects values.There is a substantial number of repetitive or non-critical input data in real-world datasets,which has no effect on model parameter updating[24].As a result,the non-essential data points are recognized and eliminated,leaving only the major determining data for training.Integration with model compression methods is the next step.We use filter level pruning strategy to remove redundant model parameters in a CNN that have a minor impact on model accuracy and output.We combine this approach of removal of redundant data points with a soft filter pruning technique [25,26] to study the improvement in efficiency of training CNN models on large scale datasets.Figure 1 represents the block diagram and flow of the data between various modules.

    Figure 1.Block diagram for the two-fold accelerated CNN system.

    3.1 Critical Data Identification and Selection

    This is further divided into two steps: Critical input data identification via aggregation and Data selection via data removal.

    3.1.1 Aggregation of Similar Data

    During the pre-training phase,the crucial dataset identifier converts the dataset into a structure of aggregated input data samples.Every data point is a portion of the initial input data points with similar properties.It is only necessary to generate aggregated data once.Firstly,the input data points having similar attributes are clustered together in a group.Every group is then represented with a representative aggregated data point.In order to provide both quick processing time and accuracy of results,groups are created at two granularities i.e finer grained and coarse-grained,using a stack of accumulated data points.First,subgroups of connected data points are created from the raw data.Then,a matching number of fine-grained compression data points (xF,yF) are produced using the original input data points with a specific compression ratio.Then,based on the fine-grained compression data points,coarse-grained compression data points (xC,yC) are generated.There are two reasons for this type of aggregated design of data points.First,using aggregated data points,computation of high-dimensional and massive data can be completed fast.Second,it can ensure that updating model parameters among consolidated data points and its associated input data points has similar impacts.The coarse-grained data points are smaller in number than the fine-grained data points because they can be evaluated faster.

    Therefore two processes are used to combine input data points with similar attribute information to create an aggregated data point.The input data is divided across n subgroups(n >1),each of which contains several input data points with comparable values.Each input data point’s class output is compared to form a group.Every aggregated data point is linked to the set of corresponding original input data values.The following stage summarises the findings of the original input data points in each segment of the input data.

    3.1.2 Redundant Data Removal

    The last step of the pre-training phase before iterative model training,entails removing input data before training the model.Its primary concept is to acquire vital data by eliminating superfluous data based on aggregated data points.The impacts score of a coarsegrained/fine-grained aggregated point is calculated using cross-entropy loss.In multi-label classification problem,P-dimensional vectoryPdenote the output of the model trained andPbe the total number of class labels.Ifymax=max(yP) be the maximum value ofyP,the effect value ofithaggregated data point is evaluated using Eq.(1),

    In this case the real category of theithinput data point corresponds toyti.

    Various features have different implications on model parameter updates,as well.To train any deep learning model,the usual method requires estimating the influence of a feature over all training data in a batch.Aggregated points can be used to approximate such training samples more rapidly,speeding up estimate.

    3.2 Compression of Model Using Soft Pruning

    In the second phase,the critical data selected from the previous model is fed into the CNN model which is coupled with filter pruning strategy.This process is done iteratively for several epochs and different data points are selected over each epochs and different filters are pruned over different epochs.Generally a CNN network is parameterized byθ(i),1≤i ≤L.θ(i)represents the weight matrix ofithconvolutional layer andLdenotes the total number of layers in the model.The input tensor is denoted byXwith a shape ofni-1×wi-1×hi-1,wherewi-1andhi-1are the width and height of the input from previous layer andni-1is the number of channels from the previous layer.Similarly,the output tensor is denoted byYwith a shape ofni×wi×hi.Thus,the convolution operation performed over theithlayer is denoted by Eq.(2),

    Yi,kdenotes thekthfeature map generated fromithlayer andfi,kdenotes thekthfilter present inithlayer.θ(i) is parameterized as four-dimensional tensorni×x×x×ni-1,wherexis the height and width of the filters,niandni-1are the output and input channels respectively.

    Pruning of filters in a convolutional layers removes the corresponding output feature maps.If the pruning rate bePiforithlayer,then the number of filters of that layer gets reduced fromnitoni(1-Pi),such that shape of the output tensor changes toni(1-Pi)×wi×hi.

    Previous filter pruning efforts had a hard time compressing the deep CNNs.The term “hard method” in this context refers to the practice of removing filters from a single layer of a pre-trained model and afterwards adjusting the trimmed model to make up for the performance loss.After that,the following layers are pruned and then the model is fine-tuned once more until the final layer is trimmed.After filter pruning is completed,they are not modified again in these circumstances.As a result of the omitted filters,the model’s capacity is substantially reduced,resulting in a smaller optimization area for the model to understand from,negatively reducing the compressed models’performance.

    In the case of soft pruning,the approach dynamically eliminates filters while also updating pruned filters.In this methodology,we continuously update the pruned filters during the training phase throughout all the epochs as shown in Figure 2.There are numerous advantages of updating the pruned filters.In addition to maintaining the capacity of the compressed neural network model at par with the original models,it also prunes all the layers at once which makes it very time-efficient.Particularly,the soft pruning strategy can perform model pruning which is being trained from scratch as well as that has already been trained.In every epoch,the model is completely trained and optimized over the training data.Thel2-norm of all convolutional filters are evaluated in every convolutional layer after each epoch and is used as a criterion in the filter selection technique.The associated filter weights will then be set to zero to prune the chosen filters,followed by another training session.After all the steps of filter selection and elimination ideologies,the deep CNNs are pruned to provide an efficient and compact model.The filter pruning strategy is divided into four stages:

    Figure 2.A soft pruning model compression technique.

    Filter selection: In order to determine the relevance of each filter,lpnorm is utilized as shown in Eq.(3).Basically,a norm is used to find distance between two points.In this case,the distance may be assumed as the activation value from each filter.If the norm value is lesser,the activation values after the convolution operations is lesser,thus resulting in a lesser impact over the deep CNN model’s final prediction.If the norm value is higher,then they produce a higher activation value,which contributes highly over the model prediction.Thus,filter with smallerlpnorm will be pruned first,followed by those with greaterlpnorm.In this proposed method,l2norm is used as a filter selection criteria which provides better results over selectingl1norm.

    Fi,krepresents thekthfilter present inithlayer andnirepresents the number of filters inithlayer andwiandhirepresents the dimensions of the filter inithlayer.

    Filter Pruning:To perform filter pruning,all the values in the chosenniPiare made zero.As a result,the filter’s contribution to model output is momentarily eliminated.But the pruned filters are updated during back propagation,thereby maintaining the model capacity to learn from and high model performance.Pruning of all the filters of all the convolutional layers are done at once.Thus,pruning is done in parallel manner,thereby reducing the computational cost of pruning all the filters making in negligible.Also,a constant pruning rate is used for all the layers,to maintain the tradeoff between accuracy and acceleration and also reduces the complexity of maintaining too many hyper-parameters.

    Reconstruction: After the pruning step,forward propagation is performed on the the model,with some of the filters being zero.But during the backpropagation stage,all the pruned filters are updated back to non-zero,thereby maintaining a larger model capacity and better optimization space.

    Obtaining Compressed Model: The filter selection,filter pruning and reconstructions is performed iteratively during each training epoch.After the training process,all the weights of a pruned filter are reset to zero,which can be used as it is for inferencing or all such filters can be removed to create a compact,smaller model.Removing a filter from one layer also impacts the input channels of the next layer.Particularly,with a pruning rate ofPiin theithlayer,only the values ofni(1-Pi)filters are not null and contributes over the final prediction.In case of the previous layer,the number of input channels of theithlayer is modified fromnitoni(1-Pi-1).As a result,theithlayer can be reconstructed into a smaller structure.Figure 3 represents the architecture of the proposed accelerated methodology in detail.

    Figure 3.Detailed module-wise architecture of the various phases of the two-fold accelerated CNN system.

    3.3 Computational Complexity

    The aggregation of data points is performed by creating groups of similar elements.Firstly,considering that there areLways to split the dataset,whereLdenotes the features of a dataset andNdenotes the total number of data points,we split the datasets equally into 2Lpartitions.The entire dataset is sorted at every point on the chosen attributeaiwhere i=1...L.For example if there exists an attributea1denoting the very first split attribute,the entire dataset will be sorted on this attribute and split into two equal halves.Then,the two subsets will be further recursively partitioned in the same way for every attributea2toaL.So the time complexity for grouping and creating aggregated data points is O(L x N).This method proves to work much faster than clustering techniques which will take time O(N x k x i) where k is the number of clusters and i is the number of iterations.It is worth noting that the grouping and aggregation is performed only once as a pre-processing step.Even though the removal of non-critical data points takes place iteratively as a part of training,the cost is negligible as compared to the aggregation.

    The most time consuming step during training is the pruning of filters.Assuming that to prune the filters of a layeri,the norm of every filter needs to be examined.As seen from Eq.(3),computing the norm of every filter in layericonsumes O(nixni-1xwixhi).

    IV.IMPLEMENTATION

    4.1 Dataset Description

    A public dataset for common image classification,CIFAR-10,is used to assess the proposed method.60000 32x32 color images in 10 classes,with 6000 images per class,make up the CIFAR-10 dataset.50,000 training images and 10,000 test images are available.Each training batch and test batch in the dataset has 10,000 images,and there are five training batches in total.Exactly 1000 images from each class were chosen at random for the test batch.A total of 5000 images from each class are included in the training batches.

    4.2 Experimental Setup

    The CNN model architecture which is used to implement our acceleration strategy is ResNet.In the case of the CIFAR-10 dataset,we test the proposed methodology on ResNet-20,ResNet-32,ResNet-56 and ResNet-110.The models are trained from scratch without using the pre-trained weights.The pruning rate is the same for all the layers and is set as 30%.Without any extra fine-tuning,a regular training cycle is utilized to prune the model from scratch.l2norm is the pruning criterion utilized for filter selection.The batch size is set at 128 and the categorical cross-entropy loss function is employed in the training of the ResNet-110 model across 200 epochs.A Stochastic Gradient Descent optimization is applied,with 0.0005 decay and 0.9 momentum.The learning rate is set to 0.01 at the beginning,and it is gradually decreased over a specified number of training epochs.In case of critical data identification and selection,the number of subsets need to be finalized.There are 50,000 training images and 5000 images fall under a single class.The total number of images in each subset is set to 100 and the total number of similar image subsets generated is 500.The compression rate used to create the fine-grained data point is 30%,a from the fine-grained points only a single coarse-grained point is generated.Each subset has one coarse-grained aggregated data point.We choose these optimization settings based on the existing literature on training original baseline models and the existing state-of-the-art filter pruning methods.A pruning ratio of 30% has been considered as the most optimal since it gives a good balance between reduction of FLOPs as well as negligible accuracy drop.The threshold values for critical and non-critical data,i.e.,Upper and lower bounds are equal to,respectively,the average of the effect values and the value divided by 5.A summary of all the parameter settings has been shown in Table 2

    Table 2.A summary of optimization settings.

    V.RESULT AND ANALYSIS

    The results for the model training utilising the proposed methodology have been evaluated and shown in this part using line graphs displaying the variation of accuracy vs.number of epochs for training and validation sets and variations of cross -entropy loss vs.number of epochs is depicted in Figure 4.This indicates that the model’s accuracy and loss converges over 200 epochs and does not improve over the next epochs.

    Figure 4.Epochs vs Loss graphs for train and validation sets over model training.

    The accuracy of training and validation are much higher,which shows that the model has been trained properly for the image classification task and their weights could predict accurately as depicted in Figure 5.

    Figure 5.Epochs vs Accuracy graphs for train and validation sets over model training.

    Table 3 compares the trade-off between the drop in accuracy and the reduced number of FLOPs.This is done for ResNet-110 model over three different pruning rates which are 10%,20%and 30%and is trained from scratch.From the results observed,30%pruning rate produces a better accuracy of 93.57% and lesser accuracy drop of 0.11 is produced but the reduction in number of FLOPs is higher which is around 40.7%when compared with the results of using 10%and 20%pruning rates.Pruning rates are set as same to all the layers,and pruning is not done for the projection shortcuts in ResNet for simplification.

    Table 3.Comparison of performance metrics of CIFAR-10 over the ResNet110 architecture.

    Table 4 shows the total number of FLOPs performed by different ResNet model architectures with pruning and without pruning.The total number of FLOPs is shown for different ResNet models such as ResNet-110,ResNet-56,ResNet-32 and ResNet-20.Also,varying pruning rates are used to calculate how pruning rates make an impact on the FLOPs reduction rate.

    Table 4.Comparison of FLOPs and FLOPs reduction rate over different pruning rates on different ResNet architectures.

    Table 5 estimates the performance of the model,by choosing different filter selection strategies which is used as pruning criterion.Different pruning rates of 10%,20%and 30%are used overl1-norm andl2-norm as filter selection methods.From the results obtained,l2-norm proves to produce only a slightly better accuracy when compared withl1-norm.

    Table 5.Comparison of overall accuracy of the model over different pruning rates and pruning criterion.

    Table 6 compares the overall performance which includes accuracy drop as compared to the complete baseline model,and FLOPs reduction.To ensure that the proposed model’s performance for classification tasks is comparable to or better than that of state-ofthe-art,it must be compared to existing models such as PFEC[17],Network slimming[19],GAL[27],Partial Least-squares[28],NISP[29]and MIL[1].For a fair comparison,we only consider the accuracy drop for comparison,as the optimization settings for different methods may vary.Comparison states that the proposed methods reduces the FLOPs by 40.7%with an accuracy drop of 0.11% which proves to be better in terms of accuracy and FLOPs reduction rate making it effective over the existing approaches where PFEC[17] reduces the FLOPs by 38.6% with an accuracy drop of 0.61%and MIL[1]reduces the FLOPs by 34.2%with an accuracy drop of 0.19%.The FLOP reduction of network slimming [19] is 27.6% with a much larger accuracy drop of 1.73%.Even though[27]and[28]have slightly better FLOPs reduction ratio,our method outperforms them when it comes to the accuracy drop.

    Table 6.Comparison of model performance of proposed and existing methods.

    The training time taken for training the model for200 epochs without pruning is 12935.50 seconds(3.59 hrs)and the time taken for training the model for 200 epochs with a pruning rate of 30%is 12285.84 seconds(3.41 hrs).This observation shows that the pruned model gets trained faster than the unpruned model,and also with a negligible accuracy drop.In case of huge datasets and bigger network architectures,this difference can be increased a lot.In the model training,pruning of filters is done at the end of every epoch.If we change the interval gap in which these filters are pruned,for instance,after every 2 epochs the filters are pruned,after 5 epochs filters are pruned,etc.This can be taken as a hyper parameter,but this varies from one dataset to other dataset and according to the model architectures.When compared with other state-of the-art methods such as[17],[19]and[1],the proposed twofold accelerated CNN system,performs much better with a very minimal loss as compared to the baseline model.Moreover,the FLOPs reduction ratio is the highest comparatively which not only makes it energy efficient but also ensures that it can be deployed on resourced constrained devices for inference.After compressing the original 50,000 data points to 500 coarsegrained aggregated points,there was a difference of approximately 6 seconds observed in the training time duration.However,not much improvement was observed in the accuracy of training with coarse-grained points.

    VI.CONCLUSION

    In this work,we implement a two-fold approach to speed up the training of deep learning models through redundant data removal and filter pruning.In the first stage,we try to identify and select only the critical data for model training,thereby reducing the overall computational costs on unimportant data points.From the critical data selected,the model is trained upon this data which is incorporated with a filterlevel pruning strategy which allows the pruned filters to be updated back thereby enforcing sparsity during the model training and attaining a better optimization space and model capacity to learn from,thereby achieving a better performance than existing methods.Without using the pre-trained weights,this strategy allows the pre-trained models to be trained from the scratch and attain a superior performance compared to the state-of-the-art approaches.Through results obtained it is evident that,our proposed method gives a minimum of 1.08×speed-up as compared to baseline models.Our proposed work has only been tested on CIFAR-10 dataset and its impact on other benchmark as well as real life datasets is yet to be studied.In future,this proposed method can be implemented in a distributed deep learning setup to overcome the communication overhead that is a severe challenge in that domain.When models are trained on multiple worker nodes simultaneosly,the parameter exchange causes a severe network bottleneck which can be significantly reduced through our proposed acceleration method.

    大话2 男鬼变身卡| 亚洲 欧美一区二区三区| 下体分泌物呈黄色| 精品一区二区三卡| 欧美精品一区二区大全| 99热网站在线观看| 咕卡用的链子| 丁香六月天网| 国产三级黄色录像| 日本一区二区免费在线视频| 中文字幕高清在线视频| 男人舔女人的私密视频| 成人亚洲精品一区在线观看| 色婷婷av一区二区三区视频| 精品久久久精品久久久| 中文字幕制服av| 亚洲精品一二三| 一级片免费观看大全| 久久久久视频综合| 亚洲国产精品成人久久小说| 两个人免费观看高清视频| 大话2 男鬼变身卡| 亚洲欧洲日产国产| 亚洲精品一卡2卡三卡4卡5卡 | 精品视频人人做人人爽| 欧美精品人与动牲交sv欧美| 国产视频首页在线观看| 性色av一级| 国产欧美日韩一区二区三 | 99香蕉大伊视频| 久久人人爽av亚洲精品天堂| 老汉色∧v一级毛片| 久久av网站| 视频在线观看一区二区三区| 久久久久久久久免费视频了| 亚洲国产欧美日韩在线播放| 国产精品一区二区在线观看99| 久久精品久久久久久噜噜老黄| 亚洲精品久久午夜乱码| 每晚都被弄得嗷嗷叫到高潮| 看十八女毛片水多多多| 18禁黄网站禁片午夜丰满| av在线播放精品| 999精品在线视频| 国产精品二区激情视频| 少妇人妻 视频| 久久国产亚洲av麻豆专区| 免费在线观看影片大全网站 | www.熟女人妻精品国产| 丰满饥渴人妻一区二区三| kizo精华| 亚洲色图 男人天堂 中文字幕| 国产日韩一区二区三区精品不卡| 亚洲中文字幕日韩| 久久鲁丝午夜福利片| 午夜av观看不卡| 天堂8中文在线网| 午夜两性在线视频| 国产欧美日韩精品亚洲av| 欧美日韩视频精品一区| 大片免费播放器 马上看| 成年av动漫网址| 19禁男女啪啪无遮挡网站| 久9热在线精品视频| 一级毛片女人18水好多 | 亚洲精品乱久久久久久| 亚洲欧美精品综合一区二区三区| www.999成人在线观看| 久久精品亚洲av国产电影网| 国产免费视频播放在线视频| 国产亚洲av片在线观看秒播厂| 国产av精品麻豆| 老司机午夜十八禁免费视频| www日本在线高清视频| 亚洲欧美色中文字幕在线| 天堂中文最新版在线下载| 精品久久蜜臀av无| 久久久久网色| 国产午夜精品一二区理论片| 久9热在线精品视频| 国产1区2区3区精品| 伊人久久大香线蕉亚洲五| 日韩熟女老妇一区二区性免费视频| 久久精品久久久久久噜噜老黄| 国产成人精品久久二区二区免费| 精品一区二区三卡| 国产黄色免费在线视频| 黄色怎么调成土黄色| 欧美 日韩 精品 国产| 大香蕉久久网| 99久久精品国产亚洲精品| 一区二区三区乱码不卡18| svipshipincom国产片| 97人妻天天添夜夜摸| 80岁老熟妇乱子伦牲交| 中文字幕人妻熟女乱码| 免费在线观看视频国产中文字幕亚洲 | 久久久久久久国产电影| 免费观看av网站的网址| 欧美另类一区| 狠狠精品人妻久久久久久综合| 亚洲人成77777在线视频| 国产97色在线日韩免费| a级毛片黄视频| 国产精品国产av在线观看| 国产精品偷伦视频观看了| 制服人妻中文乱码| 国产精品一二三区在线看| 亚洲成人国产一区在线观看 | 好男人视频免费观看在线| 国产高清国产精品国产三级| 久久青草综合色| 亚洲精品国产av成人精品| 一级黄片播放器| av在线老鸭窝| 国产一区二区激情短视频 | 亚洲熟女精品中文字幕| xxx大片免费视频| 精品少妇内射三级| 久久国产精品影院| 国产成人系列免费观看| 久久久久精品人妻al黑| 亚洲国产精品一区二区三区在线| 国产麻豆69| 黄片播放在线免费| 丝袜脚勾引网站| 国产一卡二卡三卡精品| 精品国产一区二区久久| 日本欧美国产在线视频| www.999成人在线观看| 国产欧美日韩综合在线一区二区| 老司机亚洲免费影院| 国产主播在线观看一区二区 | 亚洲精品国产av蜜桃| 色网站视频免费| 一本色道久久久久久精品综合| 欧美国产精品va在线观看不卡| 国产xxxxx性猛交| 一级,二级,三级黄色视频| 久久综合国产亚洲精品| av在线app专区| 99国产精品一区二区蜜桃av | 欧美黑人欧美精品刺激| 在线av久久热| 久久精品久久久久久久性| 大香蕉久久网| 国产成人啪精品午夜网站| 午夜两性在线视频| 精品国产乱码久久久久久男人| 色综合欧美亚洲国产小说| 国产精品一区二区精品视频观看| 在线精品无人区一区二区三| 亚洲av男天堂| 亚洲欧美激情在线| 亚洲国产日韩一区二区| 久久久国产一区二区| 亚洲中文字幕日韩| 精品国产乱码久久久久久男人| 高清av免费在线| 天堂俺去俺来也www色官网| 国产欧美日韩精品亚洲av| 日本vs欧美在线观看视频| a 毛片基地| 精品少妇黑人巨大在线播放| 婷婷丁香在线五月| 国产精品久久久久久精品古装| 天天添夜夜摸| 在线观看国产h片| 纯流量卡能插随身wifi吗| 欧美日韩亚洲综合一区二区三区_| 2018国产大陆天天弄谢| 精品国产超薄肉色丝袜足j| 亚洲成色77777| 高清av免费在线| 老汉色∧v一级毛片| 欧美日韩亚洲综合一区二区三区_| 日韩av免费高清视频| 久久久久国产精品人妻一区二区| 国产免费现黄频在线看| 日本猛色少妇xxxxx猛交久久| 久久精品久久久久久久性| 国产男女超爽视频在线观看| 男女高潮啪啪啪动态图| 91字幕亚洲| 亚洲av日韩在线播放| 在线观看国产h片| 人人妻,人人澡人人爽秒播 | 亚洲一区中文字幕在线| 成人亚洲精品一区在线观看| 成人18禁高潮啪啪吃奶动态图| 好男人电影高清在线观看| 亚洲国产精品一区二区三区在线| 真人做人爱边吃奶动态| 久久亚洲国产成人精品v| 亚洲国产av新网站| 亚洲欧美激情在线| 国产有黄有色有爽视频| 我的亚洲天堂| 国产精品香港三级国产av潘金莲 | 午夜福利视频在线观看免费| 中文欧美无线码| 精品卡一卡二卡四卡免费| 欧美黑人精品巨大| 国产成人免费观看mmmm| 欧美日韩黄片免| 免费女性裸体啪啪无遮挡网站| 中国国产av一级| 精品人妻一区二区三区麻豆| 69精品国产乱码久久久| 久久久久国产精品人妻一区二区| 国产精品一区二区在线观看99| 中文精品一卡2卡3卡4更新| 麻豆乱淫一区二区| 免费高清在线观看日韩| 国产人伦9x9x在线观看| 亚洲国产毛片av蜜桃av| 精品熟女少妇八av免费久了| 亚洲欧美激情在线| 亚洲av欧美aⅴ国产| 丁香六月欧美| 免费在线观看完整版高清| 性少妇av在线| 悠悠久久av| 人成视频在线观看免费观看| 久久精品国产综合久久久| 亚洲成人国产一区在线观看 | 亚洲精品一二三| 99热网站在线观看| 午夜精品国产一区二区电影| 日韩一本色道免费dvd| 好男人视频免费观看在线| 蜜桃在线观看..| 日韩大片免费观看网站| 日韩av免费高清视频| 999精品在线视频| 国产福利在线免费观看视频| 丰满少妇做爰视频| 国产在线免费精品| 黄片播放在线免费| 亚洲人成网站在线观看播放| 成人三级做爰电影| 午夜免费成人在线视频| 久久九九热精品免费| 国产老妇伦熟女老妇高清| 久久鲁丝午夜福利片| 国产亚洲午夜精品一区二区久久| av在线老鸭窝| 精品一区二区三区四区五区乱码 | 国产福利在线免费观看视频| 好男人视频免费观看在线| 免费在线观看完整版高清| 国产精品人妻久久久影院| 丝袜在线中文字幕| 欧美日韩成人在线一区二区| 女性被躁到高潮视频| 九草在线视频观看| 人人妻人人添人人爽欧美一区卜| 免费看不卡的av| 久久久久久亚洲精品国产蜜桃av| 伊人久久大香线蕉亚洲五| 欧美亚洲 丝袜 人妻 在线| 无限看片的www在线观看| 捣出白浆h1v1| 国产精品一区二区在线不卡| 国产精品香港三级国产av潘金莲 | 国产精品香港三级国产av潘金莲 | 考比视频在线观看| 婷婷丁香在线五月| 波多野结衣av一区二区av| 别揉我奶头~嗯~啊~动态视频 | 国产午夜精品一二区理论片| 日韩av免费高清视频| 十八禁高潮呻吟视频| 成人黄色视频免费在线看| 国产黄色视频一区二区在线观看| 国产福利在线免费观看视频| 不卡av一区二区三区| 国产免费现黄频在线看| 啦啦啦视频在线资源免费观看| 午夜免费男女啪啪视频观看| 国产男人的电影天堂91| 在线观看一区二区三区激情| 美女主播在线视频| 亚洲自偷自拍图片 自拍| 国产欧美日韩一区二区三 | 亚洲熟女精品中文字幕| 好男人视频免费观看在线| 成人午夜精彩视频在线观看| 亚洲国产av新网站| 精品少妇一区二区三区视频日本电影| 国产精品一区二区在线观看99| 国产精品三级大全| 亚洲 国产 在线| 亚洲,欧美,日韩| 亚洲国产av影院在线观看| 亚洲一区中文字幕在线| 久久鲁丝午夜福利片| 亚洲精品乱久久久久久| 久久久久网色| 久久久欧美国产精品| 香蕉国产在线看| 91精品伊人久久大香线蕉| 少妇精品久久久久久久| a级毛片在线看网站| 青青草视频在线视频观看| 日韩制服丝袜自拍偷拍| 日本a在线网址| 亚洲精品国产色婷婷电影| 丝袜在线中文字幕| 欧美日韩亚洲综合一区二区三区_| 多毛熟女@视频| 中文字幕制服av| 日本a在线网址| 人人妻人人添人人爽欧美一区卜| 丝袜喷水一区| 91国产中文字幕| 一本大道久久a久久精品| 亚洲国产精品999| 国产精品三级大全| 午夜福利乱码中文字幕| 亚洲欧洲日产国产| 久久精品aⅴ一区二区三区四区| 亚洲精品乱久久久久久| 国产精品久久久av美女十八| 亚洲伊人久久精品综合| 极品少妇高潮喷水抽搐| 十分钟在线观看高清视频www| 丝瓜视频免费看黄片| 婷婷色综合大香蕉| 国产欧美日韩一区二区三区在线| 日本av免费视频播放| 亚洲三区欧美一区| 两性夫妻黄色片| 一级黄色大片毛片| 超色免费av| 久久久久国产精品人妻一区二区| 亚洲精品日韩在线中文字幕| 精品少妇一区二区三区视频日本电影| 十八禁人妻一区二区| 少妇 在线观看| 欧美精品人与动牲交sv欧美| 国产精品99久久99久久久不卡| 亚洲一码二码三码区别大吗| 亚洲欧美中文字幕日韩二区| a级毛片在线看网站| 下体分泌物呈黄色| 男女边摸边吃奶| 水蜜桃什么品种好| 18在线观看网站| 亚洲精品一卡2卡三卡4卡5卡 | 亚洲天堂av无毛| 秋霞在线观看毛片| 久久九九热精品免费| 亚洲成人免费电影在线观看 | 乱人伦中国视频| 人人妻人人澡人人看| 中文字幕制服av| 欧美日韩一级在线毛片| 天天躁夜夜躁狠狠躁躁| 亚洲欧美成人综合另类久久久| 91字幕亚洲| 亚洲欧洲国产日韩| 新久久久久国产一级毛片| 精品久久久久久电影网| 91成人精品电影| 十分钟在线观看高清视频www| 丁香六月天网| 欧美黄色淫秽网站| 国产免费又黄又爽又色| √禁漫天堂资源中文www| 一区福利在线观看| 亚洲精品国产色婷婷电影| 午夜免费男女啪啪视频观看| 人人澡人人妻人| 亚洲成av片中文字幕在线观看| 我的亚洲天堂| 国产精品.久久久| 精品少妇内射三级| 国产有黄有色有爽视频| 成年人午夜在线观看视频| 国产成人精品无人区| 女性被躁到高潮视频| 国产成人精品久久二区二区91| 女性被躁到高潮视频| 天天操日日干夜夜撸| 一边摸一边抽搐一进一出视频| 亚洲黑人精品在线| 国产视频一区二区在线看| 18在线观看网站| 久久久精品区二区三区| 欧美日韩福利视频一区二区| av线在线观看网站| 爱豆传媒免费全集在线观看| 久久精品熟女亚洲av麻豆精品| 午夜久久久在线观看| 午夜免费观看性视频| 久久久精品94久久精品| 久久亚洲国产成人精品v| 国产成人av激情在线播放| 捣出白浆h1v1| netflix在线观看网站| 亚洲九九香蕉| 亚洲欧美中文字幕日韩二区| 嫁个100分男人电影在线观看 | 国产成人91sexporn| 丝袜喷水一区| 国产精品欧美亚洲77777| 久久毛片免费看一区二区三区| 中国国产av一级| 99国产精品免费福利视频| a 毛片基地| 一级片免费观看大全| 久久人人爽人人片av| 亚洲美女黄色视频免费看| 悠悠久久av| 国产成人一区二区三区免费视频网站 | 19禁男女啪啪无遮挡网站| 午夜福利视频精品| 免费女性裸体啪啪无遮挡网站| 五月天丁香电影| 精品福利永久在线观看| 国产视频一区二区在线看| 亚洲美女黄色视频免费看| 亚洲国产成人一精品久久久| 最新在线观看一区二区三区 | 国产成人免费观看mmmm| 久久精品熟女亚洲av麻豆精品| 国产高清videossex| 午夜福利,免费看| 啦啦啦在线免费观看视频4| videosex国产| 欧美人与性动交α欧美软件| 日韩av不卡免费在线播放| 欧美成人午夜精品| 69精品国产乱码久久久| 新久久久久国产一级毛片| 精品欧美一区二区三区在线| 91麻豆精品激情在线观看国产 | 午夜精品国产一区二区电影| 日韩大片免费观看网站| av视频免费观看在线观看| 丰满迷人的少妇在线观看| www.av在线官网国产| 亚洲欧美一区二区三区久久| 人成视频在线观看免费观看| 激情视频va一区二区三区| 精品国产乱码久久久久久小说| 国产在线一区二区三区精| 91麻豆av在线| 久久亚洲国产成人精品v| 国精品久久久久久国模美| 99精国产麻豆久久婷婷| 国产精品一区二区精品视频观看| 免费观看a级毛片全部| 国产视频一区二区在线看| 久久九九热精品免费| 日韩制服丝袜自拍偷拍| 大片电影免费在线观看免费| 欧美黑人欧美精品刺激| 久久精品国产综合久久久| 国产福利在线免费观看视频| 亚洲精品在线美女| 亚洲伊人久久精品综合| 国产淫语在线视频| 天堂8中文在线网| 人人妻人人澡人人爽人人夜夜| 亚洲自偷自拍图片 自拍| 亚洲精品自拍成人| 亚洲精品国产av蜜桃| 各种免费的搞黄视频| 另类亚洲欧美激情| 免费一级毛片在线播放高清视频 | 久久精品人人爽人人爽视色| 老司机深夜福利视频在线观看 | 精品第一国产精品| 亚洲av电影在线进入| 欧美精品啪啪一区二区三区 | av片东京热男人的天堂| 亚洲精品久久成人aⅴ小说| 黄色a级毛片大全视频| 午夜福利在线免费观看网站| a 毛片基地| 一区在线观看完整版| 国产精品二区激情视频| 五月开心婷婷网| 久久久久精品国产欧美久久久 | 777久久人妻少妇嫩草av网站| 人体艺术视频欧美日本| 91精品国产国语对白视频| 捣出白浆h1v1| 少妇猛男粗大的猛烈进出视频| 亚洲av电影在线进入| 最近手机中文字幕大全| 永久免费av网站大全| 久久人人爽人人片av| 亚洲人成电影观看| 一区二区三区激情视频| 香蕉国产在线看| 国产无遮挡羞羞视频在线观看| 国产黄色免费在线视频| 少妇 在线观看| 久久免费观看电影| 午夜激情av网站| 人成视频在线观看免费观看| 9热在线视频观看99| 久久精品国产亚洲av高清一级| 亚洲欧美一区二区三区久久| 亚洲一区二区三区欧美精品| 亚洲av成人不卡在线观看播放网 | 精品福利观看| 婷婷色综合大香蕉| 欧美97在线视频| 人人妻人人添人人爽欧美一区卜| 免费女性裸体啪啪无遮挡网站| 亚洲中文字幕日韩| 久久久久久亚洲精品国产蜜桃av| 亚洲第一青青草原| 日韩免费高清中文字幕av| 又大又爽又粗| 十八禁高潮呻吟视频| av一本久久久久| 视频区欧美日本亚洲| 欧美日韩av久久| 操出白浆在线播放| 国产淫语在线视频| 国产成人免费无遮挡视频| 纵有疾风起免费观看全集完整版| 午夜视频精品福利| 国产精品免费大片| 19禁男女啪啪无遮挡网站| 亚洲一区中文字幕在线| 精品熟女少妇八av免费久了| avwww免费| 亚洲精品av麻豆狂野| 永久免费av网站大全| 久久久久久久大尺度免费视频| 一级片'在线观看视频| 欧美日本中文国产一区发布| 美女国产高潮福利片在线看| 女警被强在线播放| av视频免费观看在线观看| 丝袜美足系列| 久久天躁狠狠躁夜夜2o2o | 欧美国产精品一级二级三级| 国产精品一区二区在线观看99| 又紧又爽又黄一区二区| 两人在一起打扑克的视频| 18禁国产床啪视频网站| 一本综合久久免费| 91国产中文字幕| 国产精品久久久av美女十八| 脱女人内裤的视频| 国产成人精品无人区| 三上悠亚av全集在线观看| 2018国产大陆天天弄谢| 性高湖久久久久久久久免费观看| 亚洲欧美一区二区三区国产| 性少妇av在线| 国产成人一区二区在线| 在线 av 中文字幕| bbb黄色大片| 国产精品国产av在线观看| 亚洲国产最新在线播放| 香蕉国产在线看| 国产xxxxx性猛交| 欧美精品av麻豆av| 成年人免费黄色播放视频| 一区二区三区乱码不卡18| 两性夫妻黄色片| 欧美精品高潮呻吟av久久| 免费av中文字幕在线| 亚洲成av片中文字幕在线观看| 午夜91福利影院| 大香蕉久久成人网| 天天影视国产精品| 国产1区2区3区精品| 亚洲精品日韩在线中文字幕| 免费黄频网站在线观看国产| 啦啦啦中文免费视频观看日本| 久久久久久久久久久久大奶| 亚洲国产精品国产精品| 久久影院123| 精品国产一区二区三区久久久樱花| 久久久久久免费高清国产稀缺| 一级片'在线观看视频| 久久精品人人爽人人爽视色| 亚洲中文字幕日韩| 自拍欧美九色日韩亚洲蝌蚪91| 性高湖久久久久久久久免费观看| 麻豆乱淫一区二区| 亚洲国产欧美在线一区| 性高湖久久久久久久久免费观看| 亚洲欧美激情在线| 色综合欧美亚洲国产小说| videosex国产| 丰满饥渴人妻一区二区三| 一级片'在线观看视频| 国产精品一区二区在线观看99| 精品国产超薄肉色丝袜足j| 久久久国产一区二区| 免费高清在线观看日韩| 国产无遮挡羞羞视频在线观看| 国产激情久久老熟女| 精品国产一区二区三区久久久樱花| 999精品在线视频| svipshipincom国产片| 中文字幕另类日韩欧美亚洲嫩草| 晚上一个人看的免费电影| 国产高清videossex| 亚洲七黄色美女视频| 999久久久国产精品视频| 亚洲av电影在线观看一区二区三区| 成年美女黄网站色视频大全免费| 久久人人97超碰香蕉20202| 久久人人爽av亚洲精品天堂|