• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Reinforcement Learning Based Quantization Strategy Optimal Assignment Algorithm for Mixed Precision

    2024-05-25 14:40:48YuejiaoWangZhongMaChaojieYangYuYangandLuWei
    Computers Materials&Continua 2024年4期

    Yuejiao Wang,Zhong Ma,Chaojie Yang,Yu Yang and Lu Wei

    Xi’an Microelectronics Technology Institute,Xi’an,710065,China

    ABSTRACT The quantization algorithm compresses the original network by reducing the numerical bit width of the model,which improves the computation speed.Because different layers have different redundancy and sensitivity to data bit width.Reducing the data bit width will result in a loss of accuracy.Therefore,it is difficult to determine the optimal bit width for different parts of the network with guaranteed accuracy.Mixed precision quantization can effectively reduce the amount of computation while keeping the model accuracy basically unchanged.In this paper,a hardware-aware mixed precision quantization strategy optimal assignment algorithm adapted to low bit width is proposed,and reinforcement learning is used to automatically predict the mixed precision that meets the constraints of hardware resources.In the state-space design,the standard deviation of weights is used to measure the distribution difference of data,the execution speed feedback of simulated neural network accelerator inference is used as the environment to limit the action space of the agent,and the accuracy of the quantization model after retraining is used as the reward function to guide the agent to carry out deep reinforcement learning training.The experimental results show that the proposed method obtains a suitable model layer-by-layer quantization strategy under the condition that the computational resources are satisfied,and the model accuracy is effectively improved.The proposed method has strong intelligence and certain universality and has strong application potential in the field of mixed precision quantization and embedded neural network model deployment.

    KEYWORDS Mixed precision quantization;quantization strategy optimal assignment;reinforcement learning;neural network model deployment

    1 Introduction

    The quantization algorithm compresses the original network by reducing the numerical bit width[1]and improving the computation speed on neural network accelerators[2].On one hand,different layers have different redundancy and accuracy requirements to bit width.Therefore,assigning the same bit width to all layers will result in a loss of accuracy.On the other hand,when different layers use flexible bit widths,the mixed-precision search space is exponential in the number of layers.It is not realistic to search for the appropriate bit width for each layer and computation resource is insufficient[3].

    Therefore,it is necessary to use mixed precision quantization for different layers to reduce the computational resource while keeping the model accuracy basically unchanged,and effectively balance the contradiction between accuracy and computational performance.

    However,there is a lack of an effective approach to determine the bit widths of different parts of the network.Traditional methods either examine only a small,artificially designed search space or utilize tedious neural structure searches to explore huge search spaces.These methods are not effective in obtaining optimal quantization schemes.So,it is necessary to automatically obtain the optimal bit width assignment scheme through the automatic neural network model mixed precision computational assignment technology.

    So,there are two technical bottlenecks that need to be solved:

    Technical Bottleneck 1:How to choose the appropriate bit width?The bit width should be set as low as possible to reduce the computational cost.

    Technical Bottleneck 2:How to improve the quantization accuracy?The loss of accuracy due to quantization should be as small as possible.

    Aiming at these two technical bottlenecks,this paper proposes a quantization strategy optimal assignment algorithm for mixed precision,which uses reinforcement learning to automatically predict the quantization bit width assignment strategy of each layer of a given model.The simulation software that simulates neural network accelerator inference is used as an environment to obtain the execution speed feedback of the model to guide the agent to meet the resource constraints.Finally,after quantization retraining,different quantization strategy optimal assignment methods are realized,and the quantization strategies suitable for all layers are output.The quantization strategy optimization technology based on reinforcement learning can minimize computational resource consumption and data access bandwidth requirements while maintaining computational accuracy.

    2 Related Work

    Quantization represents network weights with fewer bits,such as converting the data type computed by an algorithm from 32-bit floating-point to 8-bit or 4-bit low-bit integers[4].Quantization algorithms quantize both parameters(i.e.,weights)and activations of each layer in a neural network model to reduce the total memory footprint of the model during inference [5–7].Mixed precision quantization designs a certain policy to reasonably allocate the weights and activations of each layer of the model.This allows mixed precision quantization to achieve the best compromise between the performance and accuracy of deep neural networks,and has therefore been widely studied.

    The parameters of a neural network model are typically stored in a 32-bit floating-point data format[8].By converting floating-point parameters into 1,2,4,or 8-bit integer data,model operation efficiency can be improved and model deployment can be promoted on embedded devices[9].However,the reduction of the data bit width leads to a loss of accuracy,and the lower the bit width,the greater the accuracy loss.In addition,the parameters of different layers of the model have different sensitivity to the data bit width,among which the parameters that contribute to the accuracy need a higher quantization bit width,and the parameters that contribute less to the high accuracy can give a lower bit width[10].Given the overall accuracy requirements of the model,the bit width assignment technology of the neural network model searches for and assigns appropriate bit widths to different layers to achieve the optimal quantization effect.However,there are many convolutional weights and activations in the model,and the number of bit widths can be selected,so the bit width search space refers to orders of magnitude,and manual search is very difficult[11].

    Reinforcement Learning is a technique that directly learns control strategies from highdimensional raw data [12].The machine trained by reinforcement learning has reached or exceeded the level of human intelligence in many fields such as Alpha Go,video games,robotics,and ChatGPT.Researchers in the field propose some hybrid bit width automatic search methods to solve this problem,among which optimization-based methods/reinforcement learning-based methods and gradient descent-based methods have received active attention from researchers.

    For example,a Lagrange multiplier[13]has been proposed for mixed precision search,which treats mixed precision quantization as an extreme value problem of function f(x1,x2,...)under the constraint of g(x1,x2,...)=0.Each channel in a single layer is solved according to the Lagrange multiplier method.However,this method can only be used for each layer,and secondly,it cannot be allocated according to the specified optional quantization bit width,but only according to the calculated quantization bit width.Wang et al.of Massachusetts Institute of Technology (MIT) proposed a Hardware-Aware Automated Quantization(HAQ)method based on reinforcement learning[14].For a certain layer of convolution,the delay and energy feedback provided by the hardware are included in the reward function based on reinforcement learning through the hardware accelerator.When looping,the quantization bit width is updated once in each reinforcement learning action according to the reward to make the bit width accuracy better than the bit width generated by the previous action.The gradient descent-based method also achieves good results.For example,Yu et al.of the University of Science and Technology Beijing proposed a new Differentiable Neural Architecture Search(DNAS)framework[15].During the training process,the weights and activations of different layers are adjusted in their respective bit width search spaces,and their optimal bit widths are automatically explored based on the gradient descent of complexity loss.Compared with the reinforcement learning-based method,this method improves the bit width search speed and avoids separate training structures,but does not explore the root cause of the sensitivity of different layers of the model to bit width sensitivity,which may lead to suboptimal results.

    To sum up,there are some challenges in these methods:

    Challenge 1:Most of the methods do not consider the adaptability to the hardware.

    Challenge 2:These methods either still require a lot of computational resources or are very sensitive to hyperparameters or even initialization.

    Above all,we added a table comparing key characteristics and limitations of Lagrange multiplier,HAQ and DNAS to highlight their differences and limitations.At the same time,the descriptions of the potential shortcomings of our framework have also been addressed in the comparison table,which is shown in Table 1.

    Table 1: Comparison of key characteristics and limitations of prior work

    Because the operation efficiency and accuracy of the quantized neural network model on different accelerators may be quite different,it is necessary to incorporate the accuracy and speed feedback of hardware feedback into the optimization function of bit width search.Therefore,this paper proposes a simple and fast mixed precision search technique.

    3 Proposed Algorithm

    The mixed precision automatic bit width assignment technology sets different quantization bit widths for the convolution weights and activations of different layers of the model to achieve the optimal balance between model accuracy and efficiency.Traditional neural network model quantization methods adopt uniform quantization bit widths for the weights and activations of the entire network model.However,the contribution of the internal layers of the neural network model to the prediction is not exactly the same,so it is necessary to adopt different quantization bit widths for different convolutional layers with different accuracy requirements to achieve the best balance between accuracy and efficiency.

    The proposed method determines a hardware-aware mixed precision quantization strategy adapted to low bit width.Since a brute force approach is not feasible for deep networks,as the search space for mixed precision is exponential complexity in the number of layers.The proposed algorithm is a novel solution which uses mixed precision quantization to reduce the parameter size as well as computational complexity of neural network models.Our challenge is a similar factorial complexity for determining layer-wise finetuning order when quantizing the model to a target precision.Our works use execution speed feedback as metrics to measure computation complexity.

    3.1 Structure

    In this paper,a mixed precision quantization bit width assignment framework adapted to low bit widths is proposed,and reinforcement learning is used to automatically predict mixed precision that satisfies hardware resource constraints.

    For each layer,the standard deviation of the weights is used in the state-space design to measure the distribution difference of the data,and the agent receives the layer configuration and statistical information feedback from the hardware as an observation,and then outputs the action behavior of the layer,that is,the quantization bit width of the weight and activation.When all layers are quantified,the execution speed feedback of simulated neural network accelerator inference is used as an environment to limit the action space of the agent to guide the agent to meet resource constraints.Then,according to the new quantization width,the model is briefly retrained to restore performance,and the accuracy of the retrained quantization model is used as the reward function to guide the agent to carry out deep reinforcement learning training.Finally,different quantization width optimal assignment methods are realized after training,and the quantization accuracy suitable for all layers is output.The overall structure is shown in Fig.1.

    Figure 1: Overall structure

    As shown in Fig.1,this paper uses reinforcement learning to automatically search for a large quantization design space in a loop.Given the number of computing resources of the neural network accelerator,an optimal quantization strategy selection method is proposed.The agent integrates the simulation of the accelerator into the neural network model parameter update process of reinforcement learning,so that the agent can receive direct feedback from the hardware.The mixed precision quantization framework of the whole neural network model for reinforcement learning is divided into two processes:Reinforcement learning process and quantization retraining process.The former contains state-space design,reward function design and reinforcement learning training strategy,the latter contains hardware feedback strategy and symmetric quantization strategy,respectively.

    3.2 Reinforcement Learning Process

    3.2.1 State-Space Design

    In reinforcement learning,learners and decision-makers are called agents,and everything outside the agent is the environment,and the agent and environment interact at each step in a series of discrete time steps.At each time step,the agent receives the state of the environment,on the basis of which an action is selected to execute,and the agent receives a reward value and is in a new state.

    Since our agent processes neural networks in a layer-by-layer fashion,the state-space design is the set of each layer of weights/activations of all possible environmental states,as shown in Table 2.

    Table 2: State-space design table

    3.2.2 Reward Function Design

    After all layers have been quantized,the quantization model is periodically fine-tuned,and the model accuracy after short-term retraining is used as the reward value of the agent.Define a functionrewardwhose reward is directly related to precision:

    wherecost_ratio=is the compression ratio.acc_limitis the given precision constraint value.acc_orginis the top-1 accuracy of the full 8 bit-precision model before quantization on the training set,andacc_quantis the accuracy of the quantization model after finetuning.δis to prevent the occurrence of zero risk,which can be set to 0.0001.

    3.3 Quantization Retraining Process

    3.3.1 Hardware Feedback Strategy

    After the agent provides actions for all layers,the amount of resources used by the quantized model is measured,and the feedback comes directly from the delay calculation table of the hardware accelerator [16],which is used as an environment to obtain direct feedback from the hardware to guide the agent to determine the quantization strategy selection from the nuances between different layers and meet the resource constraints.The speedometer is inferred from the speed evaluation model established in advance by different models on different hardware accelerators.

    To adapt to existing accelerators,the hybrid bit width search space of weights is (1,4,8) and activation is(2,4,8).The weight bit width of the next layer is equal to the activation bit width of the previous layer.

    After the agent provides actions to all layers,the quantization model measures the number of resources that will be used.If the current quantization cost exceeds the target constraint,the bit width of each layer is reduced sequentially until the resource constraint is finally satisfied.The order in which the action space is restricted is shown in the following pseudocode:

    The proposed method encourages the agent to meet the computing resource budget by constraining the action space.The execution speed feedback comes directly from the hardware.

    3.3.2 Symmetric Quantization Strategy

    The quantization algorithm used in this paper is based on the neural network quantization software proposed by Xi’an Microelectronics Technology Institute [17].The quantization software generates a guide file for the neural network accelerator to map floating-point data into low-precision data.The schematic diagram of the quantization software is shown in the Fig.2.

    Figure 2: Schematic diagram of quantization software

    The quantization software uses the layers as the basic processing unit.A pseudo-quantization operation is inserted into the weight and output of each layer[18].Firstly,the distribution and range of input,weight,and output data of each layer of the neural network model are analyzed in advance.Then,each floating-point value is represented by a low-bit integer.The output of the current layer is used as the input of the next layer.Note here that the activation threshold is calculated from the activation equalization of a set of quantization calibration images.

    4 Numerical Simulation and Results Analysis

    In this section,simulation experiments are carried out to verify the proposed method’s effectiveness.First,we describe the architecture of the proposed method in Section 4.1.Second,quantization perceptual hyperparameter selection experiment is conducted in Section 4.2.Then,most importantly,experimental results are given from three aspects in Section 4.3: Quantization strategy assignment results,hybrid bit width search results and comparison with previous work.In addition,we perform the ablation study of the proposed algorithm itself from the perspectives of standard deviation and optimized reward in Section 4.4.The experiments were carried out on Ubuntu 18.04.The Central Processing Unit (CPU) is Intel(R) Core(TM) i7-8700K,3.70 GHz,the Graphics Processing Unit(GPU)is NVIDIA GeForce GTX1070,Deep Learning Framework is Pytorch 1.7.0,Compute Unified Device Architecture (Cuda) is 10.1.The execution speed feedback comes directly from TIANJI NPU4.0 neural network accelerator proposed by Xi’an Microelectronics Technology Institute[19].

    4.1 Experimental Architecture

    The experimental architecture is shown in Fig.3,which uses datasets and models as input,hybrid bit width search is performed on classification tasks through reinforcement learning and simple quantization retraining.It converts the initial training models such as Open Neural Network Exchange(ONNX),TensorFlow,and Caffe provided by third parties into a PyTorch format.Based on the optimal bit width obtained by search,single quantization retraining is carried out again to update the weights of the model.This completes the training phase.In the testing phase,the model with mixed bit width configuration is deployed on the inference platform to obtain accuracy calculation results.To evaluate the effectiveness of the proposed algorithm across a variety of Deep Neural Networks(DNNs),three neural network test models MobileNet-V2,ResNet-50 and LeNet are selected in different target classification task.

    Figure 3: Experimental architecture

    4.2 Hyperparameter Selection

    To adapt to existing accelerators,the weight bit width of the next layer is equal to the activation bit width of the previous layer.And when the activation bit width of the upper layer is 2-bits,the weight of the next layer can only be 1-bit.After all layers are quantified,if the current quantization policy exceeds the resource budget,resource constraints are applied.Update activations and weights layer by layer in reverse order,modifying the weights of this layer while changing the activations of the previous layer.

    There are many hyperparameters in the hybrid bit-width framework,such as the initial value of the learning rate,the rate decay,the training set size,and the quantization retraining epochs.Fig.4 lists only the most important quantization perception hyperparameters—the results of experiments with different quantization retraining epochs.In this experiment,two sets of fixed mixed precision quantization strategies of LeNet model were carried out on the Modified National Institute of Standards and Technology (MNIST) dataset.The accuracy after quantization retraining was used as the evaluation index.The initial value of learning rate is 0.01,the learning rate decay is 0.8,the number of quantization retraining is 100,and the training set size is 200.Repeating the experiment 3 times,selecting the average statistical results.

    Fig.4 shows that with the increase of quantization retraining epochs,the accuracy of different mixed precision quantization strategies increases,and the accuracy is lower than that of the first group because the quantization width of the second group of mixed precision quantization strategies is lower and more difficult to quantize.

    Figure 4: Accuracy with different quantization retraining epochs

    4.3 Experimental Results

    In this section,we first present our quantization strategy assignment results for MobileNet-V2 on ImageNet,and LeNet on MNIST.Then we present our hybrid bit width search results for MobileNet-V2 on UCMerced LandUse.Finally,comparison with prior work is carried out.

    4.3.1 Quantization Strategy Assignment Results

    The maximum compression ratio is set to 0.6,and the episodes of Reinforcement Learning(RL)training is set to 50.The number of quantization retraining epochs for this experiment is set to 100.Repeat the experiment 3 times,selecting the average statistical results.

    For MobileNet-V2 on ImageNet dataset,Table 3 shows the two sets of hybrid bit width assignment strategies with the best accuracy in multiple iterations.Due to the excessive number of layers in the MobileNet-V2 model,the hybrid bit width quantization strategy representation is simplified to layer by layer input/weight/activation bit width-[layer 1 [input,weight,activation],layer 2 [input,weight,activation]...Layer n[input,weight,activation]].It can be seen from Table 3 that after replacing the dataset and model,the influence of low bit-to-width ratio on model accuracy is also reflected.

    Table 3: Quantization strategy assignment for MobileNet-V2 on ImageNet

    For the LeNet model,a speedometer with dimension 6 ?3 ?3 is established.3 ?3 is the total number of combinations assigned bit widths of each convolutional layer,with a weight distribution of 1/4/8 bit and an activation distribution of 1/4/8 bit.In multiple iterations,select the two sets of mixed precision quantization strategies with the best accuracy.The intra-layer/inter-layer bit width assignment of different strategies and the corresponding evaluation metrics are shown in Table 4.

    Table 4: Quantization strategy assignment for LeNet on MNIST

    Experiments of LeNet on the MNIST dataset show that a group of hybrid bit width strategies with low compression ratio have low accuracy.It shows that the greater the proportion of low bit width,the greater the accuracy of the model.The larger the reward value,the higher the model’s accuracy,which shows that the more accurate the positive feedback is based on the accuracy of the quantized model after retraining,the more effective it can guide the agent to carry out deep reinforcement learning training,and the bit width strategy that satisfies the hardware resources can be obtained.

    4.3.2 Hybrid Bit Width Search Results

    To further verify the effectiveness of reinforcement learning in mixed-precision search,this section details the evolution of reward,loss,and evaluation metrics in RL training process.Evaluation metrics include the compression ratio of the search hybrid bit width and the accuracy of retrained quantization model.In order to fully improve the accuracy of the retrained quantization model,the maximum compression ratio of the search bit width is increased to 0.9,and the episodes of reinforcement learning training is set to 30.Moreover,the goal of mixed precision search technique is to find a suitable hybrid bit width assignment strategy in the complex bit width selection space.The bit width search of each time only requires simple quantization retraining,so the epochs of quantization retraining for this experiment is set to 30.

    The hybrid bit width search experiment of MobileNet-V2 on UCMerced LandUse dataset is shown in Figs.5–7.It can be seen from Fig.5 that,due to the new design of the reward function,the reward becomes larger as the RL training episodes increase,and the reward value reaches a maximum of 9.053 at the 30th iteration.This fully demonstrates that this reward function,as an evaluation criterion of the agent’s task execution,can encourage the agent to make good action decisions and discourage the agent from making action decisions that lead to bad outcomes.This constantly updates the mixed-precision quantization strategies to get as many rewards as possible.

    Figure 5: The evolution of reward in RL training

    Fig.6 shows that with the increase of RL trainings episodes,the accuracy of retrained quantization model is increasing,which reaches a maximum of 0.776 at the 30th iteration,and the error with the accuracy of 0.800 of the full-precision model is 2.5%.Meanwhile,the compression ratio of the best accuracy reaches 0.775.In particular,at the 15th iteration,the accuracy reaches 0.724,but the compression ratio reaches 0.848.In order to achieve the goal of lower compression ratio and higher accuracy,reinforcement learning once again exerts powerful decision-making capabilities.This gradually reduces the compression ratio and maintains a consistently increasing accuracy.

    According to the mixed-precision quantization framework,the model assigned new quantization widths is briefly retrained to restore performance,and the accuracy of the retrained model is used as the reward to guide the agent to carry out deep reinforcement training.To illustrate the role of quantization retraining,Fig.7 shows the evolution of loss varying seven different RL learning.On the one hand,the loss value in each RL training decreases steadily with the increase of the epochs of quantization retraining.Although there is a tendency to become larger in some places,this is due to the robustness of model training.On the other hand,with the increase of the order of 1,3,8,13,17,25 and 30 for different RL training episodes,the overall range of loss values also decreases.Furthermore,it is shown that the accuracy of the evaluated model steadily increases in the RL training process.

    Figure 6: The evolution of compression ratio and accuracy in RL training

    Figure 7: The evolution of loss varying different RL learning in quantization retraining

    4.3.3 Comparison with Prior Work

    In order to verify the effectiveness of the improved quantization strategy assignment method,the comparative experiments with related works are carried out.Experiments are performed on target classification task.We compare with the full-precision MobileNet-V2,ResNet-50 and LeNet,and popular state-of-the-art hybrid bit width search methods including Parameterized Clipping activation(PACT) [20],Han et al.[21],HAQ [14],Reinforcement Learning approach for deep Quantization(ReLeQ)[22],DoReFa[23]and Alternating Direction Method of Multipliers(ADMM)[24].Above all,we apply these methods to search for optimal hybrid bit widths.

    The comparative experiment for MobileNet-V2 on ImageNet dataset is shown in Table 5.We compare our framework with PACT and Han that uses fixed number of bits without hardware feedback.And HAQ,ReLeQ that uses flexible number of bits with hardware feedback.It can be seen from Table 5,our method performs similarly to the full-precision baseline on the same hardware platform.

    Table 5: Comparative experiment for MobileNet-V2 on ImageNet

    Table 6 shows the comparative experiment for ResNet-50 on ImageNet.Here,we add the traditional method DoReFa that uses fixed number of bits without hardware feedback as a comparison.Under 1,2,4,8 mixed precision configuration,the quantized ResNet-50 has also achieved relatively small quantization errors by 0.024 than all comparative methods shown.It has certain performance competitive advantages.

    Table 6: Comparative experiment for ResNet-50 on ImageNet

    For comparison,Table 7 records the comparative experiment for LeNet on MNIST with prior work.Our method achieves minimum quantization variations of 0.055 and 0.003 compared to the fullprecision baseline at the two-bit width allocation schemes,respectively.Again,in Table 7,we notice that a group of hybrid bit width strategies of our method with high average bit width have low accuracy.It indicates that the greater the proportion of low bit width,the greater the accuracy of the model.

    Table 7: Comparative experiment for LeNet on MNIST

    4.4 Ablation Study

    4.4.1 W/and W/O Standard Deviation

    The design adds the standard deviation of weights to the state-space design.The number of searches is set to 50,the number of quantization retraining is set to 100,in order to make the comparison with the addition of the weight standard deviation more obvious,the hybrid bit width assignment strategy with the best verified accuracy is shown in the following table for the two models of MobileNet-V2 and LeNet.

    The hybrid bit width assignment strategies selected in Table 8 are all have low bit width,that is,there is a combination of 2-bit input and 1-bit weight in one or more layers.Experiments show that reinforcement learning can fully measure the distribution differences of data in different layers in the mixed precision quantization bit width search process.Although the compression ratio of the model is slightly increased,the quantization deployment accuracy of the model is improved in the limited search space,and a suitable quantization bit width assignment strategy is found for each layer.

    Table 8: Comparative experiment with the addition of“weighted standard deviation”

    4.4.2 Comparison between Optimized Reward Function

    The goal of reinforcement learning is to obtain optimal decisions so that the agents controlling the mixed-precision quantization strategies receive the greatest rewards.The design of reward function is a crucial component of reinforcement learning as indicated in Section 3.2.In this paper,we incorporate reinforcement learning techniques by proposing a special parametric reward formulation.To evaluate the effectiveness of the proposed reward formulation,we have compared the mixed-precision search effect of the models before and after the reward function optimization in Table 9.

    Table 9: Comparative experiment with optimized reward function

    Table 9 shows the accuracy and compression ratio results of MobileNet-V2 and LeNet,where“√”indicates that the corresponding reward function has been optimized.It can be seen from the comparative results that no matter which model is evaluated,no matter which dataset is based on,although the compression ratio of the quantization strategy searched by the optimized reward function is slightly higher,the accuracy of retrained quantization model becomes higher.Under the condition that the hardware resources are satisfied,the proposed reward formulation is consistently achieving higher accuracy during the reinforcement learning training episodes.Our agent gave quite different quantization strategies for accelerators.

    5 Conclusion

    In conclusion,a quantization strategy optimal assignment algorithm for mixed precision is proposed,and reinforcement learning is used to automatically predict the optimal bit width that meets the constraints of hardware resources.The proposed method improves the classification accuracy loss of 2.6%,2.4%and 0.3%on the MobileNet-V2,ResNet-50 and LeNet compared to the full-precision baseline,respectively.It effectively achieves the compromise between the performance and accuracy of deep neural networks.

    Future research opportunities stemming from the findings presented include:(1)Mixed-precision search based on reinforcement learning can be applied to tasks such as detection,semantic segmentation,and speech recognition,etc.;(2)The execution speed feedback can also be obtained from other localized embedded neural network accelerators.These can be widely deployed on various embedded computing platforms with low power consumption and limited resources;(3) Even though we have shown benefits of ours as compared to DNAS or HAQ,it may be possible to combine these methods for more efficient AutoML search.We leave this as part of future work.

    Acknowledgement:I would like to express my sincere gratitude to Prof.Ma Zhong for his guidance and support.Our conversations inspired me to write the entire work and complete it successfully.I also want to thank the research team for their collaboration and help during gathering data for my research project.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:The authors confirm contribution to the paper as follows: Study conception and design:Y.Wang,Z.Ma;data collection:C.Yang,L.Wei;analysis and interpretation of results:Y.Wang,Y.Yang;draft manuscript preparation:Y.Wang,Z.Ma.All authors approved the final version of the manuscript.

    Availability of Data and Materials:The data that support the findings of this study are available from the corresponding author,Z.Ma,upon reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    长腿黑丝高跟| 人人妻,人人澡人人爽秒播| 美女cb高潮喷水在线观看 | 老司机午夜福利在线观看视频| 99re在线观看精品视频| 色综合婷婷激情| 69av精品久久久久久| 免费av不卡在线播放| 国产精品久久久久久精品电影| 青草久久国产| 黑人巨大精品欧美一区二区mp4| 亚洲国产精品sss在线观看| 免费在线观看日本一区| 午夜福利在线观看吧| 啦啦啦韩国在线观看视频| 欧美成人性av电影在线观看| 老汉色av国产亚洲站长工具| 国产精品久久久久久精品电影| 亚洲一区二区三区不卡视频| 性色avwww在线观看| 亚洲av片天天在线观看| 亚洲av电影不卡..在线观看| 黄色视频,在线免费观看| 人人妻人人看人人澡| 一卡2卡三卡四卡精品乱码亚洲| 麻豆国产av国片精品| 女同久久另类99精品国产91| 啦啦啦免费观看视频1| 后天国语完整版免费观看| 综合色av麻豆| 免费av毛片视频| 噜噜噜噜噜久久久久久91| 757午夜福利合集在线观看| 国产亚洲精品久久久com| 巨乳人妻的诱惑在线观看| 我的老师免费观看完整版| 欧美日韩国产亚洲二区| 女警被强在线播放| 国产一区二区在线av高清观看| 动漫黄色视频在线观看| 欧美黄色淫秽网站| 日韩欧美三级三区| 人人妻,人人澡人人爽秒播| 99国产精品99久久久久| 老鸭窝网址在线观看| a级毛片a级免费在线| 中国美女看黄片| 亚洲片人在线观看| 在线十欧美十亚洲十日本专区| e午夜精品久久久久久久| 国产一区二区激情短视频| 99久久久亚洲精品蜜臀av| 国产伦在线观看视频一区| 久久人妻av系列| 免费看日本二区| 国产亚洲欧美98| 精品无人区乱码1区二区| 宅男免费午夜| 国语自产精品视频在线第100页| 国产高清videossex| 午夜激情欧美在线| 午夜福利在线观看吧| 欧美黑人欧美精品刺激| 91字幕亚洲| 欧美不卡视频在线免费观看| 国产亚洲精品综合一区在线观看| 国产日本99.免费观看| 夜夜爽天天搞| 变态另类成人亚洲欧美熟女| 男插女下体视频免费在线播放| av天堂中文字幕网| 亚洲第一电影网av| 成人性生交大片免费视频hd| 亚洲精品中文字幕一二三四区| 亚洲精品一区av在线观看| 成人高潮视频无遮挡免费网站| 精品久久久久久成人av| 亚洲精品中文字幕一二三四区| 久久久久久久久久黄片| 一进一出好大好爽视频| www日本在线高清视频| 日韩欧美国产在线观看| 99久久国产精品久久久| 99热只有精品国产| 美女大奶头视频| 日本免费一区二区三区高清不卡| 亚洲午夜精品一区,二区,三区| 99热6这里只有精品| avwww免费| 免费无遮挡裸体视频| 夜夜躁狠狠躁天天躁| 一个人免费在线观看电影 | 国产 一区 欧美 日韩| 日本 欧美在线| 深夜精品福利| 日本 av在线| 一区二区三区高清视频在线| 久久久久久大精品| 一二三四社区在线视频社区8| 美女免费视频网站| 午夜福利高清视频| 欧美日韩综合久久久久久 | 亚洲国产欧美一区二区综合| 久久中文字幕一级| 19禁男女啪啪无遮挡网站| 精品乱码久久久久久99久播| 巨乳人妻的诱惑在线观看| 亚洲av成人不卡在线观看播放网| 亚洲精品一卡2卡三卡4卡5卡| 国产成人精品久久二区二区免费| 久久热在线av| 在线播放国产精品三级| 少妇的丰满在线观看| 国产精品亚洲av一区麻豆| 麻豆成人午夜福利视频| 夜夜爽天天搞| 欧美乱妇无乱码| 黄色片一级片一级黄色片| 草草在线视频免费看| 久久久久久九九精品二区国产| 日日夜夜操网爽| 搡老熟女国产l中国老女人| 国产精品98久久久久久宅男小说| 一级毛片精品| 国产激情偷乱视频一区二区| 欧美成狂野欧美在线观看| 最近在线观看免费完整版| tocl精华| 国产亚洲av高清不卡| 亚洲人成网站高清观看| 成人亚洲精品av一区二区| 黑人欧美特级aaaaaa片| 国产在线精品亚洲第一网站| 不卡av一区二区三区| 淫妇啪啪啪对白视频| 天堂动漫精品| 啦啦啦免费观看视频1| 久久国产乱子伦精品免费另类| 久久国产乱子伦精品免费另类| 宅男免费午夜| 久久中文字幕人妻熟女| 亚洲欧美日韩卡通动漫| АⅤ资源中文在线天堂| 亚洲精品粉嫩美女一区| 欧美日韩乱码在线| 精品国产乱码久久久久久男人| 亚洲av美国av| 搡老妇女老女人老熟妇| 欧美日韩乱码在线| 国产亚洲av嫩草精品影院| 狂野欧美白嫩少妇大欣赏| 欧美黄色淫秽网站| 亚洲av电影不卡..在线观看| 亚洲五月天丁香| 国产精品久久久人人做人人爽| 国产精品98久久久久久宅男小说| 国产一区二区三区视频了| 啦啦啦免费观看视频1| 人人妻,人人澡人人爽秒播| 99久久久亚洲精品蜜臀av| 欧美+亚洲+日韩+国产| 国产乱人视频| 老熟妇仑乱视频hdxx| 中文亚洲av片在线观看爽| 精品国产亚洲在线| 成人特级黄色片久久久久久久| 久久香蕉精品热| 手机成人av网站| 国产精品野战在线观看| 人妻丰满熟妇av一区二区三区| 黄色丝袜av网址大全| 制服人妻中文乱码| 校园春色视频在线观看| 亚洲精品乱码久久久v下载方式 | 麻豆av在线久日| 日韩高清综合在线| 成在线人永久免费视频| 美女高潮喷水抽搐中文字幕| 亚洲精品美女久久av网站| 在线观看美女被高潮喷水网站 | 天堂√8在线中文| 免费高清视频大片| 色吧在线观看| 一个人免费在线观看的高清视频| 久久久久久久午夜电影| 欧美性猛交黑人性爽| 91久久精品国产一区二区成人 | 99re在线观看精品视频| 国产99白浆流出| 国产又黄又爽又无遮挡在线| 亚洲第一电影网av| 亚洲av成人一区二区三| 全区人妻精品视频| 黄色 视频免费看| 一级黄色大片毛片| 国产视频一区二区在线看| 国内精品美女久久久久久| 人人妻人人澡欧美一区二区| 成人国产一区最新在线观看| 亚洲乱码一区二区免费版| 19禁男女啪啪无遮挡网站| 国模一区二区三区四区视频 | 欧美不卡视频在线免费观看| 三级男女做爰猛烈吃奶摸视频| 又粗又爽又猛毛片免费看| 日本五十路高清| 久久亚洲精品不卡| 哪里可以看免费的av片| 亚洲av第一区精品v没综合| 黄片大片在线免费观看| 成年女人毛片免费观看观看9| 成人鲁丝片一二三区免费| 免费人成视频x8x8入口观看| 性色av乱码一区二区三区2| tocl精华| 成熟少妇高潮喷水视频| 久久久久免费精品人妻一区二区| 国产精品精品国产色婷婷| 麻豆成人av在线观看| 欧美三级亚洲精品| 人人妻,人人澡人人爽秒播| 草草在线视频免费看| 国产精品日韩av在线免费观看| 成人午夜高清在线视频| 国产伦人伦偷精品视频| 一二三四社区在线视频社区8| 最近最新中文字幕大全免费视频| 国产主播在线观看一区二区| 中文字幕人妻丝袜一区二区| 久久久精品欧美日韩精品| 欧美在线一区亚洲| 午夜福利在线观看吧| 精品熟女少妇八av免费久了| 熟女人妻精品中文字幕| 成年免费大片在线观看| 亚洲欧美日韩高清专用| 色精品久久人妻99蜜桃| 又爽又黄无遮挡网站| 亚洲成av人片在线播放无| 窝窝影院91人妻| 国产综合懂色| 两个人看的免费小视频| 麻豆成人午夜福利视频| a级毛片在线看网站| 一级作爱视频免费观看| 久久伊人香网站| 三级男女做爰猛烈吃奶摸视频| 色在线成人网| 国产精品一及| 91麻豆精品激情在线观看国产| 国产一区二区激情短视频| 久久九九热精品免费| 国产精品永久免费网站| 午夜日韩欧美国产| 日日夜夜操网爽| 亚洲在线自拍视频| 脱女人内裤的视频| 国产精品一区二区精品视频观看| 久久久国产欧美日韩av| 亚洲av免费在线观看| 母亲3免费完整高清在线观看| 国产精华一区二区三区| av女优亚洲男人天堂 | 熟妇人妻久久中文字幕3abv| 亚洲成人久久性| 国产主播在线观看一区二区| 欧美性猛交╳xxx乱大交人| 国产淫片久久久久久久久 | 在线播放国产精品三级| 国产精品乱码一区二三区的特点| 国产精品一区二区三区四区免费观看 | 一个人免费在线观看电影 | 色尼玛亚洲综合影院| 色综合欧美亚洲国产小说| 夜夜躁狠狠躁天天躁| 不卡av一区二区三区| 亚洲av成人不卡在线观看播放网| 亚洲精品一区av在线观看| 午夜a级毛片| 黄色视频,在线免费观看| 两性夫妻黄色片| 熟女少妇亚洲综合色aaa.| 日日摸夜夜添夜夜添小说| 亚洲精品一卡2卡三卡4卡5卡| 欧美+亚洲+日韩+国产| 精品国产三级普通话版| 久久精品综合一区二区三区| 久久亚洲真实| 成人鲁丝片一二三区免费| 亚洲 欧美一区二区三区| 国产毛片a区久久久久| 精品国产亚洲在线| 亚洲电影在线观看av| 国产又色又爽无遮挡免费看| 久久中文字幕人妻熟女| 黄片大片在线免费观看| 久久精品国产99精品国产亚洲性色| 免费在线观看日本一区| 手机成人av网站| 国产三级在线视频| 老汉色av国产亚洲站长工具| 91久久精品国产一区二区成人 | 国产免费av片在线观看野外av| 久久中文字幕人妻熟女| 亚洲欧美日韩高清专用| 又爽又黄无遮挡网站| 国产1区2区3区精品| 亚洲成人免费电影在线观看| 无遮挡黄片免费观看| 成年女人永久免费观看视频| 免费看十八禁软件| 99久国产av精品| 欧美高清成人免费视频www| 亚洲一区二区三区色噜噜| 在线永久观看黄色视频| 中文字幕熟女人妻在线| 色精品久久人妻99蜜桃| 国产精品久久电影中文字幕| 成人高潮视频无遮挡免费网站| 亚洲av免费在线观看| 久久精品91无色码中文字幕| 特大巨黑吊av在线直播| 免费在线观看日本一区| 一区二区三区激情视频| 日本五十路高清| 色综合站精品国产| 国内精品久久久久久久电影| 欧美日本视频| 精品人妻1区二区| 女生性感内裤真人,穿戴方法视频| 日韩高清综合在线| 黑人欧美特级aaaaaa片| 亚洲成人久久性| 国产一区二区三区在线臀色熟女| 免费在线观看成人毛片| 高潮久久久久久久久久久不卡| 丁香六月欧美| 欧美不卡视频在线免费观看| 97碰自拍视频| 亚洲片人在线观看| 免费观看的影片在线观看| 99热这里只有是精品50| 亚洲成人久久性| 亚洲av日韩精品久久久久久密| 亚洲精品国产精品久久久不卡| 亚洲五月天丁香| 精品99又大又爽又粗少妇毛片 | 精品久久久久久久久久久久久| 变态另类成人亚洲欧美熟女| 色综合站精品国产| 亚洲成人久久爱视频| 看片在线看免费视频| 最好的美女福利视频网| 国产精品久久久久久人妻精品电影| 成熟少妇高潮喷水视频| 亚洲国产日韩欧美精品在线观看 | 不卡一级毛片| 黄色女人牲交| 欧美激情在线99| 国产激情久久老熟女| 不卡一级毛片| 久久久精品欧美日韩精品| 亚洲男人的天堂狠狠| 夜夜看夜夜爽夜夜摸| 亚洲av美国av| 岛国在线免费视频观看| 麻豆国产av国片精品| xxx96com| 欧美三级亚洲精品| 久久精品影院6| 最新在线观看一区二区三区| 国产成人精品久久二区二区免费| 国产一级毛片七仙女欲春2| 最好的美女福利视频网| 一个人免费在线观看电影 | 两个人的视频大全免费| 亚洲精品粉嫩美女一区| 女警被强在线播放| 人人妻人人看人人澡| 国产精品影院久久| 国产成年人精品一区二区| 1024手机看黄色片| 少妇的逼水好多| 亚洲自拍偷在线| 日韩av在线大香蕉| 国产精品 欧美亚洲| 亚洲人成网站在线播放欧美日韩| 亚洲国产欧洲综合997久久,| 一级作爱视频免费观看| 日本熟妇午夜| 亚洲熟妇中文字幕五十中出| 后天国语完整版免费观看| 国产伦一二天堂av在线观看| 久久这里只有精品中国| 18美女黄网站色大片免费观看| 国产精品永久免费网站| 午夜两性在线视频| 亚洲av成人av| 一级作爱视频免费观看| 色综合婷婷激情| 亚洲乱码一区二区免费版| 精品免费久久久久久久清纯| 成人亚洲精品av一区二区| 变态另类成人亚洲欧美熟女| 在线免费观看的www视频| 免费观看的影片在线观看| 日韩大尺度精品在线看网址| 人妻夜夜爽99麻豆av| 19禁男女啪啪无遮挡网站| 亚洲人成伊人成综合网2020| 特级一级黄色大片| 美女午夜性视频免费| 成人无遮挡网站| 天堂网av新在线| 国产高潮美女av| 极品教师在线免费播放| 成熟少妇高潮喷水视频| 黑人操中国人逼视频| 日韩欧美精品v在线| 成人午夜高清在线视频| 欧美中文综合在线视频| 国语自产精品视频在线第100页| 色噜噜av男人的天堂激情| cao死你这个sao货| 老司机福利观看| 亚洲va日本ⅴa欧美va伊人久久| 亚洲av美国av| 免费在线观看成人毛片| 一个人看视频在线观看www免费 | 欧美高清成人免费视频www| 黄色 视频免费看| 久久久久久大精品| 精品久久久久久久久久久久久| 亚洲精品在线美女| 岛国在线免费视频观看| 国产av麻豆久久久久久久| 亚洲自拍偷在线| 国模一区二区三区四区视频 | 日韩 欧美 亚洲 中文字幕| 最近在线观看免费完整版| 免费在线观看视频国产中文字幕亚洲| 在线免费观看的www视频| 91av网站免费观看| www国产在线视频色| 熟女电影av网| 亚洲七黄色美女视频| 亚洲人成网站高清观看| 亚洲欧美日韩无卡精品| 欧美成狂野欧美在线观看| 国产精品一区二区三区四区免费观看 | 一夜夜www| 亚洲天堂国产精品一区在线| 欧美在线黄色| 在线免费观看不下载黄p国产 | 黑人巨大精品欧美一区二区mp4| 99在线人妻在线中文字幕| 桃色一区二区三区在线观看| 亚洲国产欧美网| 国产精品久久久人人做人人爽| 性色avwww在线观看| 国产v大片淫在线免费观看| 女同久久另类99精品国产91| 色噜噜av男人的天堂激情| 久久亚洲真实| 国产欧美日韩精品一区二区| 亚洲国产精品999在线| 十八禁人妻一区二区| 成人精品一区二区免费| 97超视频在线观看视频| 亚洲成a人片在线一区二区| 国产一级毛片七仙女欲春2| 亚洲aⅴ乱码一区二区在线播放| 国产在线精品亚洲第一网站| 日本五十路高清| 国产精品久久久久久人妻精品电影| 天堂网av新在线| 日韩成人在线观看一区二区三区| a在线观看视频网站| 亚洲国产日韩欧美精品在线观看 | 午夜视频精品福利| 欧美日韩乱码在线| 国产精品一区二区免费欧美| 亚洲乱码一区二区免费版| 久久久久九九精品影院| 搡老熟女国产l中国老女人| 精品国产美女av久久久久小说| 非洲黑人性xxxx精品又粗又长| 一个人看的www免费观看视频| 国产午夜精品久久久久久| 国产97色在线日韩免费| 国产精品亚洲一级av第二区| 黄色视频,在线免费观看| 九色国产91popny在线| 日韩欧美精品v在线| 男人的好看免费观看在线视频| 好看av亚洲va欧美ⅴa在| 日韩欧美三级三区| 亚洲第一欧美日韩一区二区三区| 人妻夜夜爽99麻豆av| 久久久久久久精品吃奶| 亚洲av中文字字幕乱码综合| 国产免费av片在线观看野外av| 中文字幕精品亚洲无线码一区| 亚洲电影在线观看av| 一个人观看的视频www高清免费观看 | 成人18禁在线播放| 91av网一区二区| 美女高潮的动态| 嫩草影视91久久| 午夜免费激情av| 免费无遮挡裸体视频| 亚洲无线在线观看| 男女那种视频在线观看| 成人三级黄色视频| 国产成人av教育| 亚洲一区二区三区色噜噜| 一本久久中文字幕| 三级男女做爰猛烈吃奶摸视频| 观看美女的网站| 国产亚洲精品av在线| 欧美性猛交黑人性爽| 国产人伦9x9x在线观看| 日韩成人在线观看一区二区三区| 国产淫片久久久久久久久 | 午夜a级毛片| 亚洲欧美精品综合久久99| 午夜福利在线在线| tocl精华| 婷婷亚洲欧美| 亚洲精品色激情综合| 精品久久蜜臀av无| 久久久久亚洲av毛片大全| 欧美成狂野欧美在线观看| 午夜两性在线视频| 亚洲国产欧美一区二区综合| 国产精品国产高清国产av| 精品国产乱码久久久久久男人| 人人妻人人看人人澡| 成年女人永久免费观看视频| 啦啦啦观看免费观看视频高清| 91av网一区二区| 成人av在线播放网站| 亚洲中文字幕一区二区三区有码在线看 | 亚洲精品久久国产高清桃花| 亚洲色图av天堂| а√天堂www在线а√下载| 欧美乱码精品一区二区三区| 变态另类丝袜制服| 淫妇啪啪啪对白视频| 精品乱码久久久久久99久播| 国产高清videossex| 99久久精品热视频| 精品午夜福利视频在线观看一区| 大型黄色视频在线免费观看| 亚洲欧美日韩高清专用| 亚洲中文字幕日韩| 最近最新免费中文字幕在线| 757午夜福利合集在线观看| 国产亚洲欧美在线一区二区| 天堂网av新在线| 亚洲精品美女久久久久99蜜臀| 欧美大码av| 欧美午夜高清在线| 最新美女视频免费是黄的| 国产又色又爽无遮挡免费看| 亚洲欧美激情综合另类| 久久草成人影院| 身体一侧抽搐| 欧美性猛交黑人性爽| 蜜桃久久精品国产亚洲av| 午夜福利在线在线| 精品一区二区三区四区五区乱码| 三级国产精品欧美在线观看 | 欧美日韩国产亚洲二区| 亚洲狠狠婷婷综合久久图片| 淫妇啪啪啪对白视频| 免费电影在线观看免费观看| 成熟少妇高潮喷水视频| 久久人人精品亚洲av| 日本撒尿小便嘘嘘汇集6| 亚洲av成人不卡在线观看播放网| 后天国语完整版免费观看| 国产视频一区二区在线看| 一边摸一边抽搐一进一小说| 国产免费男女视频| 亚洲性夜色夜夜综合| 一级毛片女人18水好多| 久久精品影院6| 黑人欧美特级aaaaaa片| 他把我摸到了高潮在线观看| 欧美极品一区二区三区四区| 久久久国产欧美日韩av| 精品无人区乱码1区二区| 成人av一区二区三区在线看| 国产淫片久久久久久久久 | 丰满人妻一区二区三区视频av | 亚洲精华国产精华精| 国产毛片a区久久久久| 亚洲黑人精品在线| 国产精品1区2区在线观看.| 欧美xxxx黑人xx丫x性爽| 一级黄色大片毛片| 桃红色精品国产亚洲av| 夜夜看夜夜爽夜夜摸| 性色avwww在线观看| 久久国产精品影院| 欧美黑人巨大hd| 久久精品91蜜桃| 日韩有码中文字幕| 久久久国产欧美日韩av| 久久久久九九精品影院| 欧美一区二区精品小视频在线| 欧美日韩福利视频一区二区| 18禁观看日本| 97超级碰碰碰精品色视频在线观看|