CN113408723A - Convolutional neural network pruning and quantization synchronous compression method for remote sensing application - Google Patents

Convolutional neural network pruning and quantization synchronous compression method for remote sensing application Download PDF

Info

Publication number
CN113408723A
CN113408723A CN202110545477.8A CN202110545477A CN113408723A CN 113408723 A CN113408723 A CN 113408723A CN 202110545477 A CN202110545477 A CN 202110545477A CN 113408723 A CN113408723 A CN 113408723A
Authority
CN
China
Prior art keywords
model
training
pruning
quantization
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110545477.8A
Other languages
Chinese (zh)
Other versions
CN113408723B (en
Inventor
陈禾
齐保贵
陈亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202110545477.8A priority Critical patent/CN113408723B/en
Publication of CN113408723A publication Critical patent/CN113408723A/en
Application granted granted Critical
Publication of CN113408723B publication Critical patent/CN113408723B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Compared with other convolutional neural network pruning and quantization compression methods, the convolutional neural network pruning and quantization synchronous compression method for remote sensing application integrates the model pruning and quantization processes, realizes synchronous pruning and quantization of the convolutional neural network model, and reduces the precision loss after model compression while improving the compression ratio of model parameters; the model after pruning and quantization is retrained, so that more accurate parameter values can be obtained, and the network precision is improved; setting a rule degree constraint condition to be observed during coding, namely, when a convolutional neural network model is realized in an actual remote sensing platform processor, pruning convolutional kernels of different filters in the same layer at the same position simultaneously to improve the universality degree of a computing unit and improve the parallelism degree of computation; the compressed model can be transplanted and applied to platform processors with limited resources such as airborne and satellite-borne platforms.

Description

Convolutional neural network pruning and quantization synchronous compression method for remote sensing application
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to a convolutional neural network compression method for remote sensing application.
Background
With the wide application of various convolutional neural network models in the field of remote sensing data processing, the requirement for transplanting the convolutional neural network models to mobile platforms such as airborne platforms, satellite-borne platforms and vehicle-mounted platforms is more and more strong. When the method is applied to platforms with strict resource constraints, due to the fact that the resources of selectable special devices are limited, reliability needs to be guaranteed through multi-mode redundancy and the like, the resources which can be used for achieving the convolutional neural network model are very limited. The high-performance convolutional neural network often has a large number of parameters and floating point calculation times, is difficult to apply on platforms with limited resources, and needs a series of parameter compression work such as pruning, quantization and the like. However, the current convolutional neural network pruning and quantization compression method has the problem of large loss of model precision after compression. Both the pruning and quantization processes affect the model accuracy, and the accuracy losses are superposed after the two steps of compression. Pruning first and then quantization will result in a large loss of model accuracy.
Disclosure of Invention
In view of this, the present invention provides a convolutional neural network compression method for remote sensing applications, which improves the compression ratio of model parameters and ensures the accuracy of the compressed model.
A convolution neural network pruning and quantization synchronous compression method for remote sensing application comprises the following steps:
s1, performing parameter training on the convolutional neural network by adopting data containing the remote sensing image to obtain a model M2;
s2, pruning and coding the quantization parameters of the model M2, which are specifically as follows:
for the parameters of each layer in the convolutional neural network, its coding is defined as:
Ok=(p1,p2,...,pn,q1,q2,...,qn)
wherein n represents the number of parameters of each layer in the network; p is a radical of1,p2,...,pnThe value of 0 or 1 is used for representing whether each parameter is pruned, wherein 0 represents that the parameter is pruned, and 1 represents that the parameter is reserved and not pruned; q. q.s1,q2,...,qnRepresenting the number of quantization bits for each parameter;
s3, generating a set number of initial populations according to the codes of the step S2, wherein each population comprises a plurality of individuals;
s4, setting parameter pruning and quantification in the model M2 according to the codes of each individual in the population in the step S3, so as to obtain models corresponding to the number of the individuals, and training parameter values of the models by adopting remote sensing image data to obtain the models after parameter training;
s5, performing precision evaluation on the model obtained after parameter training, removing the part of individuals with the worst precision in each population, randomly changing the codes of the rest individuals, and performing cross and/or migration on the codes among the individuals to generate new individuals and populations;
s6, repeating the step S5 until the set conditions are met, and stopping training the model parameters;
in the training process, the parameter value of the model used in the current training is the parameter value of the last training;
and S7, performing precision evaluation on the model obtained in the S6, and reserving a part of individuals with the highest precision to finish pruning and quantification of the convolutional neural network.
Further, retraining the partial individuals with the highest accuracy in S7 specifically includes:
firstly, constructing a convolution neural network model of floating point parameters according to S7 pruning results, and training the network model parameters by adopting training data;
then, on the basis of the parameters of the post-pruning model, training the quantitative parameters of each layer of the model in turn, specifically: and (3) quantizing the parameters of the first layer of the network model according to the parameter quantization result of S7, training the whole convolutional neural network on the basis of the pruned model parameters, fixing the quantized parameters of the first layer after finishing the training of the other layers of the network model as floating point parameters, quantizing the parameters of the second layer according to the result of S7, keeping the floating point parameters of the third layer and the later layers, training the network, and so on until the last layer of the network, and finishing the final training of the model parameters.
Further, in S2, the rule constraint condition is met during encoding, specifically: when the constraint only prunes the filter level of the convolutional neural network, the same pruning coding is needed for the parameters in the same filter; when the constraint only prunes at the convolution kernel level, all parameters in the same convolution kernel require the same pruning coding.
Preferably, in S3, the method for generating the initial population and the individuals comprises:
randomly encoding each layer of the network under the condition of satisfying parameter compression ratiok=(p1,p2,...,pn,q1,q2,...,qn) Assigning the value of the middle element;
or, the existing pruning and quantization algorithm is adopted to obtain a pruned or quantized model structure, and the codes of the individuals in the initial population are assigned according to the structure to obtain the individuals.
Preferably, the setting conditions in S6 are: the training is stopped when the accuracy, the parameter amount and the calculated amount of any one individual reach a specified threshold or reach a predetermined number of times of training.
Preferably, when the model parameters are trained, the number of training cycles of each model is the same, and the training is performed for the least number of times under the condition that the model performance can be distinguished.
Preferably, when the model accuracy is evaluated in S5, if the model accuracy is consistent, the individual with a large number of model parameters is rejected.
Preferably, when the model accuracy is evaluated in S5, the evaluation method is selected according to different task types.
Preferably, in S5, encoding crossover is performed on different individuals in the same population, the optimal individual in each population is migrated to other populations, and the individual and the population are updated.
Preferably, in S1, when the remote sensing image data is insufficient, the natural scene image data is used to perform parameter training on the network to obtain a pre-training model M1; and then carrying out parameter training on the pre-training model M1 by using the remote sensing image data to obtain a model M2.
The invention has the following beneficial effects:
compared with other convolutional neural network pruning and quantization compression methods, the convolutional neural network pruning and quantization synchronous compression method for remote sensing application integrates the model pruning and quantization processes, realizes synchronous pruning and quantization of the convolutional neural network model, and reduces the precision loss after model compression while improving the compression ratio of model parameters;
the model after pruning and quantization is retrained, so that more accurate parameter values can be obtained, and the network precision is improved;
setting a rule degree constraint condition to be observed during coding, namely, when a convolutional neural network model is realized in an actual remote sensing platform processor, pruning convolutional kernels of different filters in the same layer at the same position simultaneously to improve the universality degree of a computing unit and improve the parallelism degree of computation;
the compressed model can be transplanted and applied to platform processors with limited resources such as airborne and satellite-borne platforms.
Detailed Description
The present invention will be described in detail with reference to examples.
In order to solve the problems that the precision is easy to lose and the like in the traditional convolutional neural network pruning and quantization compression method, the invention synchronously prunes and quantizes the model on the basis of the convolutional neural network compression algorithm before summarization, improves the compression ratio of the model parameters and ensures the precision of the compressed model.
The invention provides a convolutional neural network pruning and quantization synchronous compression method for remote sensing application, which mainly comprises the following steps:
s1, building a deep convolutional neural network model, and performing parameter training on the network by using natural scene image data to obtain a pre-training model M1;
s2, carrying out parameter training on the pre-training model M1 by using remote sensing image data to obtain a model M2;
the model is pre-trained by applying a large amount of natural image data to the scene with less remote sensing image data volume, so that the feature extraction capability of the model can be improved, and when the remote sensing image data volume is not less than the natural scene image data volume, the model pre-training in the step S1 is not required.
S3, pruning and coding the quantization parameters of the model M2, wherein the specific coding scheme is designed as follows:
the pruning of the model parameters contains two possibilities of whether to prune, while the quantization comprises a quantization bit number of 1 to n bits, which for a certain layer in the network can be defined as:
Ok=(p1,p2,...,pn,q1,q2,...,qn)
where n represents the number of parameters for each layer in the network. p is a radical of1To pnThe value of (1) is 0 or 1, 0 means that the parameter is pruned, and 1 means that the parameter is retained and pruning is not performed. q. q.s1To qnIs determined by the desired number of quantization bits, e.g. 4 to 8, representing the number of quantization bits of the parameter.
S4, randomly generating N (generally greater than 5) initial populations according to the encoding scheme of step S3 under the condition that the codes of the populations meet the rule constraint, each population containing K (generally greater than 20) individuals, wherein the random population generation scheme is designed according to the expected model compression method, and the typical 2 initial populations and individual generation methods are as follows:
firstly, randomly encoding O to each layer of the network under the condition of satisfying the set parameter compression ratiok=(p1,p2,...,pn,q1,q2,...,qn) Assigning the value of the middle element;
and secondly, obtaining a model structure after pruning or quantization by adopting the existing pruning and quantization algorithm, and assigning the codes of the initial population individuals according to the structure.
The rule degree constraint condition is required to be met during encoding. Pruning may be classified by granularity into filter level, convolution kernel level, kernel parameter level, etc. Considering that when a convolutional neural network model is implemented in an actual remote sensing platform processor, the general degree of a computing unit can be improved by pruning convolutional kernels of different filters in the same layer at the same position at the same time, and the parallelism of computation is improved, so that the rule degree constraint is set flexibly according to the hardware computation characteristics of the processor.E.g. when the constraint only prunes to the filter level, the parameters in the same filter need the same pruning coding. When the constraint only prunes at the convolutional kernel level, all parameters in the same convolutional kernel require the same pruning code. When the position of the pruning code is 0, the parameter quantization code corresponding to the position is also set to 0, for example, p1When q is 01=0。
S5, respectively carrying out parameter training on the models corresponding to the N populations and the K individual codes in the step S4 by adopting remote sensing image data (the data is the same as the data in the step S2);
during training, each individual retains the parameters in model M2, on the basis of which pruning and quantification are performed. The training cycles of each model are the same, and the training is carried out for a few times under the condition that the performance of the model can be distinguished, so that the training time is reduced.
S6, performing precision evaluation on the model obtained in S5, eliminating the individual with the worst precision part in each population, randomly changing the codes of the rest individuals, and performing code crossing on different individuals in the same population (exchanging codes O among different individuals)kThe element values at the same position) while migrating the optimal individual in each population to other populations, increasing the population by individuals.
The precision evaluation method is selected according to different task types, for example, for an image classification task, the classification accuracy of the model is used as a precision evaluation standard. And when the model precision is consistent, eliminating the individual with large model parameter quantity. When the encoding is changed randomly, the foregoing rule degree constraint in S4 needs to be satisfied.
S7, performing parameter training on the new population and the individuals obtained in the S6, and repeating the S6;
in the training process, the parameter value of each individual source is reserved, and training and parameter evaluation are carried out on the basis of the last training parameter. When the parameter at a certain position is pruned at the last time and no corresponding parameter value exists, the step is traced upwards until the value of the corresponding parameter in the model M2 is traced.
S8, stopping training when the precision, the parameter quantity and the calculated quantity of any individual reach a specified threshold or reach a preset training number;
and S9, performing precision evaluation on the model obtained in the S8, and reserving the T individuals with the highest precision to finish pruning and quantification of the convolutional neural network.
Whether parameters of the pruned and quantized model are pruned or not and the parameter bit number are determined, and the specific values of the parameters need to be trained and optimized continuously, which specifically comprises the following steps:
and when the training is carried out again, a layer-by-layer training method is adopted. Firstly, constructing a convolution neural network model of a floating point parameter according to an S9 pruning result, and training the network model by adopting training data, wherein the training of the pruning model of the floating point parameter can be carried out by adopting the steps S1 and S2 in sequence; then, on the basis of the parameters of the post-pruning model, the quantitative parameters of each layer of the model are trained in sequence. For example, according to the parameter quantization result of S9, the parameters of the first layer of the network model are quantized, the other layers of the network are floating point parameters, the entire convolutional neural network is trained on the basis of the pruned model parameters, after that, the quantized parameters of the first layer are fixed, the parameters are quantized according to the parameter quantization result of the second layer of S9, the parameters of the third layer and the following layers are kept as floating point parameters, then training is performed, and so on until the last layer of the network, and the training of the model parameters is completed.
In summary, the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (10)

1. A convolutional neural network pruning and quantization synchronous compression method for remote sensing application is characterized by comprising the following steps:
s1, performing parameter training on the convolutional neural network by adopting data containing the remote sensing image to obtain a model M2;
s2, pruning and coding the quantization parameters of the model M2, which are specifically as follows:
for the parameters of each layer in the convolutional neural network, its coding is defined as:
Ok=(p1,p2,...,pn,q1,q2,...,qn)
wherein n represents the number of parameters of each layer in the network; p is a radical of1,p2,...,pnThe value of 0 or 1 is used for representing whether each parameter is pruned, wherein 0 represents that the parameter is pruned, and 1 represents that the parameter is reserved and not pruned; q. q.s1,q2,...,qnRepresenting the number of quantization bits for each parameter;
s3, generating a set number of initial populations according to the codes of the step S2, wherein each population comprises a plurality of individuals;
s4, setting parameter pruning and quantification in the model M2 according to the codes of each individual in the population in the step S3, so as to obtain models corresponding to the number of the individuals, and training parameter values of the models by adopting remote sensing image data to obtain the models after parameter training;
s5, performing precision evaluation on the model obtained after parameter training, removing the part of individuals with the worst precision in each population, randomly changing the codes of the rest individuals, and performing cross and/or migration on the codes among the individuals to generate new individuals and populations;
s6, repeating the step S5 until the set conditions are met, and stopping training the model parameters;
in the training process, the parameter value of the model used in the current training is the parameter value of the last training;
and S7, performing precision evaluation on the model obtained in the S6, and reserving a part of individuals with the highest precision to finish pruning and quantification of the convolutional neural network.
2. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1, wherein the retraining is performed on the part of individuals with the highest S7 precision, specifically:
firstly, constructing a convolution neural network model of floating point parameters according to S7 pruning results, and training the network model parameters by adopting training data;
then, on the basis of the parameters of the post-pruning model, training the quantitative parameters of each layer of the model in turn, specifically: and (3) quantizing the parameters of the first layer of the network model according to the parameter quantization result of S7, training the whole convolutional neural network on the basis of the pruned model parameters, fixing the quantized parameters of the first layer after finishing the training of the other layers of the network model as floating point parameters, quantizing the parameters of the second layer according to the result of S7, keeping the floating point parameters of the third layer and the later layers, training the network, and so on until the last layer of the network, and finishing the final training of the model parameters.
3. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein in S2, the rule degree constraint condition is met during encoding, specifically: when the constraint only prunes the filter level of the convolutional neural network, the same pruning coding is needed for the parameters in the same filter; when the constraint only prunes at the convolution kernel level, all parameters in the same convolution kernel require the same pruning coding.
4. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein in S3, the method for generating initial population and individuals is as follows:
randomly encoding each layer of the network under the condition of satisfying parameter compression ratiok=(p1,p2,...,pn,q1,q2,...,qn) Assigning the value of the middle element;
or, the existing pruning and quantization algorithm is adopted to obtain a pruned or quantized model structure, and the codes of the individuals in the initial population are assigned according to the structure to obtain the individuals.
5. The convolutional neural network pruning and quantization synchronous compression method for remote sensing applications as claimed in claim 1 or 2, wherein the set conditions in S6 are: the training is stopped when the accuracy, the parameter amount and the calculated amount of any one individual reach a specified threshold or reach a predetermined number of times of training.
6. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein when training model parameters, the number of training cycles of each model is the same, and the training is performed for the minimum number of times under the condition that the performance of the model can be distinguished.
7. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein when model accuracy is evaluated in S5, individuals with large model parameter quantity are removed when model accuracy is consistent.
8. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein when evaluating model accuracy in S5, the evaluation method is selected according to different task types.
9. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein in S5, coding interleaving is performed on different individuals in the same population, the optimal individual in each population is migrated to other populations, and the individual and the population are updated.
10. The convolutional neural network pruning and quantization synchronous compression method for remote sensing application as claimed in claim 1 or 2, wherein in S1, when the remote sensing image data is insufficient, the network is firstly subjected to parameter training using natural scene image data to obtain a pre-training model M1; and then carrying out parameter training on the pre-training model M1 by using the remote sensing image data to obtain a model M2.
CN202110545477.8A 2021-05-19 2021-05-19 Convolutional neural network pruning and quantization synchronous compression method for remote sensing application Active CN113408723B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110545477.8A CN113408723B (en) 2021-05-19 2021-05-19 Convolutional neural network pruning and quantization synchronous compression method for remote sensing application

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110545477.8A CN113408723B (en) 2021-05-19 2021-05-19 Convolutional neural network pruning and quantization synchronous compression method for remote sensing application

Publications (2)

Publication Number Publication Date
CN113408723A true CN113408723A (en) 2021-09-17
CN113408723B CN113408723B (en) 2023-04-07

Family

ID=77678936

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110545477.8A Active CN113408723B (en) 2021-05-19 2021-05-19 Convolutional neural network pruning and quantization synchronous compression method for remote sensing application

Country Status (1)

Country Link
CN (1) CN113408723B (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102017124573A1 (en) * 2016-10-21 2018-04-26 Nvidia Corporation SYSTEMS AND METHOD FOR CRITING NEURONAL NETWORKS FOR AN OPERATIONAL EFFICIENT CONCLUSION
CN109635936A (en) * 2018-12-29 2019-04-16 杭州国芯科技股份有限公司 A kind of neural networks pruning quantization method based on retraining
CN110197257A (en) * 2019-05-28 2019-09-03 浙江大学 A kind of neural network structure Sparse methods based on increment regularization
CN110378468A (en) * 2019-07-08 2019-10-25 浙江大学 A kind of neural network accelerator quantified based on structuring beta pruning and low bit
CN111382863A (en) * 2018-12-28 2020-07-07 上海欧菲智能车联科技有限公司 Neural network compression method and device
CN111652366A (en) * 2020-05-09 2020-09-11 哈尔滨工业大学 Combined neural network model compression method based on channel pruning and quantitative training
CN111738401A (en) * 2019-03-25 2020-10-02 北京三星通信技术研究有限公司 Model optimization method, grouping compression method, corresponding device and equipment
CN112016674A (en) * 2020-07-29 2020-12-01 魔门塔(苏州)科技有限公司 Knowledge distillation-based convolutional neural network quantification method
CN112215353A (en) * 2020-09-29 2021-01-12 电子科技大学 Channel pruning method based on variational structure optimization network
CN112651499A (en) * 2020-12-28 2021-04-13 浙江大学 Structural model pruning method based on ant colony optimization algorithm and interlayer information

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102017124573A1 (en) * 2016-10-21 2018-04-26 Nvidia Corporation SYSTEMS AND METHOD FOR CRITING NEURONAL NETWORKS FOR AN OPERATIONAL EFFICIENT CONCLUSION
CN111382863A (en) * 2018-12-28 2020-07-07 上海欧菲智能车联科技有限公司 Neural network compression method and device
CN109635936A (en) * 2018-12-29 2019-04-16 杭州国芯科技股份有限公司 A kind of neural networks pruning quantization method based on retraining
CN111738401A (en) * 2019-03-25 2020-10-02 北京三星通信技术研究有限公司 Model optimization method, grouping compression method, corresponding device and equipment
CN110197257A (en) * 2019-05-28 2019-09-03 浙江大学 A kind of neural network structure Sparse methods based on increment regularization
CN110378468A (en) * 2019-07-08 2019-10-25 浙江大学 A kind of neural network accelerator quantified based on structuring beta pruning and low bit
CN111652366A (en) * 2020-05-09 2020-09-11 哈尔滨工业大学 Combined neural network model compression method based on channel pruning and quantitative training
CN112016674A (en) * 2020-07-29 2020-12-01 魔门塔(苏州)科技有限公司 Knowledge distillation-based convolutional neural network quantification method
CN112215353A (en) * 2020-09-29 2021-01-12 电子科技大学 Channel pruning method based on variational structure optimization network
CN112651499A (en) * 2020-12-28 2021-04-13 浙江大学 Structural model pruning method based on ant colony optimization algorithm and interlayer information

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
孙彦丽等: "基于剪枝与量化的卷积神经网络压缩方法", 《计算机科学》 *
郑哲: "稀疏与量化驱动的深度卷积网络压缩", 《中国优秀硕士学位论文全文数据库(电子期刊)信息科技辑》 *

Also Published As

Publication number Publication date
CN113408723B (en) 2023-04-07

Similar Documents

Publication Publication Date Title
RU2747445C1 (en) Data processing device, data processing method and data carrier
CN111275187A (en) Compression method and device of deep neural network model
CN113163203B (en) Deep learning feature compression and decompression method, system and terminal
WO2023207039A1 (en) Data processing method and apparatus, and device and storage medium
CN116524299A (en) Image sample generation method, device, equipment and storage medium
CN112488304A (en) Heuristic filter pruning method and system in convolutional neural network
CN115115815A (en) Training method, device and system for feature extraction network of three-dimensional grid model
WO2021038793A1 (en) Learning system, learning method, and program
CN113408723B (en) Convolutional neural network pruning and quantization synchronous compression method for remote sensing application
CN109063835B (en) Neural network compression device and method
CN112132062B (en) Remote sensing image classification method based on pruning compression neural network
CN113408704A (en) Data processing method, device, equipment and computer readable storage medium
CN110287706B (en) Security detection system and method for mimicry defense system
CN109558819B (en) Depth network lightweight method for remote sensing image target detection
Kwak et al. Quantization aware training with order strategy for CNN
CN115759238A (en) Method and device for generating quantization model, electronic equipment and storage medium
CN112200275B (en) Artificial neural network quantification method and device
CN116245162A (en) Neural network pruning method and system based on improved adaptive genetic algorithm
CN112712855A (en) Joint training-based clustering method for gene microarray containing deletion value
CN113033804A (en) Convolution neural network compression method for remote sensing image
CN113592085A (en) Nuclear pruning method, device, equipment and medium based on high-rank convolution graph
Lindmar et al. Intrinsic sparse LSTM using structured targeted dropout for efficient hardware inference
CN113569479A (en) Long-term multi-step control method, device and storage medium for rock fracture development of stone cave temple
CN111783936A (en) Convolutional neural network construction method, device, equipment and medium
CN111143641A (en) Deep learning model training method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant