CN107368885A - Network model compression method and device based on more granularity beta prunings - Google Patents

Network model compression method and device based on more granularity beta prunings Download PDF

Info

Publication number
CN107368885A
CN107368885A CN201710568710.8A CN201710568710A CN107368885A CN 107368885 A CN107368885 A CN 107368885A CN 201710568710 A CN201710568710 A CN 201710568710A CN 107368885 A CN107368885 A CN 107368885A
Authority
CN
China
Prior art keywords
beta pruning
pruning
granularity level
inessential
network model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710568710.8A
Other languages
Chinese (zh)
Inventor
曾建平
王军
李志国
班华忠
朱明�
张智鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zhi Xinyuandong Science And Technology Ltd
Original Assignee
Beijing Zhi Xinyuandong Science And Technology Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zhi Xinyuandong Science And Technology Ltd filed Critical Beijing Zhi Xinyuandong Science And Technology Ltd
Priority to CN201710568710.8A priority Critical patent/CN107368885A/en
Publication of CN107368885A publication Critical patent/CN107368885A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks

Abstract

The invention provides the network model compression method based on more granularity beta prunings, this method includes following one or two or three steps:Input channel granularity level beta pruning step, using inessential element beta pruning method, inessential element in the granularity level of the input channel of network model is subjected to beta pruning;Convolution kernel granularity level beta pruning step, using inessential element beta pruning method, beta pruning is carried out to inessential element in the granularity level of the convolution kernel of network model;Weight parameter granularity level beta pruning step, using inessential element beta pruning method, beta pruning is carried out to inessential element in the granularity level of the weight parameter of network model.Compared with prior art, the present invention can effectively solve network model compression problem by carrying out beta pruning to more granularity levels.

Description

Network model compression method and device based on more granularity beta prunings
Technical field
The present invention relates to image procossing, video monitoring and deep neural network, more particularly to based on more granularity beta prunings Network model compression method and device.
Background technology
In recent years, with the fast development of artificial intelligence, deep learning network by combining low-level image feature due to forming height Layer feature, is influenceed smaller by environmental change, breakthrough achievement is achieved in computer vision field, especially in recognition of face and Image classification etc. has surmounted the recognition accuracy of the mankind.
However, existing high-performance deep learning network typically has millions of or even more than one hundred million individual parameters, this causes its Storage and calculating consumption are all very huge, limit its equipment for being applied to store and computing resource is all limited.Therefore, to depth It is a committed step for solving this problem that learning network model compression, which is compressed,.
But existing model compression technology reduces moulded dimension typically by the rarefaction of Model Weight value, but Be can not significantly decrease operation deep learning network required for storage and computing resource.
In summary, it is necessary to propose a kind of deep learning network model compression side for reducing storage and computing resource consumption Method.
The content of the invention
In view of this, it is a primary object of the present invention to reduce storage resource and computing resource consumption, network model is realized Compression.
To reach above-mentioned purpose, according to the first aspect of the invention, there is provided the network model based on more granularity beta prunings Compression method, this method include following one or two or three steps:
Input channel granularity level beta pruning step, using inessential element beta pruning method, by the input channel of network model Inessential element carries out beta pruning in granularity level;
Convolution kernel granularity level beta pruning step, using inessential element beta pruning method, to the granularity of the convolution kernel of network model Inessential element carries out beta pruning in level;
Weight parameter granularity level beta pruning step, using inessential element beta pruning method, to the weight parameter of network model Inessential element carries out beta pruning in granularity level.
Further, the inessential element beta pruning method includes:
The zero setting step of inessential element, the importance of each element in current granularity level is calculated, by inessential element Corresponding value zero setting;
Beta pruning trim step, whole network model is finely tuned according to training data;
Judgment step is lost, calculates the loss of accuracy of the network model after fine setting, if the loss of accuracy is less than essence Exactness loses threshold value, then continues executing with the zero setting step of inessential element, otherwise terminate.
Further, the zero setting step of the inessential element includes:
Element importance sorting step, count the weight parameter vector W that beta pruning element is each treated in current granularity levele, The importance of beta pruning element is treated described in calculatingTo needed beta pruning element according to importance EIViRisen from small to large Sequence arranges, and obtains importance ascending order set, and calculate the importance summation of needed beta pruning element N is the quantity of current granularity level weight parameter;
Pruning threshold calculation procedure, according to energy beta pruning rate threshold value EPR, calculate beta pruning ENERGY E P=EIVS × EPR, statistics The cumulative distribution of importance in ascending order set, the importance corresponding to the cumulative distribution equal with beta pruning energy is chosen as beta pruning Threshold value;
Loss function influence amount calculation procedure, one group of test data is inputted, the loss function value Loss of calculating network, is calculated Loss function influence amount
Zero setting step, for each treating beta pruning element i in current granularity level, if treating the importance EIV of beta pruning elementi Less than pruning threshold, and loss function influence amount ELiLess than 0, then this is treated to be worth zero setting corresponding to beta pruning element i.
According to another aspect of the present invention, there is provided the network model compression set based on more granularity beta prunings, the device Including following one or two or three modules:
Input channel granularity level pruning module, for using inessential element pruning module, by the input of network model Inessential element carries out beta pruning in the granularity level of passage;
Convolution kernel granularity level pruning module, for using inessential element pruning module, to the convolution kernel of network model Granularity level in inessential element carry out beta pruning;
Weight parameter granularity level pruning module, for using inessential element pruning module, to the weight of network model Inessential element carries out beta pruning in the granularity level of parameter.
Further, the inessential element pruning module includes:
The zero setting module of inessential element, will be inessential for calculating the importance of each element in current granularity level Value zero setting corresponding to element;
Module is finely tuned in beta pruning, for finely tuning whole network model according to training data;
Judge module is lost, for calculating the loss of accuracy of the network model after finely tuning, if the loss of accuracy is small In loss of accuracy's threshold value, then the zero setting module of inessential element is continued executing with, is otherwise terminated.
Further, the zero setting module of the inessential element includes:
Element importance sorting module, the weight parameter vector W of beta pruning element is each treated in current granularity level for countinge, The importance of beta pruning element is treated described in calculatingTo needed beta pruning element according to importance EIViAscending order is carried out from small to large Arrangement, obtains importance ascending order set, and calculate the importance summation of needed beta pruning element N is the quantity of current granularity level weight parameter;
Pruning threshold computing module, for according to energy beta pruning rate threshold value EPR, calculating beta pruning ENERGY E P=EIVS × EPR, The cumulative distribution of importance in ascending order set is counted, chooses the importance conduct corresponding to the cumulative distribution equal with beta pruning energy Pruning threshold;
Loss function influence amount computing module, for input one group of test data, the loss function value Loss of calculating network, Counting loss function influences amount
Zero setting module, for for each treating beta pruning element i in current granularity level, if treating the importance of beta pruning element EIViLess than pruning threshold, and loss function influence amount ELiLess than 0, then this is treated to be worth zero setting corresponding to beta pruning element i.
Compared with existing network model compression method, the model compression method of the invention based on more granularity beta prunings, adopt , not only can be with compression network moulded dimension, and due to network model with the method for one or more kinds of granularity level beta prunings Sparse format is regular, it is possible to reduce the calculating consumption of network.
Brief description of the drawings
Fig. 1 shows the flow of the network model compression method based on more granularity beta prunings according to one embodiment of the invention Figure.
Fig. 2 shows the structure of the network model compression set based on more granularity beta prunings according to one embodiment of the invention Schematic diagram.
Embodiment
To enable your auditor to further appreciate that structure, feature and the other purposes of the present invention, in conjunction with appended preferably real Apply example describe in detail it is as follows, illustrated preferred embodiment is merely to illustrate technical scheme, and the non-limiting present invention.
According to the present invention the network model compression method based on more granularity beta prunings include it is following one or two or Three steps:
Input channel granularity level beta pruning step, using inessential element beta pruning method, by the input channel of network model Inessential element carries out beta pruning in granularity level;
Convolution kernel granularity level beta pruning step, using inessential element beta pruning method, to the granularity of the convolution kernel of network model Inessential element carries out beta pruning in level;
Weight parameter granularity level beta pruning step, using inessential element beta pruning method, to the weight parameter of network model Inessential element carries out beta pruning in granularity level.
Fig. 1 gives the flow of the network model compression method based on more granularity beta prunings according to one embodiment of the invention Figure.As shown in figure 1, include according to the network model compression method based on more granularity beta prunings of the present invention:
Input channel granularity level beta pruning step S1, using inessential element beta pruning method S10, the input of network model is led to Inessential element carries out beta pruning in the granularity level in road;
Convolution kernel granularity level beta pruning step S2, using inessential element beta pruning method S10, to the convolution kernel of network model Inessential element carries out beta pruning in granularity level;
Weight parameter granularity level beta pruning step S3, using inessential element beta pruning method S10, the weight of network model is joined Inessential element carries out beta pruning in several granularity levels.
Further, the inessential element beta pruning method S10 includes:
The zero setting step S11 of inessential element, the importance of each element in current granularity level is calculated, by inessential member Value zero setting corresponding to plain;
Beta pruning trim step S12, whole network model is finely tuned according to training data;
Judgment step S13 is lost, the loss of accuracy of the network model after fine setting is calculated, if the loss of accuracy is less than Loss of accuracy's threshold value, then the zero setting step S11 of inessential element is continued executing with, is otherwise terminated.
Further, the zero setting step S11 of the inessential element includes:
Element importance sorting step S111, count the weight parameter vector W that beta pruning element is each treated in current granularity levele, meter The importance of beta pruning element is treated described in calculationTo needed beta pruning element according to importance EIViAscending order row is carried out from small to large Row, obtain importance ascending order set, and calculate the importance summation of needed beta pruning element N is the quantity of current granularity level weight parameter;
Pruning threshold calculation procedure S112, according to energy beta pruning rate threshold value EPR, beta pruning ENERGY E P=EIVS × EPR is calculated, The cumulative distribution of importance in ascending order set is counted, chooses the importance conduct corresponding to the cumulative distribution equal with beta pruning energy Pruning threshold;
Loss function influence amount calculation procedure S113, one group of test data of input, the loss function value Loss of calculating network, Counting loss function influences amount
Zero setting step S114, for each treating beta pruning element i in current granularity level, if treating the importance of beta pruning element EIViLess than pruning threshold, and loss function influence amount ELiLess than 0, then this is treated to be worth zero setting corresponding to beta pruning element i.
Further, the span of the energy beta pruning rate threshold value EPR is 0.01~0.2.Preferably, the energy is cut Branch rate threshold value EPR span is 0.05~0.18.
Specifically, the current grain in the inessential element beta pruning method S10 of the input channel granularity level beta pruning step S1 Spend granularity level of the level for the input channel of network model.The inessential element of the convolution kernel granularity level beta pruning step S2 Current granularity level in beta pruning method S10 is the granularity level of the convolution kernel of network model.The weight parameter granularity level is cut Current granularity level in branch step S3 inessential element beta pruning method S10 is the granularity level of the weight parameter of network model.
Further, the beta pruning trim step S12 is finely adjusted using gradient descent method to whole network model.It is excellent Selection of land, the beta pruning trim step S12 are finely adjusted using stochastic gradient descent method to whole network model.
Further, the span of loss of accuracy's threshold value is 0.01~0.1 in the loss judgment step S13.It is excellent Selection of land, the span of loss of accuracy's threshold value is 0.05~0.08.
Further, the network is deep learning network.Preferably, the network include but is not limited to it is following a kind of or The a variety of combination of person:Convolutional neural networks, depth belief network, recurrent neural network.
According to the present invention the network model compression set based on more granularity beta prunings include it is following one or two or Three modules:
Input channel granularity level pruning module, for using inessential element pruning module, by the input of network model Inessential element carries out beta pruning in the granularity level of passage;
Convolution kernel granularity level pruning module, for using inessential element pruning module, to the convolution kernel of network model Granularity level in inessential element carry out beta pruning;
Weight parameter granularity level pruning module, for using inessential element pruning module, to the weight of network model Inessential element carries out beta pruning in the granularity level of parameter.
Fig. 2 gives the structure of the network model compression set based on more granularity beta prunings according to one embodiment of the invention Schematic diagram.As shown in Fig. 2 include according to the network model compression set based on more granularity beta prunings of the present invention:
Input channel granularity level pruning module 1, for using inessential element pruning module, by the input of network model Inessential element carries out beta pruning in the granularity level of passage;
Convolution kernel granularity level pruning module 2, for using inessential element pruning module, to the convolution kernel of network model Granularity level in inessential element carry out beta pruning;
Weight parameter granularity level pruning module 3, for using inessential element pruning module, to the weight of network model Inessential element carries out beta pruning in the granularity level of parameter.
Further, the inessential element pruning module 10 includes:
The zero setting module 11 of inessential element, for calculating the importance of each element in current granularity level, it will not weigh Want value zero setting corresponding to element;
Module 12 is finely tuned in beta pruning, for finely tuning whole network model according to training data;
Judge module 13 is lost, for calculating the loss of accuracy of the network model after finely tuning, if the loss of accuracy Less than loss of accuracy's threshold value, then the zero setting module 11 of inessential element is continued executing with, is otherwise terminated.
Further, the zero setting module 11 of the inessential element includes:
Element importance sorting module 111, the weight parameter vector W of beta pruning element is each treated in current granularity level for countinge, The importance of beta pruning element is treated described in calculatingTo needed beta pruning element according to importance EIViAscending order row is carried out from small to large Row, obtain importance ascending order set, and calculate the importance summation of needed beta pruning element N is the quantity of current granularity level weight parameter;
Pruning threshold computing module 112, for according to energy beta pruning rate threshold value EPR, calculate beta pruning ENERGY E P=EIVS × EPR, the cumulative distribution of importance in ascending order set is counted, choose the importance corresponding to the cumulative distribution equal with beta pruning energy As pruning threshold;
Loss function influence amount computing module 113, for inputting one group of test data, the loss function value of calculating network Loss, counting loss function influences amount
Zero setting module 114, for for each treating beta pruning element i in current granularity level, if treating the weight of beta pruning element Spend EIViLess than pruning threshold, and loss function influence amount ELiLess than 0, then this is treated to be worth zero setting corresponding to beta pruning element i.
Compared with existing network model compression method, the model compression method of the invention based on more granularity beta prunings, adopt , not only can be with compression network moulded dimension, and due to network model with the method for one or more kinds of granularity level beta prunings Sparse format is regular, it is possible to reduce the calculating consumption of network.
The foregoing is only a preferred embodiment of the present invention, is not intended to limit the scope of the present invention, should Understand, the present invention is not limited to implementation as described herein, and the purpose of these implementations description is to help this area In technical staff put into practice the present invention.Any those of skill in the art are easy to do not departing from spirit and scope of the invention In the case of be further improved and perfect, therefore the present invention is only by the content of the claims in the present invention and limiting for scope System, its intention cover all alternatives being included in the spirit and scope of the invention being defined by the appended claims and waited Same scheme.

Claims (10)

1. the network model compression method based on more granularity beta prunings, it is characterised in that this method include it is following one or two Or three steps:
Input channel granularity level beta pruning step, using inessential element beta pruning method, by the granularity of the input channel of network model Inessential element carries out beta pruning in level;
Convolution kernel granularity level beta pruning step, using inessential element beta pruning method, to the granularity level of the convolution kernel of network model In inessential element carry out beta pruning;
Weight parameter granularity level beta pruning step, using inessential element beta pruning method, to the granularity of the weight parameter of network model Inessential element carries out beta pruning in level.
2. the method as described in claim 1, it is characterised in that the inessential element beta pruning method includes:
The zero setting step of inessential element, the importance of each element in current granularity level is calculated, inessential element is corresponding Value zero setting;
Beta pruning trim step, whole network model is finely tuned according to training data;
Judgment step is lost, the loss of accuracy of the network model after fine setting is calculated, if the loss of accuracy is less than accuracy Threshold value is lost, then continues executing with the zero setting step of inessential element, otherwise terminates.
3. method as claimed in claim 2, it is characterised in that the zero setting step of the inessential element includes:Element is important Sequence step is spent, counts the weight parameter vector W that beta pruning element is each treated in current granularity levele, beta pruning member is treated described in calculating The importance of elementTo needed beta pruning element according to importance EIViAscending order arrangement is carried out from small to large, obtains weight Ascending order set is spent, and calculates the importance summation of needed beta pruning elementN For the quantity of current granularity level weight parameter;
Pruning threshold calculation procedure, according to energy beta pruning rate threshold value EPR, beta pruning ENERGY E P=EIVS × EPR is calculated, counts ascending order The cumulative distribution of importance in set, the importance corresponding to the cumulative distribution equal with beta pruning energy is chosen as beta pruning threshold Value;
Loss function influence amount calculation procedure, input one group of test data, the loss function value Loss of calculating network, counting loss Function influences amount
Zero setting step, for each treating beta pruning element i in current granularity level, if treating the importance EIV of beta pruning elementiIt is less than Pruning threshold, and loss function influence amount ELiLess than 0, then this is treated to be worth zero setting corresponding to beta pruning element i.
4. method as claimed in claim 2, the inessential element beta pruning method of the input channel granularity level beta pruning step Current granularity level is the granularity level of the input channel of network model;The convolution kernel granularity level beta pruning step it is inessential Current granularity level in element beta pruning method is the granularity level of the convolution kernel of network model;The weight parameter granularity level is cut Current granularity level in the inessential element beta pruning method of branch step is the granularity level of the weight parameter of network model.
5. method as claimed in claim 3, the span of the energy beta pruning rate threshold value EPR is 0.01~0.2.
6. method as claimed in claim 2, the span of loss of accuracy's threshold value is 0.01~0.1.
7. method as claimed in claim 4, the network is deep learning network;Preferably, the network includes but unlimited In following one or more kinds of combination:Convolutional neural networks, depth belief network, recurrent neural network.
8. the network model compression set based on more granularity beta prunings, it is characterised in that the device include it is following one or two Or three modules:
Input channel granularity level pruning module, for using inessential element pruning module, by the input channel of network model Granularity level in inessential element carry out beta pruning;
Convolution kernel granularity level pruning module, for using inessential element pruning module, to the grain of the convolution kernel of network model Spend inessential element in level and carry out beta pruning;
Weight parameter granularity level pruning module, for using inessential element pruning module, to the weight parameter of network model Granularity level in inessential element carry out beta pruning.
9. device as claimed in claim 8, it is characterised in that the inessential element pruning module includes:
The zero setting module of inessential element, for calculating the importance of each element in current granularity level, by inessential element Corresponding value zero setting;
Module is finely tuned in beta pruning, for finely tuning whole network model according to training data;
Judge module is lost, for calculating the loss of accuracy of the network model after finely tuning, if the loss of accuracy is less than essence Exactness loses threshold value, then continues executing with the zero setting module of inessential element, otherwise terminate.
10. device as claimed in claim 9, it is characterised in that the zero setting module of the inessential element includes:Element is important Order module is spent, each treats the weight parameter vector W of beta pruning element in current granularity level for countinge, wait to cut described in calculating The importance of branch elementTo needed beta pruning element according to importance EIViAscending order arrangement is carried out from small to large, is obtained To importance ascending order set, and calculate the importance summation of needed beta pruning element N is the quantity of current granularity level weight parameter;Pruning threshold computing module, for according to energy beta pruning rate threshold value EPR, calculating Beta pruning ENERGY E P=EIVS × EPR, the cumulative distribution of importance in ascending order set is counted, choose equal with beta pruning energy add up The corresponding importance of distribution is as pruning threshold;
Loss function influence amount computing module, for inputting one group of test data, the loss function value Loss of calculating network, calculate Loss function influence amount
Zero setting module, for for each treating beta pruning element i in current granularity level, if treating the importance EIV of beta pruning elementi Less than pruning threshold, and loss function influence amount ELiLess than 0, then this is treated to be worth zero setting corresponding to beta pruning element i.
CN201710568710.8A 2017-07-13 2017-07-13 Network model compression method and device based on more granularity beta prunings Pending CN107368885A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710568710.8A CN107368885A (en) 2017-07-13 2017-07-13 Network model compression method and device based on more granularity beta prunings

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710568710.8A CN107368885A (en) 2017-07-13 2017-07-13 Network model compression method and device based on more granularity beta prunings

Publications (1)

Publication Number Publication Date
CN107368885A true CN107368885A (en) 2017-11-21

Family

ID=60308043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710568710.8A Pending CN107368885A (en) 2017-07-13 2017-07-13 Network model compression method and device based on more granularity beta prunings

Country Status (1)

Country Link
CN (1) CN107368885A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108629288A (en) * 2018-04-09 2018-10-09 华中科技大学 A kind of gesture identification model training method, gesture identification method and system
CN108898168A (en) * 2018-06-19 2018-11-27 清华大学 The compression method and system of convolutional neural networks model for target detection
CN109344921A (en) * 2019-01-03 2019-02-15 湖南极点智能科技有限公司 A kind of image-recognizing method based on deep neural network model, device and equipment
CN109523017A (en) * 2018-11-27 2019-03-26 广州市百果园信息技术有限公司 Compression method, device, equipment and the storage medium of deep neural network
CN109583586A (en) * 2018-12-05 2019-04-05 东软睿驰汽车技术(沈阳)有限公司 A kind of convolution kernel processing method and processing device
CN109634401A (en) * 2018-12-29 2019-04-16 联想(北京)有限公司 A kind of control method and electronic equipment
WO2019127362A1 (en) * 2017-12-29 2019-07-04 清华大学 Neural network model block compression method, training method, computing device and system
CN110059823A (en) * 2019-04-28 2019-07-26 中国科学技术大学 Deep neural network model compression method and device
WO2019226686A3 (en) * 2018-05-23 2020-02-06 Movidius Ltd. Deep learning system
CN111695375A (en) * 2019-03-13 2020-09-22 上海云从企业发展有限公司 Face recognition model compression algorithm based on model distillation, medium and terminal
CN112036563A (en) * 2019-06-03 2020-12-04 国际商业机器公司 Deep learning model insights using provenance data
CN112132062A (en) * 2020-09-25 2020-12-25 中南大学 Remote sensing image classification method based on pruning compression neural network
CN112329931A (en) * 2021-01-04 2021-02-05 北京智源人工智能研究院 Countermeasure sample generation method and device based on proxy model
KR20210015990A (en) * 2019-05-18 2021-02-10 주식회사 디퍼아이 Convolution neural network parameter optimization method, neural network computing method and apparatus
CN113408561A (en) * 2020-03-17 2021-09-17 北京京东乾石科技有限公司 Model generation method, target detection method, device, equipment and storage medium
TWI768167B (en) * 2017-12-30 2022-06-21 大陸商中科寒武紀科技股份有限公司 Integrated circuit chip device and related products
US11651202B2 (en) 2017-12-30 2023-05-16 Cambricon Technologies Corporation Limited Integrated circuit chip device and related product
US11704544B2 (en) 2017-12-30 2023-07-18 Cambricon Technologies Corporation Limited Integrated circuit chip device and related product
US11710031B2 (en) 2017-12-30 2023-07-25 Cambricon Technologies Corporation Limited Parallel processing circuits for neural networks
US11734548B2 (en) 2017-12-30 2023-08-22 Cambricon Technologies Corporation Limited Integrated circuit chip device and related product

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019127362A1 (en) * 2017-12-29 2019-07-04 清华大学 Neural network model block compression method, training method, computing device and system
US11704544B2 (en) 2017-12-30 2023-07-18 Cambricon Technologies Corporation Limited Integrated circuit chip device and related product
US11710031B2 (en) 2017-12-30 2023-07-25 Cambricon Technologies Corporation Limited Parallel processing circuits for neural networks
US11734548B2 (en) 2017-12-30 2023-08-22 Cambricon Technologies Corporation Limited Integrated circuit chip device and related product
TWI768167B (en) * 2017-12-30 2022-06-21 大陸商中科寒武紀科技股份有限公司 Integrated circuit chip device and related products
US11651202B2 (en) 2017-12-30 2023-05-16 Cambricon Technologies Corporation Limited Integrated circuit chip device and related product
CN108629288A (en) * 2018-04-09 2018-10-09 华中科技大学 A kind of gesture identification model training method, gesture identification method and system
CN108629288B (en) * 2018-04-09 2020-05-19 华中科技大学 Gesture recognition model training method, gesture recognition method and system
WO2019226686A3 (en) * 2018-05-23 2020-02-06 Movidius Ltd. Deep learning system
US11900256B2 (en) 2018-05-23 2024-02-13 Intel Corporation Deep learning system
CN108898168A (en) * 2018-06-19 2018-11-27 清华大学 The compression method and system of convolutional neural networks model for target detection
CN108898168B (en) * 2018-06-19 2021-06-01 清华大学 Compression method and system of convolutional neural network model for target detection
CN109523017B (en) * 2018-11-27 2023-10-17 广州市百果园信息技术有限公司 Gesture detection method, device, equipment and storage medium
CN109523017A (en) * 2018-11-27 2019-03-26 广州市百果园信息技术有限公司 Compression method, device, equipment and the storage medium of deep neural network
CN109583586A (en) * 2018-12-05 2019-04-05 东软睿驰汽车技术(沈阳)有限公司 A kind of convolution kernel processing method and processing device
CN109583586B (en) * 2018-12-05 2021-03-23 东软睿驰汽车技术(沈阳)有限公司 Convolution kernel processing method and device in voice recognition or image recognition
CN109634401A (en) * 2018-12-29 2019-04-16 联想(北京)有限公司 A kind of control method and electronic equipment
CN109634401B (en) * 2018-12-29 2023-05-02 联想(北京)有限公司 Control method and electronic equipment
CN109344921A (en) * 2019-01-03 2019-02-15 湖南极点智能科技有限公司 A kind of image-recognizing method based on deep neural network model, device and equipment
CN111695375A (en) * 2019-03-13 2020-09-22 上海云从企业发展有限公司 Face recognition model compression algorithm based on model distillation, medium and terminal
CN110059823A (en) * 2019-04-28 2019-07-26 中国科学技术大学 Deep neural network model compression method and device
KR102247896B1 (en) 2019-05-18 2021-05-04 주식회사 디퍼아이 Convolution neural network parameter optimization method, neural network computing method and apparatus
KR20210015990A (en) * 2019-05-18 2021-02-10 주식회사 디퍼아이 Convolution neural network parameter optimization method, neural network computing method and apparatus
CN112036563A (en) * 2019-06-03 2020-12-04 国际商业机器公司 Deep learning model insights using provenance data
WO2021185121A1 (en) * 2020-03-17 2021-09-23 北京京东乾石科技有限公司 Model generation method and apparatus, object detection method and apparatus, device, and storage medium
CN113408561A (en) * 2020-03-17 2021-09-17 北京京东乾石科技有限公司 Model generation method, target detection method, device, equipment and storage medium
CN112132062A (en) * 2020-09-25 2020-12-25 中南大学 Remote sensing image classification method based on pruning compression neural network
CN112329931B (en) * 2021-01-04 2021-05-07 北京智源人工智能研究院 Countermeasure sample generation method and device based on proxy model
CN112329931A (en) * 2021-01-04 2021-02-05 北京智源人工智能研究院 Countermeasure sample generation method and device based on proxy model

Similar Documents

Publication Publication Date Title
CN107368885A (en) Network model compression method and device based on more granularity beta prunings
CN105512289B (en) Image search method based on deep learning and Hash
CN103810101B (en) A kind of Software Defects Predict Methods and software defect forecasting system
CN109146076A (en) model generating method and device, data processing method and device
CN107273836A (en) A kind of pedestrian detection recognition methods, device, model and medium
CN109754068A (en) Transfer learning method and terminal device based on deep learning pre-training model
CN110263863A (en) Fine granularity mushroom phenotype recognition methods based on transfer learning Yu bilinearity InceptionResNetV2
CN108614997B (en) Remote sensing image identification method based on improved AlexNet
CN108388957B (en) Medium and small river flood forecasting method and forecasting system based on multi-feature fusion technology
CN110322453A (en) 3D point cloud semantic segmentation method based on position attention and auxiliary network
WO2020238237A1 (en) Power exponent quantization-based neural network compression method
CN105354595A (en) Robust visual image classification method and system
CN104616029B (en) Data classification method and device
CN103313047B (en) A kind of method for video coding and device
CN109993100A (en) The implementation method of facial expression recognition based on further feature cluster
CN106780737A (en) A kind of method that utilization digital elevation model calculates Geomorphologic Instantaneous Unit Hydrograph probability
CN111435463A (en) Data processing method and related equipment and system
CN110060489A (en) A kind of traffic signal timing scheme recommended method neural network based
CN109284761A (en) A kind of image characteristic extracting method, device, equipment and readable storage medium storing program for executing
CN104318241A (en) Local density spectral clustering similarity measurement algorithm based on Self-tuning
CN105631469A (en) Bird image recognition method by multilayer sparse coding features
CN109472352A (en) A kind of deep neural network model method of cutting out based on characteristic pattern statistical nature
CN110070106A (en) Smog detection method, device and electronic equipment
Cheni et al. Reducing latency in a converted spiking video segmentation network
CN103281555B (en) Half reference assessment-based quality of experience (QoE) objective assessment method for video streaming service

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20171121

RJ01 Rejection of invention patent application after publication