CN109146000A - A kind of method and device for improving convolutional neural networks based on frost weight - Google Patents

A kind of method and device for improving convolutional neural networks based on frost weight Download PDF

Info

Publication number
CN109146000A
CN109146000A CN201811044605.5A CN201811044605A CN109146000A CN 109146000 A CN109146000 A CN 109146000A CN 201811044605 A CN201811044605 A CN 201811044605A CN 109146000 A CN109146000 A CN 109146000A
Authority
CN
China
Prior art keywords
hidden layer
layer node
weight
convolutional neural
neural networks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811044605.5A
Other languages
Chinese (zh)
Other versions
CN109146000B (en
Inventor
韩宇铭
朱立东
冉普航
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201811044605.5A priority Critical patent/CN109146000B/en
Publication of CN109146000A publication Critical patent/CN109146000A/en
Application granted granted Critical
Publication of CN109146000B publication Critical patent/CN109146000B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/32Digital ink

Abstract

The invention discloses a kind of method and device for improving convolutional neural networks based on frost weight, this method improves traditional BP convolutional neural networks by the theoretical of frost weight.This method acquires enough training and test sample data first, it is pre-processed, and establish convolutional neural networks model, processing is optimized to convolutional neural networks according to frost weight theory, the hidden layer node in convolutional neural networks is analyzed by being introduced into entropy weight method, freeze to contribute little hidden layer node to network output in convolutional neural networks training process, the renewal process of hidden layer node in reasonably optimizing convolutional neural networks, it is effectively reduced the computational complexity of convolutional neural networks, shortens the training duration of convolutional neural networks.This method has the advantages that efficient, reliable, precision is higher.

Description

A kind of method and device for improving convolutional neural networks based on frost weight
Technical field
The invention belongs to nerual network technique algorithm fields, and in particular to one kind improves convolutional Neural net based on frost weight The method and device of network.
Background technique
Convolutional neural networks (Convolutional Neural Network, CNN) are to be applied to deep learning in recent years A kind of neural network be also widely used for image in recent years by convolutional neural networks to the high-performance of image procossing The fields such as identification.The concept of convolutional neural networks is proposed by LeCun earliest, and LeCun et al. is proposed in Fukushima Neocognitron (Neocognitron) on the basis of, using simple cell layer (S-layer) and complex cell layer (C- Layer) the structure being alternately superimposed, and gone out warp using BP (Back Propagation back-propagation algorithm) algorithm successful design The character identification system LeNet-5 model of allusion quotation.As the convolutional neural networks of first appearance, which has stronger Shandong Stick, convolutional neural networks are reached by amplifying the space characteristics between data using the spatial coherence of data to each other To the purpose for shortening the training time.
But convolutional neural networks equally also have many problems, and main problem therein is made from convolutional neural networks Conventional counter propagation algorithm (Back Propagation, BP).Since the error function in traditional BP neural network is phase For the first derivative of hidden layer node weight each in network, it can not accomplish that carrying out unlimited number of minimum asks in reality Topic just will appear the problem of step-length is chosen.Secondly, traditional BP neural network objective function to be optimized be it is extremely complex, Therefore it necessarily will appear " zigzag phenomenon ", this makes traditional BP neural network computing inefficient;Again due to traditional BP neural network The objective function of middle optimization is very complicated, it some flat regions will necessarily occur in the case where neuron is exported close to 0 or 1, In these regions, node weight changes very little, and weight error changes very little, network training process is made almost to pause.These problems The convolutional neural networks training time will be caused longer, for complicated nerve network system of construction itself, this, which is undoubtedly, is avenged Add frost.
In a kind of " the parametric source modeling method based on improved BP " of Publication No. CN103077267B In state's patent, a kind of improvement BP optimized based on genetic algorithm to structure and parameter in convolutional neural networks model is disclosed Neural network, by find preferably convolutional neural networks imply initial weight between the number of plies, hidden layer node and threshold value come Optimize convolutional neural networks algorithm.This method optimizes convolutional neural networks to a certain extent, but this method introduces heredity Algorithm increases the training duration of convolutional neural networks actually, and there is no the training durations for really shortening convolutional neural networks.
Summary of the invention
It is an object of the invention to overcome the above-mentioned deficiency in the presence of the prior art, a kind of introducing frost weight reason is provided By the method for improving traditional BP convolutional neural networks with entropy weight data processing method, while this method can effectively promote network fortune Defeated efficiency reduces the training time, guarantees that convolutional neural networks have stronger robustness.
In order to achieve the above-mentioned object of the invention, the present invention provides following technical schemes:
A method of convolutional neural networks are improved based on frost weight, comprising the following steps:
Step 101, pretreatment image obtains multiple batch training image data;
Step 102, convolutional neural networks are constructed, and the hidden layer node weight amount in convolutional neural networks are carried out initial Change assignment;
Step 103, the training image data for inputting a batch carry out convolutional neural networks by forward conduction algorithm It calculates, obtains the activation weight of each hidden layer node in convolutional neural networks, each implied according to the activation weight computing The weight difference of node layer;
Step 104, the weight difference is analyzed based on entropy weight method, to obtain the entropy weight of each hidden layer node, and The evaluation of estimate that hidden layer node is determined with this screens the hidden layer node based on the evaluation of estimate of hidden layer node, freezes Evaluation of estimate is tied in evaluation of estimate threshold value hidden layer node below to maintain its weight amount as current weight amount, and transmits and is not frozen Hidden layer node;
Step 105, by reverse conduction algorithm update transmitting come be not frozen hidden layer node weight amount, obtain It is not frozen the updated weight amount of hidden layer node, to get the weight amount of current each hidden layer node;
Step 106, error judgment is carried out according to current each hidden layer node weight amount, if error is not default first In range, then repeatedly step 103~105, until error stops the training of this batch in the first preset range;
Step 107, it uses multiple batch training image data instead to be trained convolutional neural networks, so that the error of network Index is in the second preset range.
Further, the method for improving convolutional neural networks, the pretreatment image are normalized from hand-written The a certain number of image datas got in volume data identification library, by all image data scaleds to unified scale.
Further, in the method for improving convolutional neural networks, the convolutional neural networks include an input layer, At least two convolutional layers and at least two sample levels, an output layer, wherein output layer merges full articulamentum.
Further, in the method for improving convolutional neural networks, the weight difference is gained hidden layer node Activate the absolute value of the difference of weight and hidden layer node current weight amount.
Further, in the method for improving convolutional neural networks, institute is analyzed based on entropy weight method described in step 104 Weight difference is stated, to obtain the entropy weight of each hidden layer node, and with this step of this evaluation of estimate for determining hidden layer node It specifically includes:
According to hidden layer node number and its weight difference, data matrix is established;The normalized data matrix, obtains Normalization matrix;According to normalization matrix, the entropy of hidden layer node weight difference is calculated;According to the entropy meter of gained hidden layer node Calculate coefficient of variation;The entropy weight of hidden layer node is calculated according to gained coefficient of variation;Hidden layer is calculated according to the entropy weight acquired The evaluation of estimate of node.
Further, the data matrix are as follows:
Wherein, m is hidden layer node number, and n is hidden layer node weight difference.
Further, in the method for improving convolutional neural networks, the weight amount of current each hidden layer node Initial weight amount and the not frozen updated weight amount of hidden layer node including frozen hidden layer node.
Further, the error judgment formula are as follows:
Wherein, E (w, v) is the current error amount of convolutional neural networks, and wherein k is the quantity of mode, and C is hidden layer node Quantity, tpiIt is the output target value of hidden layer node p, spiIt is reality output of the hidden layer node p in convolutional neural networks Value.
Further, the method for improving convolutional neural networks, step 107 includes, by the training image of multiple batches Data are separately input in convolutional neural networks, are repeated step 103- step 106, are executed multiple batches of training image data to network Training.
Preferably, a kind of device improving convolutional neural networks based on frost weight, including at least one processor, and The memory being connect at least one described processor communication;The memory, which is stored with, to be held by least one described processor Capable instruction, described instruction are executed by least one described processor, so that at least one described processor is able to carry out right It is required that method described in any one of 1 to 6.
Compared with prior art, beneficial effects of the present invention:
1, the structure for optimizing convolutional neural networks incorporates full articulamentum and output layer in convolutional neural networks, mentions High convolutional neural networks operation efficiency.
2, it is theoretical to introduce frost weight, improves original BP convolutional neural networks model.In BP convolutional neural networks In every batch of training process, the hidden layer node of frost and output result degree of correlation difference, in subsequent reverse conduction no longer The weight for the hidden layer node being frozen is updated, so as to improve the complexity of convolutional neural networks training process, promotes network instruction Practice efficiency, simplifies the calculating process that each reverse conduction updates hidden layer node weight, network training duration is reduced with this.
3, the destination node that entropy weight data processing method filters out required frost in every batch of training is introduced, is utilized The entropy weight between node weight value difference value in every batch of training is in the frost theory of original Relative Fuzzy as assessment reference The selection of node proposes better solution.
4, the experimental results showed that, training process in network not only can be effectively reduced in the improved convolutional neural networks of the present invention Complexity, shorten network training duration, moreover it is possible to avoid falling into step-length problem and training time too long problem in network training.
Detailed description of the invention:
Fig. 1 is the flow chart of convolutional neural networks improved method according to an exemplary embodiment of the present invention;
Fig. 2 is LeNet-5 convolutional neural networks structure chart;
Fig. 3 is the convolutional neural networks structure chart after the optimization according to an exemplary embodiment of the present invention based on LeNet-5;
Fig. 4 is the flow chart of entropy weight method according to an exemplary embodiment of the present invention;
Fig. 5 is convolutional neural networks improved method device according to an exemplary embodiment of the present invention.
Specific embodiment
Below with reference to test example and specific embodiment, the present invention is described in further detail.But this should not be understood It is all that this is belonged to based on the technology that the content of present invention is realized for the scope of the above subject matter of the present invention is limited to the following embodiments The range of invention.
Embodiment 1
A method of convolutional neural networks are improved based on frost weight, comprising the following steps:
Step 101, pretreatment image obtains multiple batch training image data;
70000 image datas are obtained in library specifically, identifying first from the hand-written volume data of MNIST, wherein training image Data 60000 are opened, and test image data 10000 are opened.To all image datas (including training image data and test image number According to) processing that is normalized, it is the input matrix of 28 × 28 pixels by all image real time transfers, every batch of training is schemed As data sample size is set as 50, to obtain the training image data of 1200 trained batches.
Step 102, convolutional neural networks are constructed, and the hidden layer node weight amount in convolutional neural networks are carried out initial Change assignment;
Specifically, we are based on character identification system LeNet-5 building convolutional neural networks in this example, and optimize The network structure of LeNet-5 merges the full articulamentum and output layer of the network, to reduce convolutional neural networks computational complexity. As shown in Fig. 2, the LeNet-5 network architecture on basis is by 1 input layer, 3 convolutional layers, 2 pond layers, 1 full articulamentum and 1 A output layer is constituted, and the convolutional neural networks after optimizing in the present invention are as shown in figure 3, comprising 1 input layer, 2 convolutional layers, and 2 A sample level (also referred to as pond layer), 1 output layer, sequence are as follows: the first convolutional layer, first are sequentially connected after input layer Pond layer, the second convolutional layer, the second pond layer, output layer (output layer incorporates full articulamentum).
Further, the input layer of the convolutional neural networks constructed in this example is according to test diagram data picture and training image Designed by data, 50, while one in each hidden layer node processing image are set by hidden layer node number in input layer A data point, thus the structure of input layer is 28 × 28 × 50.There are 65 × 5 volumes in subsequent first convolutional layer of input layer Product core, the data matrix that the image data transmitted from input layer is 24 × 24 by the processing of the first convolutional layer, and at each A biasing is added after convolution kernel.There are 61 × 12 convolution kernels in the second convolutional layer, is 12 8 × 8 by image procossing Image, it is same that 12 biasings are added.
The design of first sample level and the second sample level will match the framework of previous convolutional layer, first sample level In, the size that setting sampling window is 6 12 × 12 is biased to 0.Second sample level setting window is 12 4 × 4 big It is small, equally it is biased to 0.Each sample level both for previous convolutional layer data structure, by the result of convolution at each Dimension all reduces one times.A last output layer is similarly full articulamentum.
In this example, 150 hidden layer nodes are arranged in the first convolutional layer of the convolutional neural networks in we, second 1800 hidden layer nodes are arranged in convolutional layer.Then we carry out the weight amount of hidden layer nodes all in convolutional neural networks Assignment is initialized, all hidden layer nodes are input layer in convolutional neural networks, convolutional layer, pond layer, output layer (containing complete Articulamentum) in all hidden layer node.Here assignment is to determine in range (generally 0-100) to all hidden layer sections Point carries out random assignment.In addition, the setting of hidden layer node all in input layer, output layer all after initializing assignment not Become, and the hidden layer node in the layer of pond is not the target of this algorithm training, therefore, in this example, only in convolutional layer 1950 hidden layer nodes can be by innovatory algorithm of the invention training.
Step 103, the training image data for inputting a batch carry out convolutional neural networks by forward conduction algorithm It calculates, obtains the activation weight of the hidden layer node, believed according to the weight difference of the activation weight computing hidden layer node Breath;
Specifically, by training image data (50) the input convolutional neural networks of a batch, according to BP volumes Forward conduction training method in product neural network calculates the convolutional neural networks, available in this calculating The activation weight of each hidden layer node (activates the hidden layer node, the hidden layer node is made to obtain optimum Working institute The weight amount needed).Then with the initial power of the activation weight and each hidden layer node of obtained each hidden layer node Value amount does subtraction, and obtained absolute value of the difference is the weight difference information of the hidden layer node (namely in usual definition Residual error)
The principle of traditional BP convolutional neural networks is: a convolutional neural networks is given, by a batch training image After data input convolutional neural networks, what is carried out first is that " forward conduction " calculates, and is calculated according to forward conduction calculation method Each hidden layer node is from the second layer (first layer is input layer) in convolutional neural networks to output layer in convolutional neural networks Activation value calculate its residual error δ then for each hidden layer node ii (l), which shows each hidden layer section How many contribution produced to network final output for point.The residual information of hidden layer node carries out back transfer from output layer at this time, Then the weight amount of all hidden layer nodes is updated by " reverse conduction more new algorithm ".In the training of different batches, own The weight amount of hidden layer node can be continuously updated, and the output result of network can more level off to required for us as a result, network Stability and accuracy can also increase accordingly therewith, but at the same time, the related complexity and training duration of algorithm can also be shown The promotion of work.
The experimental results show during being trained to traditional BP convolutional neural networks, convolutional neural networks All hidden layer nodes in a part of hidden layer node it is in fact in fact and insensitive to a batch training image data, i.e., this The residual values that part hidden layer node calculates in the forward conduction training of network are smaller, it is to network output accordingly It influences also just smaller.Also mean that the increase with the training time, the variation of this part hidden layer node weight is little.Therefore such as Fruit directly carries out the right value update of hidden layer node, this part after completing forward conduction algorithm and calculating using reverse conduction algorithm What the lesser hidden layer node of residual values continued to be updated has little significance.Continue to be updated actually this part hidden layer node It is to flog a dead horse, increases with when training with apparent so as to cause the related complexity of algorithm.
The present invention is the improvement made on this basis to traditional BP convolutional neural networks, passes through forward conduction in network After training calculates the residual error of each hidden layer node, further, according to the theoretical thought of shannon entropy, we can lead to The entropy for the residual error for calculating each hidden layer node is crossed to evaluate it, it is subsequent reversed to decide whether to carry out the node Conduction updates weight process, and in evaluation criteria system index weights, entropy is a highly desirable scale.It is based on entropy weight method It determines an evaluation criterion, meaningless hidden layer node is continued to update in this part with this evaluation criterion and screens and carries out Freeze (i.e. this part of nodes no longer carry out weight amount update), so, not frozen node need to only be carried out subsequent Reverse conduction calculate update its weight, greatly reduce the computational complexity of network, save a large amount of net training time.
Step 104, the weight difference is analyzed based on entropy weight method, to obtain the entropy weight of each hidden layer node, and The evaluation of estimate that hidden layer node is determined with this screens the hidden layer node based on the evaluation of estimate of hidden layer node, freezes Evaluation of estimate is tied in evaluation of estimate threshold value hidden layer node below to maintain its weight amount as current weight amount, and transmits and is not frozen Hidden layer node;
Specifically, analyze the weight difference based on entropy weight method, to obtain the entropy weight of each hidden layer node, and with This determines that this step of the evaluation of estimate of hidden layer node specifically includes:
An indicator evaluation system is initially set up, (n weight of hidden layer node is poor if wherein there is n evaluation index Value), m are evaluated object (hidden layer node number), are evaluated the raw data matrix of the corresponding index of object, as follows:
Then we need that raw data matrix is normalized, and are denoted as matrix S=(sij)m*n
S is normalized, is denoted as
All S values obtained in this way will be in [0,1] section.
The entropy of j-th of weight difference is at this time
Wherein,
The coefficient of variation of j-th of weight difference are as follows:
αj=1-Hj(j=1,2, n)
The entropy weight of j-th of weight difference are as follows:
Under the setting of such entropy weight, we are it can be concluded that a judgment criteria, when obtained hidden layer node Entropy it is bigger when, then its entropy weight will be smaller, also can accordingly reduce to the contribution of the training process of this batch.When gained entropy Maximum value be 1, when entropy weight is 0, it is believed that the hidden layer node does not provide for this training any effective Contribution.
As shown in figure 4, specific entropy weight analytic approach can be summarized as follows,
Step 401, according to hidden layer node number and its weight difference, data matrix is established;
Step 402, normalized data matrix, obtains normalization matrix;
Step 403, according to normalization matrix, the entropy of hidden layer node weight difference is calculated;
Step 404, coefficient of variation is calculated according to the entropy of gained hidden layer node;
Step 405, the entropy weight of hidden layer hiding node is calculated according to gained coefficient of variation;
Step 406, the evaluation of estimate of hidden layer node is calculated with following formula according to the entropy weight acquired.
Thus the evaluation of estimate of all hidden layer nodes is found out, the value of X is smaller, then the contribution that the node updates weight is got over It is small.Then the evaluation of estimate of each hidden layer node is analyzed, by hidden layer node evaluation of estimate the smallest 20 percent Hidden layer node, which screens, to be freezed, and institute's evaluation values threshold value is the evaluation of estimate at 20 percent this separation, In every batch of training simultaneously, evaluation of estimate threshold value can change, in generally 0-0.05 a number.Freeze entropy weight in entropy Weight threshold hidden layer node below no longer updates this in subsequent algorithm to maintain its weight amount as initial weight amount The weight amount of the frozen hidden layer node in part, the weight amount of this part hidden layer node will be maintained at its initial weight amount. Simultaneously from the output layer of convolutional neural networks to the hidden layer of the front (be the system to convolutional layer and pond layer before output layer Claim) hidden layer node that is not frozen of back transfer.
Step 105, by reverse conduction algorithm update transmitting come be not frozen hidden layer node weight amount, obtain It is not frozen the updated weight amount of hidden layer node, to get the weight amount of current each hidden layer node;
Specifically, utilizing reverse conduction formula to not frozen hidden layer node, its weight amount is updated, has been frozen at this time The hidden layer node of knot keeps its current weight amount, along with the not frozen updated weight amount of hidden layer node, is The weight amount of current each hidden layer node.
Step 106, error judgment is carried out according to current each hidden layer node weight amount, if error is not default first In range, then repeatedly step 103~105, until error stops the training of this batch in the first preset range;
Specifically, the network for working as previous training is calculated by following formula according to the weight amount of current each hidden layer node Error E (w):
Wherein k is the quantity of mode, and C is the quantity of hidden layer node, tpiIt is the output target value of hidden layer node p.Spi It is real output value of the hidden layer node p in convolutional neural networks.
Wherein, h is the node in hidden layer in network, xiIt is a n dimension input pattern, i=1,2 ..., k, vmIt is to connect Connect the C dimensional vector of the weight of the arc of m-th of hidden layer node and output layer.The activation primitive of output layer is sigmoid function σ (y)=1/ (1+e-y), the activation primitive of hidden layer is hyperbolic tangent function:
δ (y)=(ey-e-y)/(ey+e-y)
At this point, if the error amount acquired within the first preset range (generally 0~0.05), i.e. hidden layer node at this time The output target value t of ppiWith hidden layer node p convolutional neural networks real output value SpiBetween almost without deviation, say Bright batch training objective is smoothly completed, and can terminating the training of this batch, (step 103- step 105 is the training of this batch Overall process).Conversely, this batch training objective is not reached if the error amount acquired is not within tolerance interval, then repeat Step 103- step 105 again inputs this batch training image data into convolutional neural networks, to convolutional neural networks In hidden layer node carry out forward calculation training, be based on entropy weight method frozen fraction hidden layer node, reversed update is not frozen The weight amount for tying hidden layer node, until the obtained error amount of weight amount according to all hidden layer nodes is in the first default model Enclose interior this batch of end training.
Step 107, it uses multiple batch training datas instead to be trained convolutional neural networks, so that the error criterion of network In the second preset range.
Specifically, inputting the 2nd, the 3rd, the the 4th ... the 120th training image data after the training of a batch Step 103- step 106 is repeated, the training of remaining batch training image data is completed, so that so that the error criterion of network is the In two preset ranges.Each of this example image data is all one obtained from MNIST hand-written volume data identification library The image of handwritten numeral, since the writing of different people is different, so how by convolutional neural networks to identify that writing habit is poor Not biggish same numbers are that our problems to be solved that is, by training network pair want that network can be made to carry out it correctly Identification.Therefore we are trained network using training image data (60000) after so that network performance is tended towards stability, and utilize Test image data (10000) are tested for the property network, to show that (i.e. can network correctly identify writing to test result It is accustomed to the biggish same numbers of difference), the fault rate of network is found out by test result.The default error range of second at this time is It is a restriction to the fault rate of network overall operation result, generally requires network fault rate within 1 15.
Further, we by original BP convolutional neural networks and improve BP convolutional neural networks operation result (including instruction Practice time and fault rate) it is compared, obtain comparison result as shown in Table 1.
Table 1
It can be seen that improved BP convolutional neural networks through the invention from above-mentioned chart, when averagely training every time Between can reduce 25 seconds, be greatly shortened convolutional neural networks training duration.The improved BP convolutional Neural net of the present invention simultaneously Although (fault rate improves 2% or so, within an acceptable range) the generally training time that increases on network error rate is reduced 13.9%.Therefore improved convolutional neural networks significantly improve net in the case where sacrificing partially acceptable accuracy The arithmetic speed of network makes network training complexity reach 5% the above object of expected reduction.Secondly the improved volume of the present invention Product neural network can be working properly, network does not fall into endless loop or error rate is excessively high, illustrate present invention effectively prevents The problems such as step-length and local minimum for being easy to appear in traditional BP convolutional neural networks, has stronger robustness.
Embodiment 2
Fig. 5 show it is according to an embodiment of the present invention improve convolutional neural networks device, i.e., electronic equipment 310 (such as Have program execute function computer server) comprising at least one processor 311, power supply 314, and with it is described extremely The memory 312 and input/output interface 313 of a few processor 311 communication connection;The memory 312 is stored with can be by institute State the instruction of at least one processor 311 execution, described instruction executed by least one described processor 311 so that it is described extremely A few processor 311 is able to carry out method disclosed in aforementioned any embodiment;The input/output interface 313 may include Display, keyboard, mouse and USB interface are used for inputoutput data;Power supply 314 is used to provide electricity for electronic equipment 310 Energy.
It will be appreciated by those skilled in the art that: realize that all or part of the steps of above method embodiment can pass through program Relevant hardware is instructed to complete, program above-mentioned can store in computer-readable storage medium, which is executing When, execute step including the steps of the foregoing method embodiments;And storage medium above-mentioned includes: movable storage device, read-only memory The various media that can store program code such as (Read Only Memory, ROM), magnetic or disk.
When the above-mentioned integrated unit of the present invention be realized in the form of SFU software functional unit and as the sale of independent product or In use, also can store in a computer readable storage medium.Based on this understanding, the skill of the embodiment of the present invention Substantially the part that contributes to existing technology can be embodied in the form of software products art scheme in other words, the calculating Machine software product is stored in a storage medium, including some instructions are used so that a computer equipment (can be individual Computer, server or network equipment etc.) execute all or part of each embodiment the method for the present invention.And it is aforementioned Storage medium include: the various media that can store program code such as movable storage device, ROM, magnetic or disk.

Claims (10)

1. a kind of method for improving convolutional neural networks based on frost weight, which comprises the following steps:
Step 101, pretreatment image obtains multiple batch training image data;
Step 102, convolutional neural networks are constructed, and initialization tax is carried out to the hidden layer node weight amount in convolutional neural networks Value;
Step 103, the training image data for inputting a batch calculate convolutional neural networks by forward conduction algorithm, The activation weight of each hidden layer node in convolutional neural networks is obtained, according to each hidden layer node of the activation weight computing Weight difference;
Step 104, the weight difference is analyzed based on entropy weight method, to obtain the entropy weight of each hidden layer node, and with this The evaluation of estimate for determining hidden layer node screens the hidden layer node based on the evaluation of estimate of hidden layer node, freezes to comment Value in evaluation of estimate threshold value hidden layer node below to maintain its weight amount as current weight amount, and transmit be not frozen it is hidden Containing node layer;
Step 105, by reverse conduction algorithm update transmitting come be not frozen hidden layer node weight amount, obtain not by Freeze the updated weight amount of hidden layer node, to get the weight amount of current each hidden layer node;
Step 106, error judgment is carried out according to current each hidden layer node weight amount, if error is not in the first preset range It is interior, then repeatedly step 103~105, until error stops the training of this batch in the first preset range;
Step 107, it uses multiple batch training image data instead to be trained convolutional neural networks, so that the error criterion of network In the second preset range.
2. the method as described in claim 1, which is characterized in that the pretreatment image is normalized from hand-written volume data The a certain number of image datas got in identification library, by all image data scaleds to unified scale.
3. the method as described in claim 1, which is characterized in that the convolutional neural networks include an input layer, at least two A convolutional layer and at least two sample levels, an output layer, wherein output layer merges full articulamentum.
4. the method as described in claim 1, which is characterized in that the weight difference is the activation weight of gained hidden layer node With the absolute value of the difference of hidden layer node current weight amount.
5. the method as described in claim 1, which is characterized in that analyze the weight based on entropy weight method described in step 104 Difference to obtain the entropy weight of each hidden layer node, and is specifically wrapped with this step of this evaluation of estimate for determining hidden layer node It includes:
According to hidden layer node number and its weight difference, data matrix is established;The normalized data matrix, obtains normalizing Change matrix;According to normalization matrix, the entropy of hidden layer node weight difference is calculated;It is poor to be calculated according to the entropy of gained hidden layer node Different coefficient;The entropy weight of hidden layer node is calculated according to gained coefficient of variation;Hidden layer node is calculated according to the entropy weight acquired Evaluation of estimate.
6. method as claimed in claim 4, which is characterized in that the data matrix are as follows:
Wherein, m is hidden layer node number, and n is hidden layer node weight difference.
7. the method as described in claim 1, which is characterized in that the weight amount of current each hidden layer node includes being frozen The initial weight amount of the hidden layer node of knot and the not frozen updated weight amount of hidden layer node.
8. the method as described in claim 1, which is characterized in that the error judgment formula are as follows:
Wherein, E (w, v) is the current error amount of convolutional neural networks, and wherein k is the quantity of mode, and C is the number of hidden layer node Amount, tpiIt is the output target value of hidden layer node p, spiIt is real output value of the hidden layer node p in convolutional neural networks.
9. the method as described in claim 1, which is characterized in that step 107 includes, by the training image data of multiple batches point It is not input in convolutional neural networks, repeats step 103- step 106, execute multiple batches of training image data to the instruction of network Practice.
10. a kind of device for improving convolutional neural networks based on frost weight, which is characterized in that including at least one processor, And the memory being connect at least one described processor communication;The memory is stored with can be by least one described processing The instruction that device executes, described instruction is executed by least one described processor, so that at least one described processor is able to carry out Method described in any one of claims 1 to 9.
CN201811044605.5A 2018-09-07 2018-09-07 Method and device for improving convolutional neural network based on freezing weight Active CN109146000B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811044605.5A CN109146000B (en) 2018-09-07 2018-09-07 Method and device for improving convolutional neural network based on freezing weight

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811044605.5A CN109146000B (en) 2018-09-07 2018-09-07 Method and device for improving convolutional neural network based on freezing weight

Publications (2)

Publication Number Publication Date
CN109146000A true CN109146000A (en) 2019-01-04
CN109146000B CN109146000B (en) 2022-03-08

Family

ID=64823890

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811044605.5A Active CN109146000B (en) 2018-09-07 2018-09-07 Method and device for improving convolutional neural network based on freezing weight

Country Status (1)

Country Link
CN (1) CN109146000B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110263925A (en) * 2019-06-04 2019-09-20 电子科技大学 A kind of hardware-accelerated realization framework of the convolutional neural networks forward prediction based on FPGA
CN111222465A (en) * 2019-11-07 2020-06-02 深圳云天励飞技术有限公司 Image analysis method based on convolutional neural network and related equipment
CN111624964A (en) * 2020-05-27 2020-09-04 甬矽电子(宁波)股份有限公司 Dynamic authority management and control method, device, server and readable storage medium
TWI731511B (en) * 2019-12-12 2021-06-21 國立中興大學 Finger vein or palm vein identification processing and neural network training method
CN116989510A (en) * 2023-09-28 2023-11-03 广州冰泉制冷设备有限责任公司 Intelligent refrigeration method combining frosting detection and hot gas defrosting

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5845051A (en) * 1995-09-15 1998-12-01 Electronics And Telecommunications Research Institute Learning method for multilayer perceptron neural network with N-bit data representation
US20130061114A1 (en) * 2011-09-02 2013-03-07 Samsung Electronics Co., Ltd. Freezing-based ldpc decoder and method
CN105654729A (en) * 2016-03-28 2016-06-08 南京邮电大学 Short-term traffic flow prediction method based on convolutional neural network
US20170109243A1 (en) * 2015-10-16 2017-04-20 Business Objects Software, Ltd. Model-Based System and Method for Undoing Actions in an Application
CN107194376A (en) * 2017-06-21 2017-09-22 北京市威富安防科技有限公司 Mask fraud convolutional neural networks training method and human face in-vivo detection method
CN107301396A (en) * 2017-06-21 2017-10-27 北京市威富安防科技有限公司 Video fraud convolutional neural networks training method and human face in-vivo detection method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5845051A (en) * 1995-09-15 1998-12-01 Electronics And Telecommunications Research Institute Learning method for multilayer perceptron neural network with N-bit data representation
US20130061114A1 (en) * 2011-09-02 2013-03-07 Samsung Electronics Co., Ltd. Freezing-based ldpc decoder and method
US20170109243A1 (en) * 2015-10-16 2017-04-20 Business Objects Software, Ltd. Model-Based System and Method for Undoing Actions in an Application
CN105654729A (en) * 2016-03-28 2016-06-08 南京邮电大学 Short-term traffic flow prediction method based on convolutional neural network
CN107194376A (en) * 2017-06-21 2017-09-22 北京市威富安防科技有限公司 Mask fraud convolutional neural networks training method and human face in-vivo detection method
CN107301396A (en) * 2017-06-21 2017-10-27 北京市威富安防科技有限公司 Video fraud convolutional neural networks training method and human face in-vivo detection method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
HINTON G E等: "Improving neural networks by preventing co-adaptation of feature detectors", 《COMPUTER SCIENCE》 *
RUI ZHANG等: "Extreme Learning Machine with Adaptive Growth of Hidden Nodes and Incremental Updating of Output Weights", 《AIS 2011: AUTONOMOUS AND INTELLIGENT SYSTEMS》 *
高大文等: "人工神经网络中隐含层节点与训练次数的优化", 《哈尔滨工业大学学报》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110263925A (en) * 2019-06-04 2019-09-20 电子科技大学 A kind of hardware-accelerated realization framework of the convolutional neural networks forward prediction based on FPGA
CN110263925B (en) * 2019-06-04 2022-03-15 电子科技大学 Hardware acceleration implementation device for convolutional neural network forward prediction based on FPGA
CN111222465A (en) * 2019-11-07 2020-06-02 深圳云天励飞技术有限公司 Image analysis method based on convolutional neural network and related equipment
TWI731511B (en) * 2019-12-12 2021-06-21 國立中興大學 Finger vein or palm vein identification processing and neural network training method
CN111624964A (en) * 2020-05-27 2020-09-04 甬矽电子(宁波)股份有限公司 Dynamic authority management and control method, device, server and readable storage medium
CN116989510A (en) * 2023-09-28 2023-11-03 广州冰泉制冷设备有限责任公司 Intelligent refrigeration method combining frosting detection and hot gas defrosting

Also Published As

Publication number Publication date
CN109146000B (en) 2022-03-08

Similar Documents

Publication Publication Date Title
CN109146000A (en) A kind of method and device for improving convolutional neural networks based on frost weight
Cui et al. Efficient human motion prediction using temporal convolutional generative adversarial network
Chen et al. Dynamical channel pruning by conditional accuracy change for deep neural networks
Zhou et al. Context-reinforced semantic segmentation
CN112052948B (en) Network model compression method and device, storage medium and electronic equipment
CN112613581B (en) Image recognition method, system, computer equipment and storage medium
CN111294646A (en) Video processing method, device, equipment and storage medium
Liu et al. CNN-FCM: System modeling promotes stability of deep learning in time series prediction
CN109817276A (en) A kind of secondary protein structure prediction method based on deep neural network
CN110383299A (en) The generation time model of memory-enhancing effect
CN105550749A (en) Method for constructing convolution neural network in novel network topological structure
Maliar et al. Will artificial intelligence replace computational economists any time soon?
CN108596327A (en) A kind of seismic velocity spectrum artificial intelligence pick-up method based on deep learning
US11538178B2 (en) Machine learning-based 2D structured image generation
CN111199216B (en) Motion prediction method and system for human skeleton
CN106796533A (en) It is adaptive selected the system and method for execution pattern
CN113343901A (en) Human behavior identification method based on multi-scale attention-driven graph convolutional network
WO2020030052A1 (en) Animal count identification method, device, medium, and electronic apparatus
CN110033089A (en) Deep neural network parameter optimization method and system based on Distributed fusion algorithm
CN112529149A (en) Data processing method and related device
de Prado et al. Automated design space exploration for optimized deployment of dnn on arm cortex-a cpus
Jaiswal et al. Deep neural network compression via knowledge distillation for embedded applications
US11568621B2 (en) Dynamic character model fitting of three-dimensional digital items
CN113111308B (en) Symbolic regression method and system based on data-driven genetic programming algorithm
CN114863508A (en) Expression recognition model generation method, medium and device of adaptive attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant