CN109816092A - Deep neural network training method, device, electronic equipment and storage medium - Google Patents

Deep neural network training method, device, electronic equipment and storage medium Download PDF

Info

Publication number
CN109816092A
CN109816092A CN201811528375.XA CN201811528375A CN109816092A CN 109816092 A CN109816092 A CN 109816092A CN 201811528375 A CN201811528375 A CN 201811528375A CN 109816092 A CN109816092 A CN 109816092A
Authority
CN
China
Prior art keywords
training sample
neural network
training
loss function
network model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811528375.XA
Other languages
Chinese (zh)
Other versions
CN109816092B (en
Inventor
柴振华
孟欢欢
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sankuai Online Technology Co Ltd
Original Assignee
Beijing Sankuai Online Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sankuai Online Technology Co Ltd filed Critical Beijing Sankuai Online Technology Co Ltd
Priority to CN201811528375.XA priority Critical patent/CN109816092B/en
Publication of CN109816092A publication Critical patent/CN109816092A/en
Application granted granted Critical
Publication of CN109816092B publication Critical patent/CN109816092B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

This application discloses a kind of deep neural network training methods, belong to field of computer technology, for solving the problems, such as that neural network trained in the prior art performance under complex scene is lower.The described method includes: obtaining several training samples for being provided with pre-set categories label, it is based on several training samples, training neural network model;Wherein, the loss function of the neural network model is used to be weighted according to first weight directly proportional to the differentiation difficulty of each training sample, determines the penalty values of the neural network model.Deep neural network training method disclosed in the embodiment of the present application, importance of the biggish training sample of difficulty in training sample is distinguished by adaptive boosting, it avoids distinguishing the neural network mistake classification that the biggish sample of difficulty is trained to, facilitates the performance for promoting the neural network.

Description

Deep neural network training method, device, electronic equipment and storage medium
Technical field
This application involves field of computer technology, more particularly to a kind of deep neural network training method, device, electronics Equipment and storage medium.
Background technique
In recent years, deep learning achieves significant progress in area of pattern recognition, and key factor may include: network Model is abundant flexibly, operational capability is strong, more adapts to big data processing.As neural network is applied to different task, for mind Improvement through network model is also the critical issue of those skilled in the art's research.Neural network model is changed in the prior art Into be concentrated mainly on two aspect: network structure and loss function.Wherein, the loss function for being usually used in disaggregated model training is main It is on Softmax loss and 2016 worlds Nian top meeting ECCV based on the improved Center loss of Softmax.However, Applicant by the prior art use Center loss as loss function neural network progress the study found that if The training sample weaker there are the biggish sample of noise or distinction, utilizes existing damage in the training sample of training neural network It loses the model that function training obtains to be tested, the classification of neural network or the promotion of recognition result for obtain training all have Limit.Therefore, deep neural network training is carried out the characteristics of combined training sample, it can the obtained neural network of training for promotion Performance, and then promote the accuracy that the obtained neural network of application training carries out Classification and Identification.
Summary of the invention
The application provides a kind of deep neural network training method, facilitates the property for the neural network that training for promotion obtains Can, to promote the accuracy that the neural network that application training obtains carries out Classification and Identification.
To solve the above-mentioned problems, in a first aspect, the embodiment of the present application provides a kind of deep neural network training method, Include:
Obtain several training samples for being provided with pre-set categories label;
Based on several training samples, training neural network model;
Wherein, the loss function of the neural network model is used for according to the differentiation difficulty with each training sample at just First weight of ratio is weighted, and determines the penalty values of the neural network model, the differentiation of each training sample Away from directly proportional at a distance from respective classes center, the respective classes center is to several training for difficulty and the training sample The class center of the classification in classification comprising the training sample is obtained after sample clustering.
Second aspect, the embodiment of the present application provide a kind of deep neural network training device, comprising:
Training sample obtains module, for obtaining several training samples for being provided with pre-set categories label;
Model training module, for being based on several training samples, training neural network model;
Wherein, the loss function of the neural network model is used for according to the differentiation difficulty with each training sample at just First weight of ratio is weighted, and determines the penalty values of the neural network model.
The third aspect, the embodiment of the present application also disclose a kind of electronic equipment, including memory, processor and are stored in institute The computer program that can be run on memory and on a processor is stated, the processor realizes this when executing the computer program Apply for deep neural network training method described in embodiment.
Fourth aspect, the embodiment of the present application provide a kind of computer readable storage medium, are stored thereon with computer journey Sequence, when which is executed by processor disclosed in the embodiment of the present application the step of deep neural network training method.
Deep neural network training method disclosed in the embodiment of the present application, if being provided with pre-set categories label by obtaining Dry training sample is based on several training samples, training neural network model;Wherein, the loss of the neural network model Function is used to be weighted according to first weight directly proportional to the differentiation difficulty of each training sample, determines the mind Penalty values through network model, the performance under complex scene for solving the neural network that training obtains in the prior art are lower Problem.Deep neural network training method disclosed in the embodiment of the present application is mentioned automatically by improving the loss function of neural network The importance for distinguishing the biggish training sample of difficulty in training sample is risen, the differentiation biggish training sample of difficulty is avoided to be trained to Obtained neural network mistake classification, facilitates the performance for the neural network that training for promotion obtains, to promote what application obtained The accuracy of neural network progress Classification and Identification.
Detailed description of the invention
Technical solution in ord to more clearly illustrate embodiments of the present application, below will be in embodiment or description of the prior art Required attached drawing is briefly described, it should be apparent that, the accompanying drawings in the following description is only some realities of the application Example is applied, it for those of ordinary skill in the art, without any creative labor, can also be attached according to these Figure obtains other attached drawings.
Fig. 1 is the deep neural network training method flow chart of the embodiment of the present application one;
Fig. 2 is cluster result schematic diagram in the deep neural network training method of the embodiment of the present application one;
Fig. 3 is the flow chart that the neural network training method based on the embodiment of the present application one carries out object classification identification;
Fig. 4 is one of deep neural network training device structural schematic diagram of the embodiment of the present application three;
Fig. 5 is the deep neural network training device second structural representation of the embodiment of the present application three;
Fig. 6 is the deep neural network training device third structural representation of the embodiment of the present application three.
Specific embodiment
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application carries out clear, complete Site preparation description, it is clear that described embodiment is some embodiments of the present application, instead of all the embodiments.Based on this Shen Please in embodiment, every other implementation obtained by those of ordinary skill in the art without making creative efforts Example, shall fall in the protection scope of this application.
Embodiment one
A kind of deep neural network training method disclosed in the present embodiment, as shown in Figure 1, this method comprises: step 110 and Step 120.
Step 110, several training samples for being provided with pre-set categories label are obtained.
Before training neural network, it is necessary first to obtain several training samples for being provided with pre-set categories label.
Different according to specific application scenarios, the form of training sample is different.For example, in work clothes identification application, Training sample is work clothes image;In face In vivo detection application scenarios, training sample image acquires the living body people of equipment acquisition The image and non-living body face (such as faceform, human face photo) of face;In voice recognition application scene, training sample is one section Audio.
Difference is exported according to specific identification mission, the class label of training sample is different.Work clothes are executed with training For the neural network of identification mission, exported according to specific identification mission, the classification of training sample may include being used to indicate Meituan takes out work clothes and the class label of the different work clothes classifications such as work clothes is taken out by Baidu.The mind of voice recognition task is executed with training For network, exported according to specific identification mission, the classification of training sample may include being used to indicate boy student and schoolgirl etc. The class label of alternative sounds classification.Again by taking training executes the neural network of face In vivo detection task as an example, training sample Classification may include the class label for being used to indicate living body faces and non-living body face two categories.
Step 120, several training samples, training neural network model are based on.
Wherein, the loss function of the neural network model is used for according to the differentiation difficulty with each training sample at just First weight of ratio is weighted, and determines the penalty values of the neural network model.
In some embodiments of the present application, before training neural network model, further includes: to the same category mark Several training samples of label are clustered, and determine the class center of category training sample.Each training sample Difficulty is distinguished with the training sample away from directly proportional at a distance from respective classes center.The specific method that training sample is clustered Referring to the prior art, repeated no more in the embodiment of the present application.
After getting and being provided with the training sample of pre-set categories label, deep neural network is further constructed, then, Based on the training sample, the neural network model of training building.
The application is when it is implemented, major network can choose ResNet (residual error network), such as ResNet50 residual error network, Then, by improving the loss function and training process of neural network, with the performance for the neural network that training for promotion obtains.Example Such as, it sets the loss function of the neural network to by softmax loss function and the center based on attention mechanism is lost The associated losses function that function is constituted, by improving the loss function of neural network, in the output penalty values for calculating loss function When, adaptive adjustment and cluster centre must influence the penalty values that loss function exports apart from biggish sample, i.e., adaptive to adjust Difficulty biggish sample in main plot point must influence the penalty values that loss function exports, to avoid the biggish sample of difficulty is distinguished in mould It is ignored in type training process, so as to cause the obtained category of model of training or the decline of the accuracy rate of identification.
When it is implemented, with the loss function of neural network model are as follows:Citing, whereinIt indicates Loss function,For softmax loss function, indicate are as follows: For based on note The center loss function for power mechanism of anticipating indicates are as follows:
In above-mentioned formula, i indicates the mark of training sample, and m indicates training The sum of sample, yiIndicate the classification logotype that the loss function is input in the neural network model, xiExpression belongs to classification yiA training sample,The jth column of the weight matrix of the last one full articulamentum, b before indicating the loss functionj The jth of the last one full articulamentum deviation b arranges before indicating loss function,Indicate the last one before the loss function The y of the weight matrix of full articulamentumiColumn,The y of the last one full articulamentum deviation b before expression loss functioniColumn,Indicate training sample xiThe first weight, with cluster centreBetween distance it is directly proportional, m indicate training sample quantity, T Indicate that transposition, λ indicate scalar.γ indicates scalar, and value is the number greater than 0, as γ is equal to 2.
When it is implemented, the closer sample of distance classification hyperplane is easy mistake under influence of noise after for preliminary classification It is categorized into other classifications and system is caused to judge incorrectly.Preliminary classification result schematic diagram as shown in Figure 2, wherein sample 211, 212,221,222 distance classification hyperplane are closer, i.e., farther out apart from respective class center, under influence of noise, 211 He of sample 212 easy misclassifications are into classification 22, and similarly, the easy misclassification of sample 221 and 222 is into classification 21.If neural network Softmax function is used only in loss function, i.e., in above-mentioned formulaThen can due to only accounting for the distinction between sample, And identification (not only to separate, it is also necessary to have certain interval) is ignored, neural network low so as to cause classification effectiveness Model performance reduces.Because in a practical situation, for the sample of some updates, some closer samples of distance classification hyperplane Misclassification is easy under influence of noise causes system to judge incorrectly to other classifications.
Therefore training sample is comprehensively considered by increasing the center loss function based on attention mechanism in the prior art Influence of the distance between the classification to penalty values.However, unanimously treating to each training sample in the prior art, no matter sample is trained , apart from class center apart from size, influence of the training sample to the penalty values that loss function exports all is identical for this.And it passes through Repetition test is crossed, the class center distance of inventor's discovery such as training sample 211 and 212 and classification 21 in Fig. 2 is to loss letter The influence of several penalty values should not be identical.For training the neural network of work clothes classification task, Meituan work clothes and puts me and reach Work clothes belong to the classification that Meituan takes out work clothes, and Meituan work clothes put me up to work usually apart from the closer work clothes of cluster centre Clothes are due to negligible amounts, after classification, put me up to work clothes as apart from the farther away work clothes of cluster centre.If calculating loss function Penalty values when, it is all identical for putting the influence for the penalty values that I exports up to work clothes and Meituan work clothes, that is, puts me up to work clothes and beauty Group's work clothes have same sample weights, then will lead to trained neural network when carrying out work clothes identification, and mistake will put me and reach Work clothes and Meituan work clothes are categorized into different classes of.Therefore, the present embodiment passes through adaptively according to training sample and cluster centre Distance is that corresponding weight is arranged in training sample, distinguishes the biggish training sample of difficulty to loss function to be promotedPenalty values Influence.When it is implemented, training sample xiWeight, with cluster centreBetween distance it is directly proportional.
In specific training process, computer is corresponding by each training sample for executing above-mentioned loss function calculating input Loss function value, and compare the difference between the loss function value being calculated and sample label;Then, by constantly adjusting damage The parameter for losing function, calculates the loss function value of the corresponding different parameters of each training sample of input repeatedly, final to determine completely The smallest parameter of difference between sufficient loss function value and sample label, the parameter as loss function.And based on determining ginseng Number calculates the corresponding output valve of sample of input neural network.The output valve of neural network corresponds to the classification results of input sample.
It, can be by reverse conduction and gradient descent method optimization neural network, so that loss function is defeated when specific training Penalty values out are minimum, thus the neural network after being optimized.
In some embodiments of the present application, the method that further can also adjust cluster centre to gradient descent method carries out excellent Change, by considering that influence of the training sample when adjusting cluster centre is arranged in the differentiation difficulty of training sample simultaneously, with further The performance for the network model that training for promotion obtains.
The specific training process of neural network may refer to the prior art, repeat no more in the present embodiment.
The application using infrastructure network of other network structures as neural network when it is implemented, can also be appointed All there is the optimization training of loss function in the training process of what network structure, the application to the specific network structure of neural network not It limits, only the realization of loss function and optimization method is defined.
Deep neural network training method disclosed in the embodiment of the present application, if being provided with pre-set categories label by obtaining Dry training sample, and the identical several training samples of label are clustered, determine class center;Then, based on described Several training samples, training neural network model;Wherein, the loss function of the neural network model be used for according to it is each described The first weight that the differentiation difficulty of training sample is directly proportional is weighted, and determines the penalty values of the neural network model, Solve the problems, such as that the performance under complex scene for the neural network that training obtains in the prior art is lower.The embodiment of the present application is public The deep neural network training method opened, by improving the loss function of neural network, it is biggish that adaptive boosting distinguishes difficulty Importance of the training sample in training sample avoids distinguishing the neural network mistake that the biggish training sample of difficulty is trained to Misclassification facilitates the performance for the neural network that training for promotion obtains.
Embodiment two
Based on embodiment one, a kind of prioritization scheme of deep neural network training method disclosed in the present embodiment.
When it is implemented, building is neural first after getting several training samples for being provided with pre-set categories label Network.Still it uses ResNet50 (residual error network) as basic network in the present embodiment, constructs neural network, the nerve net Network includes multiple feature extraction layers.Neural network successively calls each feature extraction layer (as entirely by carrying out the propagated forward stage Articulamentum) forward direction function, obtain layer-by-layer output, the last layer obtains loss function compared with objective function, calculates error Updated value.Then, first layer is successively reached by backpropagation, all weights update together at the end of backpropagation.Finally For one feature extraction layer using the feature of extraction as the predicted value input value loss function of neural network, loss function passes through a system Column count obtains the difference of predicted value and true tag, is determined as the penalty values of neural network.The purpose of training neural network is just It is to keep predicted value and the difference of true tag minimum.
In other preferred embodiments of the application, the first weightCan by training sample and class center away from From normal distyribution function indicate, such as:Wherein, σ is constant, xiExpression belongs to class Other yiA training sample,Indicate classification yiCenter.The differentiation difficulty of training sample it can be seen from above formula It is bigger, i.e. training sample xiWith its generic centerDistance it is bigger, the first weightValue is bigger.That is training sample Differentiation difficulty it is bigger, in training neural network, importance needs corresponding promoted.
In the other embodiments of the application, calculating the first weight can also be by training sample and class center distance Other proportional relationship formula indicate that the present embodiment does not enumerate.
Specific in the present embodiment, during backpropagation, by constantly adjusting cluster centre, so that loss function The penalty values of error are minimum between the prediction codomain true value of the expression training sample of output.When it is implemented, usually passing through public affairs FormulaTo update cluster centreThe application is in the renewal amount for updating the class centerWhen, weight Point considers the training sample of wanting close to class center, weakens the training sample far from class center.Therefore, the classification is being updated When center, the neural network model is used to be carried out according to the second weight that the differentiation difficulty with each training sample is inversely proportional Ranking operation determines the variable quantity at respective classes center.For the classification results described in Fig. 2, updated in calculated value described in The renewal amount of the class center of classification 22When, it is smaller similar to the differentiation difficulty of sample 222 that emphasis examines rate, i.e., in classification Contribution of the closer training sample of the heart to renewal amount.And the differentiation difficulty of similar sample 221 is larger, i.e., farther out apart from class center Training sample the contribution of renewal amount will be weakened.When it is implemented, can be by for different training samples, according to the instruction Practice sample and it is arranged corresponding weight at a distance from its generic center, to calculate the renewal amount of class center
For example, the second weight that the basis and the differentiation difficulty of each training sample are inversely proportional is weighted, The variable quantity for determining respective classes center includes: according to formula
Determine class center cjVariable quantity, wherein i indicates training sample Mark, m indicate the sum of training sample, j and yiIndicate the classification that the loss function is input in the neural network model Mark, xiExpression belongs to classification yiA training sample, qiIndicate the second weight, qiWith training sample xiWith class center cjIt Between distance be inversely proportional, δ () be Dirac function, δ=1 when condition meets in bracket, otherwise=0, αcFor in control category The scalar of the learning rate of the heart, value range are [0,1].
In conjunction with the embodiments one, it is indicated with loss function are as follows:Wherein, the center based on attention mechanism Loss function indicates are as follows: Example illustrates the concrete scheme for adjusting cluster centre.
By calculating center penalty values to the partial derivative of sample, the variable quantity of class center, specific derivation of equation mistake are determined Journey is as follows:
Further,Thus Out:
Wherein,σcFor constant, xiTable Show and belongs to classification yiA training sample,Indicate classification yiCenter.I.e. in some embodiments of the present application, second Weight qiIt can be indicated by training sample and the normal distyribution function of class center distance.
In the other embodiments of the application, the second weight qiIt can also be by its of training sample and class center distance He inversely prroportional relationship formula indicates that the present embodiment does not enumerate.Preferably, the calculation method of the second weight and the first weight Calculation method matching.
By when updating class center, weakening influence of the more indistinguishable training sample to class center, promoted simultaneously Be easier to influence of the training sample to class center distinguished, can neural network that further training for promotion obtains classification it is quasi- True property, promotes the performance of neural network model.
In other preferred embodiments of the application, the loss function of the neural network model is used for: in the mind When calculating the penalty values of the training sample through network model, the third that is inversely proportional by the classification accounting with the training sample Weight adjusts the penalty values of the training sample.
Specifically, the loss function of the neural network model indicates are as follows:
Wherein,It indicates Loss function, i indicate the mark of training sample, and m indicates the sum of training sample, yiIt indicates to input in the neural network model To the classification logotype of the loss function, xiExpression belongs to classification yiA training sample,Indicate the loss function it The jth of the weight matrix of the last one preceding full articulamentum arranges, bjThe last one full articulamentum deviation b before expression loss function Jth column,The y of the weight matrix of the last one full articulamentum before indicating the loss functioniColumn,Indicate loss The y of the last one full articulamentum deviation b before functioniColumn, (1-ki)γIndicate training sample xiThe first weight, kiValue is Number greater than 0 and less than 1, and kiWith training sample xiWith training sample xiThe distance at generic center is inversely proportional,It indicates Training sample xiThird weight,Value and classification yiAccounting of the training sample in full dose training sample be inversely proportional, T indicates that transposition, λ and γ indicate scalar.When it is implemented, kiOther inverse ratios of training sample and class center distance can be passed through Example relation formula indicates.
Embodiment three
The embodiment of the present application also discloses a kind of deep neural network training method, is applied in classification application.Such as Fig. 3 institute Show, which comprises step 310 to step 370.
Step 310, several training samples for being provided with pre-set categories label are obtained.
The application is when it is implemented, the training sample includes following any one: image, text, voice.It is directed to not Same object to be sorted needs to obtain the training sample of corresponding object to be sorted in neural network model.In the present embodiment, By taking training is for the neural network model of work clothes identification as an example, firstly, the work clothes image for being provided with different platform label is obtained, Such as: being provided with Meituan and take out the work clothes image of platform label, the work clothes image for being provided with the platform label that is hungry, be provided with Baidu Take out the work clothes image etc. of platform label.
Step 320, several training samples with the same category label are clustered, determines category training sample This class center.
The differentiation difficulty of each training sample is with the training sample away from directly proportional at a distance from respective classes center.
Several training samples with the same category label are clustered, determine the classification of category training sample Center is referring to embodiment two, and this embodiment is not repeated.
Step 330, several training samples, training neural network model are based on.
Wherein, the loss function of the neural network model is used for according to the differentiation difficulty with each training sample at just First weight of ratio is weighted, and determines the penalty values of the neural network model, the differentiation of each training sample Away from directly proportional at a distance from respective classes center, the respective classes center is to several training for difficulty and the training sample The class center of the classification in classification comprising the training sample is obtained after sample clustering.
Later, the work clothes image training neural network model based on acquisition.
When work clothes image based on acquisition trains neural network model, first to the same category mark in the training sample of acquisition The training sample of label is clustered, and determines the corresponding class center of training sample of every kind of class label.Cluster is obtained Each classification in training sample, sample characterizes the differentiation difficulty of the sample: sample and classification at a distance from class center The distance at center is bigger, and the differentiation difficulty of the sample is bigger;Conversely, sample is smaller at a distance from class center, the area of the sample Divide difficulty smaller.
The training process of neural network is exactly that computer calculates the trained sample of each of input by executing above-mentioned loss function This corresponding loss function value, and compare the difference between the loss function value being calculated and sample label;Then, by not The parameter of disconnected adjustment loss function, calculates the loss function value of the corresponding different parameters of each training sample of input, most repeatedly The smallest parameter of difference met between loss function value and sample label, the process of the parameter as loss function are determined eventually.
The specific embodiment of work clothes image training neural network model based on acquisition is referring to previous embodiment one and in fact Example two is applied, this embodiment is not repeated.
Step 340, the object data with the matched object to be sorted of the training sample is obtained by data acquisition equipment.
For different objects to be sorted, the number of objects of the object to be sorted is obtained by corresponding data acquisition equipment According to.For example, when object to be sorted is work clothes the image of take-away personnel can be acquired, by camera to obtain work clothes image Data.
Step 350, the characteristic of division of the object data is obtained.
Same method when according to training pattern obtains the feature vector of work clothes image data, as characteristic of division.
Step 360, the characteristic of division is input to the neural network model of training completion, obtains the nerve net The output result of network model.
The characteristic of division that will acquire is input to the neural network model that signature step training is completed, to obtain the mind Output result through network model.
Step 370, predetermined registration operation is executed according to the output result.
In some embodiments of the present application, it is described according to the output result execute predetermined registration operation include following any one Item is multinomial: showing the classification results of the object to be sorted corresponding with the output result;It is defeated according to the output result Access control signal out;Worksheet processing operation is executed according to the output result.
Specifically, identifying application scenarios for work clothes, when object to be identified is work clothes, this step acquires output Work clothes image classification is the confidence level of respective classes label.When it is implemented, platform is taken out for Meituan, it is a certain outer when recognizing When the personnel of selling wear Meituan work clothes, then take-away personnel can be executed with worksheet processing operation;It is worn when recognizing a certain take-away personnel When Baidu's work clothes, then worksheet processing operation is not executed to take-away personnel.
By neural network training method disclosed in the present application, in mistake of the training for the neural network of work clothes identification mission Cheng Zhong, since the training sample of Meituan work clothes is very more, and the training sample for putting me up to work clothes is fewer, Meituan work clothes and puts me Up to work clothes class label having the same, therefore, can suitably be promoted it is each put I up to work clothes training sample in sample training mistake Influence in journey is promoted and each puts me up to influence of the work clothes to the penalty values of the output of loss function, to avoid due to certain class Such sample caused by sample is less is ignored in assorting process, so as to cause trained obtained neural network to such sample The phenomenon that classification error, occurs.
Inventor is by being based on identical training dataset, the execution work clothes identification that training is constructed based on different loss functions The neural network model of task, and test discovery is carried out by the neural network that same test data set obtains training, it is based on The neural network model classification accuracy of the execution work clothes identification mission of the building of loss function disclosed in the present embodiment has very big It is promoted.For example, that is, using Meituan work clothes characteristics of image as test sample, being tested with the practical sampling observation figure of Meituan special delivery jockey When, the classification accuracy of the neural network based on the building of softmax loss function is 97.39%, to be based in the prior art Softmax loss function and the classification accuracy of the neural network of the center loss function building based on attention mechanism are 97.41%, and with it is disclosed in the present application based on softmax loss function and combine ranking operation based on attention mechanism The classification accuracy of the neural network of center loss function building is 98.17%.Figure is inspected by random samples with non-Meituan special delivery jockey (comes from portion Focus packet jockey, registration figure and the jockey's sampling observation figure for wearing other work clothes), i.e., with non-Meituan work clothes characteristics of image and non-dots, I reaches Work clothes characteristics of image is as test sample, when carrying out attack test, point of the neural network based on the building of softmax loss function Class is that the probability of Meituan work clothes is 0.41%, in the prior art based on softmax loss function and based on attention mechanism The probability that the classification of the neural network of center loss function building is similarly Meituan work clothes is 0.41%, and with disclosed in the present application Nerve net based on softmax loss function and the center loss function building based on attention mechanism for combining ranking operation The probability for being classified as Meituan work clothes of network is 0.40%, and classification error rate is decreased obviously.
Neural network training method disclosed in the embodiment of the present application is provided with several instructions of pre-set categories label by obtaining Practice sample, and several training samples with the same category label are clustered, determines the class of category training sample Then other center is based on several training samples, training neural network model during application on site, is acquired by data Equipment obtains the object data with the matched object to be sorted of the training sample, and further obtains point of the object data Category feature;Later, the characteristic of division is input to the neural network model of training completion, obtains the neural network mould The output of type is as a result, finally, execute predetermined registration operation according to the output result, due to the loss function of the neural network model For being weighted according to first weight directly proportional to the differentiation difficulty of each training sample, the nerve net is determined The penalty values of network model, also, the differentiation difficulty of each training sample and the training sample are away from respective classes center Apart from directly proportional, therefore, the accuracy that the model obtained based on training determines object classification can be promoted, to accurately execute pre- If operation.
Example IV
A kind of deep neural network training device disclosed in the present embodiment, as shown in figure 4, described device includes:
Training sample obtains module 410, for obtaining several training samples for being provided with pre-set categories label;
Model training module 420, for being based on several training samples, training neural network model;
Wherein, the loss function of the neural network model is used for according to the differentiation difficulty with each training sample at just First weight of ratio is weighted, and determines the penalty values of the neural network model.
Optionally, as shown in figure 5, described device further include:
Cluster module 430 determines such for clustering to several training samples with the same category label The class center of other training sample;
The differentiation difficulty of each training sample is with the training sample away from directly proportional at a distance from respective classes center.
Optionally, when updating the class center, the neural network model is used for basis and each training sample The second weight for being inversely proportional of differentiation difficulty be weighted, determine the variable quantity at respective classes center.
Optionally, the loss function of the neural network model is used for: calculating the training in the neural network model When the penalty values of sample, the third weight being inversely proportional by the classification accounting with the training sample adjusts the training sample Penalty values.
Further alternative, the loss function of the neural network model indicates are as follows:
Wherein, whereinIndicate that loss function, i indicate the mark of training sample, m indicates the sum of training sample, yiIndicate the neural network model In be input to the classification logotype of the loss function, xiExpression belongs to classification yiA training sample,Indicate the loss The jth column of the weight matrix of the last one full articulamentum, b before functionjThe last one full articulamentum before expression loss function The jth of deviation b arranges,The y of the weight matrix of the last one full articulamentum before indicating the loss functioniColumn,Table The y of the last one full articulamentum deviation b before showing loss functioniColumn, (1-ki)γIndicate training sample xiThe first weight, ki Value is number greater than 0 and less than 1, and kiWith training sample xiWith training sample xiThe distance at generic center is inversely proportional,Indicate training sample xiThird weight,Value and classification yiAccounting of the training sample in full dose training sample It is inversely proportional, T indicates that transposition, λ and γ indicate scalar.
In other preferred embodiments of the application,Wherein, σ is constant, xiIt indicates Belong to classification yiA training sample,Indicate classification yiCenter.The area of training sample it can be seen from above formula Divide difficulty bigger, i.e. training sample xiWith its generic centerDistance it is bigger, kiIt is smaller, the first weightValue is got over Greatly.That is the differentiation difficulty of training sample is bigger, and in training neural network, importance needs corresponding promoted.
During backpropagation, by constantly adjusting cluster centre, so that the expression training sample of loss function output The penalty values of error are minimum between this prediction codomain true value.When it is implemented, usually passing through formulaTo update cluster centreThe application is in the renewal amount for updating the class centerWhen, emphasis Consider the training sample of wanting close to class center, weakens the training sample far from class center.It is further alternative, the basis The second weight being inversely proportional with the differentiation difficulty of each training sample is weighted, and determines the variation at respective classes center Amount includes:
According to formulaDetermine class center cjVariable quantity, wherein i indicate instruction Practice the mark of sample, m indicates the sum of training sample, j and yiIt indicates to be input to the loss letter in the neural network model Several classification logotypes, xiExpression belongs to classification yiA training sample, qiIndicate the second weight, qiWith training sample xiAnd classification Center cjThe distance between be inversely proportional, δ () be Dirac function, δ=1 when condition meets in bracket, otherwise=0, αcFor control The scalar of the learning rate of class center processed, value range are [0,1].
In some embodiments of the present application,σcFor constant, xiExpression belongs to classification yi A training sample,Indicate classification yiCenter.
Optionally, the training sample includes following any one: image, text, voice.
Optionally, as shown in fig. 6, described device further include:
Data acquisition module 440, it is matched to be sorted right with the training sample for being obtained by data acquisition equipment The object data of elephant;
Feature obtains module 450, for obtaining the characteristic of division of the object data;
Model calling module 460 is obtained for the characteristic of division to be input to the neural network model of training completion Obtain the output result of the neural network model;
Execution module 470, for executing predetermined registration operation according to the output result.
Optionally, it is described according to the output result execute predetermined registration operation include following any one or more:
Show the classification results of the object to be sorted corresponding with the output result;
Access control signal is exported according to the output result;
Worksheet processing operation is executed according to the output result.
Deep neural network training device disclosed in the embodiment of the present application, for realizing the embodiment of the present application one and embodiment Each step of deep neural network training method described in two, the specific embodiment of each module of device is referring to corresponding step Suddenly, details are not described herein again.
Deep neural network training device disclosed in the embodiment of the present application, if being provided with pre-set categories label by obtaining Dry training sample, and several training samples with the same category label are clustered, determine category training sample Class center, be based on several training samples, training neural network model;Wherein, the loss of the neural network model Function is used to be weighted according to first weight directly proportional to the differentiation difficulty of each training sample, determines the mind Penalty values through network model, the performance under complex scene for solving the neural network that training obtains in the prior art are lower Problem.Deep neural network training device disclosed in the embodiment of the present application, by improving the loss function of neural network, adaptively The importance for distinguishing the biggish training sample of difficulty in training sample is promoted, the differentiation biggish training sample of difficulty is avoided to be instructed The network model mistake classification got, facilitates the performance for the neural network that training for promotion obtains.
Neural metwork training device disclosed in the embodiment of the present application during application on site, is obtained by data acquisition equipment The object data with the matched object to be sorted of the training sample is taken, and the classification for further obtaining the object data is special Sign;Later, the characteristic of division is input to the neural network model of training completion, obtains the neural network model Output is as a result, finally, execute predetermined registration operation according to the output result, since the loss function of the neural network model is used for It is weighted according to first weight directly proportional to the differentiation difficulty of each training sample, determines the neural network mould The penalty values of type, also, the differentiation difficulty of each training sample and the training sample away from respective classes center at a distance from It is directly proportional, therefore, the accuracy that the model obtained based on training determines object classification can be promoted, to accurately execute default behaviour Make.
Correspondingly, disclosed herein as well is a kind of electronic equipment, including memory, processor and it is stored in the memory Computer program that is upper and can running on a processor, the processor are realized when executing the computer program as the application is real Apply deep neural network training method described in example one to embodiment three.The electronic equipment can be PC machine, mobile terminal, a Personal digital assistant, tablet computer etc..
Disclosed herein as well is a kind of computer readable storage mediums, are stored thereon with computer program, which is located Manage the step of realizing the deep neural network training method as described in the embodiment of the present application one to embodiment three when device executes.
All the embodiments in this specification are described in a progressive manner, the highlights of each of the examples are with The difference of other embodiments, the same or similar parts between the embodiments can be referred to each other.For Installation practice For, since it is basically similar to the method embodiment, so being described relatively simple, referring to the portion of embodiment of the method in place of correlation It defends oneself bright.
A kind of deep neural network training method provided by the present application and device are described in detail above, herein Applying specific case, the principle and implementation of this application are described, and the explanation of above example is only intended to help Understand the present processes and its core concept;At the same time, for those skilled in the art, according to the thought of the application, There will be changes in the specific implementation manner and application range, in conclusion the content of the present specification should not be construed as to this The limitation of application.
Through the above description of the embodiments, those skilled in the art can be understood that each embodiment can It realizes by means of software and necessary general hardware platform, naturally it is also possible to pass through hardware realization.Based on such reason Solution, substantially the part that contributes to existing technology can embody above-mentioned technical proposal in the form of software products in other words Come, which may be stored in a computer readable storage medium, such as ROM/RAM, magnetic disk, CD, including Some instructions are used so that a computer equipment (can be personal computer, server or the network equipment etc.) executes respectively Method described in certain parts of a embodiment or embodiment.

Claims (20)

1. a kind of deep neural network training method characterized by comprising
Obtain several training samples for being provided with pre-set categories label;
Based on several training samples, training neural network model;
Wherein, the loss function of the neural network model is used for according to directly proportional to the differentiation difficulty of each training sample First weight is weighted, and determines the penalty values of the neural network model.
2. the method according to claim 1, wherein described be based on several training samples, training nerve net Before the step of network model, further includes:
Several training samples with the same category label are clustered, in the classification for determining category training sample The heart;
The differentiation difficulty of each training sample is with the training sample away from directly proportional at a distance from respective classes center.
3. according to the method described in claim 2, it is characterized in that, when updating the class center, the neural network mould Type is configured as being weighted according to the second weight that the differentiation difficulty with each training sample is inversely proportional, and determines corresponding The variable quantity of class center.
4. according to the method described in claim 3, it is characterized in that, the loss function of the neural network model is used for: in institute When stating neural network model and calculating the penalty values of the training sample, it is inversely proportional by the classification accounting with the training sample Third weight adjusts the penalty values of the training sample.
5. according to the method described in claim 4, it is characterized in that, the loss function of the neural network model indicates are as follows:
Wherein,Indicate loss Function, i indicate the mark of training sample, and m indicates the sum of training sample, yiIt indicates to be input to institute in the neural network model State the classification logotype of loss function, xiExpression belongs to classification yiA training sample, Wj TIndicate the loss function before most The jth of the weight matrix of the full articulamentum of the latter arranges, bjThe jth of the last one full articulamentum deviation b before expression loss function Column,The y of the weight matrix of the last one full articulamentum before indicating the loss functioniColumn,Indicate loss function The y of the last one full articulamentum deviation b beforeiColumn, (1-ki)γIndicate training sample xiThe first weight, kiValue be greater than 0 and the number less than 1, and kiWith training sample xiWith training sample xiThe distance at generic center is inversely proportional,Indicate training Sample xiThird weight,Value and classification yiAccounting of the training sample in full dose training sample be inversely proportional, T table Show that transposition, λ and γ indicate scalar.
6. according to the method described in claim 3, it is characterized in that, the differentiation difficulty of the basis and each training sample at Second weight of inverse ratio is weighted, and determines that the variable quantity at respective classes center includes:
According to formulaDetermine class center cjVariable quantity, wherein i indicate training sample This mark, m indicate the sum of training sample, j and yiIt indicates to be input to the loss function in the neural network model Classification logotype, xiExpression belongs to classification yiA training sample, qiIndicate the second weight, qiWith training sample xiAnd class center cjThe distance between be inversely proportional, δ () be Dirac function, δ=1 when condition meets in bracket, otherwise=0, αcTo control class The scalar of the learning rate at other center, value range are [0,1].
7. method according to any one of claims 1 to 6, which is characterized in that the training sample includes following any one : image, text, voice.
8. the method according to the description of claim 7 is characterized in that the method also includes:
The object data with the matched object to be sorted of the training sample is obtained by data acquisition equipment;
Obtain the characteristic of division of the object data;
The characteristic of division is input to the neural network model of training completion, obtains the output of the neural network model As a result;
Predetermined registration operation is executed according to the output result.
9. according to the method described in claim 8, it is characterized in that, described include according to output result execution predetermined registration operation It is any one or more below:
Show the classification results of the object to be sorted corresponding with the output result;
Access control signal is exported according to the output result;
Worksheet processing operation is executed according to the output result.
10. a kind of deep neural network training device characterized by comprising
Training sample obtains module, for obtaining several training samples for being provided with pre-set categories label;
Model training module, for being based on several training samples, training neural network model;
Wherein, the loss function of the neural network model is used for according to directly proportional to the differentiation difficulty of each training sample First weight is weighted, and determines the penalty values of the neural network model.
11. device according to claim 10, which is characterized in that described device further include:
Cluster module determines category training for clustering to several training samples with the same category label The class center of sample;
The differentiation difficulty of each training sample is with the training sample away from directly proportional at a distance from respective classes center.
12. device according to claim 11, which is characterized in that when updating the class center, the neural network Model is used to be weighted according to the second weight that the differentiation difficulty with each training sample is inversely proportional, and determines respective class The variable quantity at other center.
13. device according to claim 12, which is characterized in that the loss function of the neural network model is used for: When the neural network model calculates the penalty values of the training sample, it is inversely proportional by the classification accounting with the training sample Third weight, adjust the penalty values of the training sample.
14. device according to claim 13, which is characterized in that the loss function of the neural network model indicates are as follows:
Wherein,Indicate loss Function, i indicate the mark of training sample, and m indicates the sum of training sample, yiIt indicates to be input to institute in the neural network model State the classification logotype of loss function, xiExpression belongs to classification yiA training sample, Wj TIndicate the loss function before most The jth of the weight matrix of the full articulamentum of the latter arranges, bjThe jth of the last one full articulamentum deviation b before expression loss function Column,The y of the weight matrix of the last one full articulamentum before indicating the loss functioniColumn,Indicate loss function The y of the last one full articulamentum deviation b beforeiColumn, (1-ki)γIndicate training sample xiThe first weight, kiValue be greater than 0 and the number less than 1, and kiWith training sample xiWith training sample xiThe distance at generic center is inversely proportional,Indicate training Sample xiThird weight,Value and classification yiAccounting of the training sample in full dose training sample be inversely proportional, T table Show that transposition, λ and γ indicate scalar.
15. device according to claim 10, which is characterized in that the differentiation difficulty of the basis and each training sample The second weight being inversely proportional is weighted, and determines that the variable quantity at respective classes center includes:
According to formulaDetermine class center cjVariable quantity, wherein i indicate training sample This mark, m indicate the sum of training sample, j and yiIt indicates to be input to the loss function in the neural network model Classification logotype, xiExpression belongs to classification yiA training sample, qiIndicate the second weight, qiWith training sample xiAnd class center cjThe distance between be inversely proportional, δ () be Dirac function, δ=1 when condition meets in bracket, otherwise=0, αcTo control class The scalar of the learning rate at other center, value range are [0,1].
16. device according to any one of claims 10 to 15, which is characterized in that the training sample includes following any One: image, text, voice.
17. device according to claim 16, which is characterized in that described device further include:
Data acquisition module, for obtaining the object with the matched object to be sorted of the training sample by data acquisition equipment Data;
Feature obtains module, for obtaining the characteristic of division of the object data;
Model calling module, for the characteristic of division to be input to the neural network model of training completion, described in acquisition The output result of neural network model;
Execution module, for executing predetermined registration operation according to the output result.
18. device according to claim 17, which is characterized in that described to execute predetermined registration operation packet according to the output result It includes following any one or more:
Show the classification results of the object to be sorted corresponding with the output result;
Access control signal is exported according to the output result;
Worksheet processing operation is executed according to the output result.
19. a kind of electronic equipment, including memory, processor and it is stored on the memory and can runs on a processor Computer program, which is characterized in that the processor realizes claim 1 to 9 any one when executing the computer program The deep neural network training method.
20. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the program is by processor The step of deep neural network training method described in claim 1 to 9 any one is realized when execution.
CN201811528375.XA 2018-12-13 2018-12-13 Deep neural network training method and device, electronic equipment and storage medium Active CN109816092B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811528375.XA CN109816092B (en) 2018-12-13 2018-12-13 Deep neural network training method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811528375.XA CN109816092B (en) 2018-12-13 2018-12-13 Deep neural network training method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN109816092A true CN109816092A (en) 2019-05-28
CN109816092B CN109816092B (en) 2020-06-05

Family

ID=66602960

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811528375.XA Active CN109816092B (en) 2018-12-13 2018-12-13 Deep neural network training method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN109816092B (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110288085A (en) * 2019-06-20 2019-09-27 厦门市美亚柏科信息股份有限公司 A kind of data processing method, device, system and storage medium
CN110309856A (en) * 2019-05-30 2019-10-08 华为技术有限公司 Image classification method, the training method of neural network and device
CN110428052A (en) * 2019-08-01 2019-11-08 江苏满运软件科技有限公司 Construction method, device, medium and the electronic equipment of deep neural network model
CN110427466A (en) * 2019-06-12 2019-11-08 阿里巴巴集团控股有限公司 Training method and device for the matched neural network model of question and answer
CN110481561A (en) * 2019-08-06 2019-11-22 北京三快在线科技有限公司 Automatic driving vehicle automatic control signal generation method and device
CN110490054A (en) * 2019-07-08 2019-11-22 北京三快在线科技有限公司 Detection method, device, electronic equipment and the readable storage medium storing program for executing of target area
CN110532562A (en) * 2019-08-30 2019-12-03 联想(北京)有限公司 Neural network training method, Chinese idiom misuse detection method, device and electronic equipment
CN110929785A (en) * 2019-11-21 2020-03-27 中国科学院深圳先进技术研究院 Data classification method and device, terminal equipment and readable storage medium
CN111191769A (en) * 2019-12-25 2020-05-22 中国科学院苏州纳米技术与纳米仿生研究所 Self-adaptive neural network training and reasoning device
CN111310814A (en) * 2020-02-07 2020-06-19 支付宝(杭州)信息技术有限公司 Method and device for training business prediction model by utilizing unbalanced positive and negative samples
CN111429414A (en) * 2020-03-18 2020-07-17 腾讯科技(深圳)有限公司 Artificial intelligence-based focus image sample determination method and related device
CN111475618A (en) * 2020-03-31 2020-07-31 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN111488985A (en) * 2020-04-08 2020-08-04 华南理工大学 Deep neural network model compression training method, device, equipment and medium
CN111507396A (en) * 2020-04-15 2020-08-07 广州大学 Method and device for relieving error classification of neural network on unknown samples
CN111680631A (en) * 2020-06-09 2020-09-18 广州视源电子科技股份有限公司 Model training method and device
CN111783996A (en) * 2020-06-18 2020-10-16 杭州海康威视数字技术股份有限公司 Data processing method, device and equipment
CN111861909A (en) * 2020-06-29 2020-10-30 南京理工大学 Network fine-grained image denoising and classifying method
CN112085041A (en) * 2019-06-12 2020-12-15 北京地平线机器人技术研发有限公司 Training method and training device for neural network and electronic equipment
CN112381161A (en) * 2020-11-18 2021-02-19 厦门市美亚柏科信息股份有限公司 Neural network training method
CN112398862A (en) * 2020-11-18 2021-02-23 深圳供电局有限公司 Charging pile attack clustering detection method based on GRU model
WO2021036397A1 (en) * 2019-08-30 2021-03-04 华为技术有限公司 Method and apparatus for generating target neural network model
CN112561050A (en) * 2019-09-25 2021-03-26 杭州海康威视数字技术股份有限公司 Neural network model training method and device
CN112836816A (en) * 2021-02-04 2021-05-25 南京大学 Training method suitable for crosstalk of photoelectric storage and calculation integrated processing unit
CN113242547A (en) * 2021-04-02 2021-08-10 浙江大学 Method and system for filtering user behavior privacy in wireless signal based on deep learning and wireless signal receiving and transmitting device
WO2021238586A1 (en) * 2020-05-27 2021-12-02 华为技术有限公司 Training method and apparatus, device, and computer readable storage medium
CN114118413A (en) * 2021-11-30 2022-03-01 上海商汤临港智能科技有限公司 Network training and equipment control method, device, equipment and storage medium
CN114241260A (en) * 2021-12-14 2022-03-25 四川大学 Open set target detection and identification method based on deep neural network
WO2022174805A1 (en) * 2021-02-22 2022-08-25 上海商汤智能科技有限公司 Model training method and apparatus, image processing method and apparatus, electronic device and storage medium
CN116660389A (en) * 2023-07-21 2023-08-29 山东大禹水务建设集团有限公司 River sediment detection and repair system based on artificial intelligence

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107832700A (en) * 2017-11-03 2018-03-23 全悉科技(北京)有限公司 A kind of face identification method and system
CN108256450A (en) * 2018-01-04 2018-07-06 天津大学 A kind of supervised learning method of recognition of face and face verification based on deep learning
US20180293678A1 (en) * 2017-04-07 2018-10-11 Nicole Ann Shanahan Method and apparatus for the semi-autonomous management, analysis and distribution of intellectual property assets between various entities

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180293678A1 (en) * 2017-04-07 2018-10-11 Nicole Ann Shanahan Method and apparatus for the semi-autonomous management, analysis and distribution of intellectual property assets between various entities
CN107832700A (en) * 2017-11-03 2018-03-23 全悉科技(北京)有限公司 A kind of face identification method and system
CN108256450A (en) * 2018-01-04 2018-07-06 天津大学 A kind of supervised learning method of recognition of face and face verification based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YANDONG WEN, EL AT: "A Discriminative Feature Learning Approach for Deep Face Recognition", 《COMPUTER VISION-ECCV 2016》 *

Cited By (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110309856A (en) * 2019-05-30 2019-10-08 华为技术有限公司 Image classification method, the training method of neural network and device
CN110427466A (en) * 2019-06-12 2019-11-08 阿里巴巴集团控股有限公司 Training method and device for the matched neural network model of question and answer
CN112085041A (en) * 2019-06-12 2020-12-15 北京地平线机器人技术研发有限公司 Training method and training device for neural network and electronic equipment
CN110427466B (en) * 2019-06-12 2023-05-26 创新先进技术有限公司 Training method and device for neural network model for question-answer matching
CN110288085A (en) * 2019-06-20 2019-09-27 厦门市美亚柏科信息股份有限公司 A kind of data processing method, device, system and storage medium
CN110288085B (en) * 2019-06-20 2022-06-03 厦门市美亚柏科信息股份有限公司 Data processing method, device and system and storage medium
CN110490054A (en) * 2019-07-08 2019-11-22 北京三快在线科技有限公司 Detection method, device, electronic equipment and the readable storage medium storing program for executing of target area
CN110490054B (en) * 2019-07-08 2021-03-09 北京三快在线科技有限公司 Target area detection method and device, electronic equipment and readable storage medium
CN110428052A (en) * 2019-08-01 2019-11-08 江苏满运软件科技有限公司 Construction method, device, medium and the electronic equipment of deep neural network model
CN110481561A (en) * 2019-08-06 2019-11-22 北京三快在线科技有限公司 Automatic driving vehicle automatic control signal generation method and device
WO2021036397A1 (en) * 2019-08-30 2021-03-04 华为技术有限公司 Method and apparatus for generating target neural network model
CN110532562A (en) * 2019-08-30 2019-12-03 联想(北京)有限公司 Neural network training method, Chinese idiom misuse detection method, device and electronic equipment
CN112561050B (en) * 2019-09-25 2023-09-05 杭州海康威视数字技术股份有限公司 Neural network model training method and device
CN112561050A (en) * 2019-09-25 2021-03-26 杭州海康威视数字技术股份有限公司 Neural network model training method and device
CN110929785A (en) * 2019-11-21 2020-03-27 中国科学院深圳先进技术研究院 Data classification method and device, terminal equipment and readable storage medium
CN110929785B (en) * 2019-11-21 2023-12-05 中国科学院深圳先进技术研究院 Data classification method, device, terminal equipment and readable storage medium
WO2021098618A1 (en) * 2019-11-21 2021-05-27 中国科学院深圳先进技术研究院 Data classification method and apparatus, terminal device and readable storage medium
CN111191769B (en) * 2019-12-25 2024-03-05 中国科学院苏州纳米技术与纳米仿生研究所 Self-adaptive neural network training and reasoning device
CN111191769A (en) * 2019-12-25 2020-05-22 中国科学院苏州纳米技术与纳米仿生研究所 Self-adaptive neural network training and reasoning device
CN111310814A (en) * 2020-02-07 2020-06-19 支付宝(杭州)信息技术有限公司 Method and device for training business prediction model by utilizing unbalanced positive and negative samples
CN111429414A (en) * 2020-03-18 2020-07-17 腾讯科技(深圳)有限公司 Artificial intelligence-based focus image sample determination method and related device
CN111429414B (en) * 2020-03-18 2023-04-07 腾讯科技(深圳)有限公司 Artificial intelligence-based focus image sample determination method and related device
CN111475618A (en) * 2020-03-31 2020-07-31 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN111488985A (en) * 2020-04-08 2020-08-04 华南理工大学 Deep neural network model compression training method, device, equipment and medium
CN111488985B (en) * 2020-04-08 2023-11-14 华南理工大学 Deep neural network model compression training method, device, equipment and medium
CN111507396A (en) * 2020-04-15 2020-08-07 广州大学 Method and device for relieving error classification of neural network on unknown samples
CN111507396B (en) * 2020-04-15 2023-08-08 广州大学 Method and device for relieving error classification of unknown class samples by neural network
WO2021238586A1 (en) * 2020-05-27 2021-12-02 华为技术有限公司 Training method and apparatus, device, and computer readable storage medium
CN111680631B (en) * 2020-06-09 2023-12-22 广州视源电子科技股份有限公司 Model training method and device
CN111680631A (en) * 2020-06-09 2020-09-18 广州视源电子科技股份有限公司 Model training method and device
CN111783996B (en) * 2020-06-18 2023-08-25 杭州海康威视数字技术股份有限公司 Data processing method, device and equipment
CN111783996A (en) * 2020-06-18 2020-10-16 杭州海康威视数字技术股份有限公司 Data processing method, device and equipment
CN111861909A (en) * 2020-06-29 2020-10-30 南京理工大学 Network fine-grained image denoising and classifying method
CN111861909B (en) * 2020-06-29 2023-06-16 南京理工大学 Network fine granularity image classification method
CN112381161A (en) * 2020-11-18 2021-02-19 厦门市美亚柏科信息股份有限公司 Neural network training method
CN112381161B (en) * 2020-11-18 2022-08-30 厦门市美亚柏科信息股份有限公司 Neural network training method
CN112398862A (en) * 2020-11-18 2021-02-23 深圳供电局有限公司 Charging pile attack clustering detection method based on GRU model
CN112836816A (en) * 2021-02-04 2021-05-25 南京大学 Training method suitable for crosstalk of photoelectric storage and calculation integrated processing unit
CN112836816B (en) * 2021-02-04 2024-02-09 南京大学 Training method suitable for crosstalk of photoelectric storage and calculation integrated processing unit
WO2022174805A1 (en) * 2021-02-22 2022-08-25 上海商汤智能科技有限公司 Model training method and apparatus, image processing method and apparatus, electronic device and storage medium
CN113242547B (en) * 2021-04-02 2022-10-04 浙江大学 Method and system for filtering user behavior privacy in wireless signal based on deep learning and wireless signal receiving and transmitting device
CN113242547A (en) * 2021-04-02 2021-08-10 浙江大学 Method and system for filtering user behavior privacy in wireless signal based on deep learning and wireless signal receiving and transmitting device
CN114118413A (en) * 2021-11-30 2022-03-01 上海商汤临港智能科技有限公司 Network training and equipment control method, device, equipment and storage medium
CN114241260B (en) * 2021-12-14 2023-04-07 四川大学 Open set target detection and identification method based on deep neural network
CN114241260A (en) * 2021-12-14 2022-03-25 四川大学 Open set target detection and identification method based on deep neural network
CN116660389A (en) * 2023-07-21 2023-08-29 山东大禹水务建设集团有限公司 River sediment detection and repair system based on artificial intelligence
CN116660389B (en) * 2023-07-21 2023-10-13 山东大禹水务建设集团有限公司 River sediment detection and repair system based on artificial intelligence

Also Published As

Publication number Publication date
CN109816092B (en) 2020-06-05

Similar Documents

Publication Publication Date Title
CN109816092A (en) Deep neural network training method, device, electronic equipment and storage medium
CN108536681B (en) Intelligent question-answering method, device, equipment and storage medium based on emotion analysis
CN109817246B (en) Emotion recognition model training method, emotion recognition device, emotion recognition equipment and storage medium
WO2021155706A1 (en) Method and device for training business prediction model by using unbalanced positive and negative samples
Li et al. Identifying disaster damage images using a domain adaptation approach
CN109829358A (en) Micro- expression loan control method, device, computer equipment and storage medium
CN106295591A (en) Gender identification method based on facial image and device
CN108806792A (en) Deep learning facial diagnosis system
Zakariah et al. Sign language recognition for Arabic alphabets using transfer learning technique
CN111831826B (en) Training method, classification method and device of cross-domain text classification model
JP2022141931A (en) Method and device for training living body detection model, method and apparatus for living body detection, electronic apparatus, storage medium, and computer program
CN111444873A (en) Method and device for detecting authenticity of person in video, electronic device and storage medium
CN112767386B (en) Image aesthetic quality evaluation method and system based on theme feature and score distribution
CN115620384B (en) Model training method, fundus image prediction method and fundus image prediction device
CN111144566A (en) Neural network weight parameter training method, characteristic classification method and corresponding device
CN112418059A (en) Emotion recognition method and device, computer equipment and storage medium
CN111694954A (en) Image classification method and device and electronic equipment
Sakthimohan et al. Detection and Recognition of Face Using Deep Learning
CN109192226A (en) A kind of signal processing method and device
CN109101984A (en) A kind of image-recognizing method and device based on convolutional neural networks
CN111368524A (en) Microblog viewpoint sentence recognition method based on self-attention bidirectional GRU and SVM
EP4198906A1 (en) Image generation based on ethical viewpoints
CN110738985A (en) Cross-modal biometric feature recognition method and system based on voice signals
CN115731620A (en) Method for detecting counter attack and method for training counter attack detection model
CN109657710A (en) Data screening method, apparatus, server and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant