CN109902722A - Classifier, neural network model training method, data processing equipment and medium - Google Patents
Classifier, neural network model training method, data processing equipment and medium Download PDFInfo
- Publication number
- CN109902722A CN109902722A CN201910082386.8A CN201910082386A CN109902722A CN 109902722 A CN109902722 A CN 109902722A CN 201910082386 A CN201910082386 A CN 201910082386A CN 109902722 A CN109902722 A CN 109902722A
- Authority
- CN
- China
- Prior art keywords
- neural network
- network model
- training
- sample
- training sample
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses classifier, neural network model training method, data processing equipment and media.A kind of classifier includes: weighting loss value calculation apparatus, so that obtaining weighting loss value of the training sample in each class categories via it from the forecast confidence that neural network model exports;And parameter adjustment controls, the weighting loss value is reversely output to the parameter adjustment controls, so that its parameter for adjusting the neural network model based on weighting loss value, thus the training of the neural network model of the classifier is completed, wherein, the weighting loss value calculation apparatus adds weighted factor, weighting loss value of the training sample in each class categories is thus calculated in the forecast confidence for being used to measure training sample and the loss function of the loss cost of mark true value.By adding weighted factor when penalty values calculate, the parameter of neural network model can be adjusted, optimizes network model, improves the accuracy of identification of classifier.
Description
Technical field
The present invention relates to sorting technique more particularly to classifier, neural network model training method, data processing equipment and
Medium.
Background technique
Deep learning is currently that CRT technology field precision is outstanding, widely used main stream approach.In depth
In the application process for practising model, it is necessary first to be trained deep learning model on big training dataset, by making by oneself
Loss function and back-propagation algorithm (such as gradient decreasing function of the back-propagation algorithm as successive ignition) study of justice
The optimal neural network model parameter of iteration is deployed in practical application and carries out after then verifying touches the mark on verifying collection
Classification or prediction.
For deep learning in real data training, often will appear leads to classification essence since the quantity of training sample is unbalanced
Spend not high problem.For example, the number of the sample of some common class may have very much, such as can have for object identification
Thousands of to up to ten thousand or more, still, the sample of common class is not since the frequency of occurrences is low, is difficult to collect, their number is then very
It is few, for example, may only several hundred a samples it is even less.Since traditional loss function (such as cross entropy, mean square error) is not examined
The difference for considering the number of this different classes of training sample in reality, will lead to the general classification knot of deep learning model
Fruit is more biased towards in the more classifications of sample number, although having kept the penalty values of loss function minimum, for small sample classification
The precision of classification but still may be very low.For example, it is assumed that a training data concentration has two class sample A and B, the number of sample A
Be far longer than sample B, even if then some article to be sorted belong to sample B it is similar, but by this training dataset training after
Deep learning model classification results may more maximum probability be sample A, lead here it is the quantity due to training sample is unbalanced
The not high problem of the nicety of grading of cause.
A kind of solution of accuracy of identification for improving classifier is needed as a result,.
Summary of the invention
One of in order to solve problem above, the present invention provides a kind of classifiers, neural network model training method, data
Processing equipment and medium, the accuracy of identification to overcome the problems, such as classifier is not high, thus improves its accuracy of identification.
According to one embodiment of present invention, a kind of classifier based on neural network model is provided, comprising: weighting loss
Value calculation apparatus is configured for so that tentative prediction export from the neural network model, as neural network model
As a result, for indicating that training sample belongs to the forecast confidence of the probability of each class categories of sample, via the weighting loss
Value calculation apparatus obtains weighting loss value of the training sample in each class categories;And parameter adjustment controls, it is damaged by weighting
The weighting loss value for losing value calculation apparatus output is reversely output to the parameter adjustment controls, so that parameter adjustment dress
The parameter for adjusting the neural network model based on the weighting loss value is set, the neural network model of the classifier is thus completed
Training, wherein the weighting loss value calculation apparatus is in the forecast confidence and mark true value for being used to measure training sample
Loss cost loss function in, add weighted factor, be thus calculated the training sample in each class categories plus
Weigh penalty values.
Optionally, the weighted factor is related at least one of:
1) training data concentrates the number of the training sample of each class categories;And
2) as the tentative prediction result of the neural network model, for indicating that training sample belongs to each point of sample
The forecast confidence of the probability of class classification.
Optionally, the weighted factor and training data concentrate the number of the training sample of each class categories to be inversely proportional.
Optionally, the weighted factor with as neural network model tentative prediction result, for indicate train sample
Originally belong to the forecast confidence of the probability of each class categories of sample in monotone decreasing relationship.
Optionally, the loss function includes cross entropy loss function.
Optionally, the parameter adjustment controls are based on the weighting loss value, pass through the Back Propagation Algorithm of successive ignition
To adjust the parameter of neural network model.
According to one embodiment of present invention, a kind of neural network model training method for classifier is provided, comprising:
Training sample is input to neural network model;Neural network model output as tentative prediction result, for described in indicating
Training sample belongs to the forecast confidence of the probability of each class categories of sample;By being used to measure the described pre- of training sample
Survey confidence level and mark in the loss function of the loss cost of true value and add weighted factor, and using this be added to weighting because
The loss function of son, to calculate weighting loss value of the training sample in each class categories;And existed based on the training sample
The weighting loss value in each class categories, adjusts the parameter of the neural network model, thus completes neural network model
Training.
Optionally, the weighted factor is related at least one of:
1) training data concentrates the number of the training sample of each class categories;And
2) as the tentative prediction result of the neural network model, for indicating that training sample belongs to each point of sample
The forecast confidence of the probability of class classification.
Optionally, the weighted factor and training data concentrate the number of the training sample of each class categories to be inversely proportional.
Optionally, the weighted factor with as neural network model tentative prediction result, for indicate train sample
Originally belong to the forecast confidence of the probability of each class categories of sample in monotone decreasing relationship.
Optionally, the loss function includes cross entropy loss function.
Optionally, it is based on the weighting loss value, neural network mould is adjusted by the Back Propagation Algorithm of successive ignition
The parameter of type.
Still another embodiment in accordance with the present invention provides a kind of data processing equipment, comprising: processor;And memory,
It is stored thereon with executable code, when the executable code is executed by the processor, executes the processor above
One of method of description.
According to still another embodiment of the invention, a kind of non-transitory machinable medium is provided, is stored thereon with
Executable code makes the processor execute one of method described above when the executable code is executed by processor.
The present invention can adjust the parameter of neural network model by adding weighted factor when penalty values calculate, optimization
Neural network model, hence for the image classification for using neural network model to be classified and predicting, article identifies, image divides
It cuts, edge extracting, for the application such as speech recognition, accuracy of identification can be effectively improved.
Detailed description of the invention
Disclosure illustrative embodiments are described in more detail in conjunction with the accompanying drawings, the disclosure above-mentioned and its
Its purpose, feature and advantage will be apparent, wherein in disclosure illustrative embodiments, identical appended drawing reference
Typically represent same parts.
Fig. 1 gives the schematic block diagram of the classifier of an exemplary embodiment according to the present invention.
Fig. 2 gives the schematic diagram of the neural network model training process of an exemplary embodiment according to the present invention.
Fig. 3 gives the schematic stream of the neural network model training method of an exemplary embodiment according to the present invention
Cheng Tu.
Fig. 4 gives the schematic block diagram of the data processing equipment of an exemplary embodiment according to the present invention.
Specific embodiment
The preferred embodiment of the disclosure is more fully described below with reference to accompanying drawings.Although showing the disclosure in attached drawing
Preferred embodiment, however, it is to be appreciated that may be realized in various forms the disclosure without the embodiment party that should be illustrated here
Formula is limited.On the contrary, these embodiments are provided so that this disclosure will be more thorough and complete, and can be by the disclosure
Range is completely communicated to those skilled in the art.What needs to be explained here is that number, serial number and attached drawing in the application
Mark it is merely for convenience description and occur, for step of the invention, sequence etc. be not limited in any way, unless
The execution that step has been explicitly pointed out in specification has specific sequencing.
It is n output node of setting that neural network classification model, which solves more most common methods of classification problem, and wherein n is class
Other number.For each sample, via the available n dimension group of neural network as output result.In array
The output result of the corresponding classification of each dimension (namely each output node).In the ideal case, if one
Sample belongs to classification k, then the output valve of output node corresponding to this classification should be 1, and the output of other nodes is all
It is 0.In practical applications, usually output result can not reach 1, but a probability numbers, such as 0.5,0.3 etc., this
Kind situation usually takes final classification result of that the corresponding classification of the output node of maximum probability as this sample.
As described in the background art, the problem low presently, there are the nicety of grading of classifier, such as due to number of training
The low problem of nicety of grading caused by mesh is unbalanced, More specifically, because the number of the sample of some or certain classifications is big,
The output probability value of corresponding output node may also can be big, the classification results caused are partial to the big sample class of number
Not, lead to the type that the small sample of number possibly can not be recognized accurately.In addition, there is also the accuracy of identification of difficult sample is low
Problem.
One of to solve the above-mentioned problems, the present invention proposes a kind of neural network model of the use based on Weighted Loss Function
Training method and corresponding classifier.
Firstly, Fig. 1 gives the schematic block diagram of the classifier of an exemplary embodiment according to the present invention, Fig. 2 is given
The schematic diagram of the neural network model training process of an exemplary embodiment according to the present invention is gone out.
Specifically, as shown in Figure 1, the classification based on neural network model of an exemplary embodiment according to the present invention
Device 100 includes weighting loss value calculation apparatus 110 and parameter adjustment controls 120.
Wherein, weighting loss value calculation apparatus 110 can be configured for, so that from the neural network mould of classifier 100
Type output, tentative prediction result as neural network model, for indicating that training sample belongs to each classification class of sample
The not forecast confidence of the probability of (as shown in Fig. 2, classifier shares K output node, so sharing K class categories), warp
By the weighting loss value calculation apparatus 110, weighting loss value of the training sample in each class categories is obtained (such as Fig. 2 institute
Show).
Then, the weighting loss value exported by weighting loss value calculation apparatus 110 is reversely output to parameter adjustment controls
120, so that parameter adjustment controls 120 adjust the ginseng of each node in the neural network model based on these weighting loss values
Number, thus completes the training of the neural network model of the classifier.
Wherein, above-mentioned weighting loss value calculation apparatus 110 is in the forecast confidence and mark for being used to measure training sample
In the loss function of the loss cost of true value, weighted factor is added, the training sample is thus calculated in each class categories
On weighting loss value.
Here, " forecast confidence " of training sample is exported from neural network model for the preliminary of the training sample
Prediction result, for indicating that the training sample belongs to the probability of each class categories.For example, for a 3 class classifiers, it is initial
One training sample of Neural Network model predictive belong to the 1st class categories probability be 0.3, belong to the general of the 2nd class categories
Rate is 0.5, and the probability for belonging to the 3rd class categories is 0.2, then 0.3,0.5 and 0.2 here is respectively that the training sample belongs to the
The forecast confidence of 1 class, the 2nd class and the 3rd class.
On the other hand, " mark true value " of training sample is it is meant that when the training sample belongs to certain class categories, the instruction
Practicing mark true value of the sample in the class categories should be " 1 ", when the training sample is not belonging to certain class categories, the instruction
Practicing mark true value of the sample in the class categories should be " 0 ".
The present invention can adjust the parameter of neural network model by adding weighted factor when penalty values calculate, optimization
Neural network model, hence for the image classification for using neural network model to be classified and predicting, article identifies, image divides
It cuts, edge extracting, for the application such as speech recognition, accuracy of identification can be effectively improved.
Fig. 3 gives the schematic stream of the neural network model training method of an exemplary embodiment according to the present invention
Cheng Tu.
As shown in figure 3, training sample is input to neural network model, then neural network model is defeated in step S101
Out as tentative prediction result, for indicating that the training sample belongs to the forecast confidence of the probability of each class categories of sample
(can refer to shown in the above-mentioned example and Fig. 2).
Then, in step S110, by the forecast confidence and mark true value for being used to measure training sample
It loses in the loss function of cost and adds weighted factor, and be added to the loss function of weighted factor using this, to calculate this
Weighting loss value of the training sample in each class categories.
Then, in step S120, weighting loss value based on the training sample in each class categories adjusts the nerve
Thus the parameter of network model completes the training (as shown in Figure 2) of neural network model.
Here it is possible to it is based on weighting loss value, by the Back Propagation Algorithm (such as gradient descent method etc.) of successive ignition,
To adjust the parameter of neural network model.
The present invention can adjust the parameter of neural network model by adding weighted factor when penalty values calculate, optimization
Neural network model, hence for the image classification for using neural network model to be classified and predicting, article identifies, image divides
It cuts, edge extracting, for the application such as speech recognition, accuracy of identification can be effectively improved.
In illustrative embodiments above, weighting function may include intersecting entropy function.
For the disaggregated model of a M class, traditional cross entropy loss function is defined as:
Wherein, pi(xk) indicate training sample xkBelong to the true probability (i.e. above-mentioned " mark true value ") of the i-th class,
Value is usually 0 or 1.For example, as described above, if training sample xkBelong to certain class, then the probability is in the value of the category
1, it is otherwise 0.
Formula 1) in qi(xk) indicate prediction training sample xkBelong to the confidence level probability (i.e. above-mentioned " prediction of the i-th class
Confidence level "), it is exactly the output result of the output node corresponding with the i-th class sample of classifier.As described above, M output section
Point can export M confidence level probability, for indicating the probability of a classification that a sample is belonging respectively in this M classification.
As described above, usually taking that classification of confidence level maximum probability as classification results in classification.
In the present invention, above-mentioned weighted factor can be related at least one of:
1) training data concentrates the number of the training sample of each class categories;And
2) as the tentative prediction result of the neural network model, for indicating that training sample belongs to each point of sample
The forecast confidence of the probability of class classification.
Assuming that there is one, containing K training sample, (here, sample can be denoted as xk, wherein k=1,2 ..., K) data
Collect XK, wherein XK={ x1,x2,……,xK}。
It is different from above-mentioned traditional entropy function that intersects for loss function, in the present invention, it is possible, firstly, to using with class
The sample number weight w that the number of the sample of other i is inversely proportionali, to adjust the penalty values of loss function, thus by sample number weight shape
At Weighted Loss Function it is as follows.
As described above, sample number weight wiIt is inversely proportional with the number of the sample of classification i, it can be for example with following form
Alternatively, for example, following form
Here, niIt is the sample size of the i-th class training sample, N is the quantity of all training samples, and β is appointing greater than 0
Meaning real number, it is preferable that β is positive integer (calculating simpler).
As described above, an exemplary embodiment according to the present invention, defines the number of one with the sample of classification i first
The weighted factor w being inversely proportionali(it is properly termed as " sample number weight wi"), then reversely adjusted according to the number size of Different categories of samples
Whole penalty values, the i.e. penalty values of the training data of the few classification of increase sample size, inhibit the training of the classification more than sample size
The penalty values of data, so that neural network model can obtain better nicety of grading on equalization data.
That is, the problem that the present exemplary embodiment is low for number of samples nicety of grading caused by uneven, proposes use
Sample number weight wiCarry out regulation loss value, withFor, when the sample number very little of a certain classification, weight wiValue
Level off to 1, the penalty values result of the category is basically unchanged;But when there are many number of samples of a certain classification, weight wiValue become
It is bordering on 0, the penalty values of the category can all reduce.In this way, the type more than number of samples can be greatly reduced by sample number weight
Sample loss contribution, thus, it is possible to solve to a certain extent number of samples it is uneven caused by nicety of grading is low asks
Topic.
In summary, by utilizing above-mentioned " sample number weight wi" adjust Different categories of samples penalty values, i.e., increase sample
The penalty values of the training data of the few classification of quantity, while inhibiting the penalty values of the training data of classification more than sample size, make
Higher nicety of grading can be obtained by obtaining classification method according to the present invention.
The weight of another exemplary embodiment according to the present invention, penalty values can also be according to the prediction confidence of the i-th class
Spend qi(xk) be adjusted, at this time can the weight be referred to as confidence weight ui(weight uiWith the forecast confidence q of the i-th classi
(xk) related), so that for forecast confidence qi(xk) bad sample class, it gives higher penalty values weight and otherwise gives
Lower weight is given, so that classification method according to the present invention or classifier can obtain higher nicety of grading.
As a result, by confidence weight uiThe Weighted Loss Function of formation is as follows.
Wherein, uiFor above-mentioned confidence weight.
As described above, confidence weight uiWith the forecast confidence q of the i-th classi(xk) related, specifically, for example, confidence level
The size of weight can be in the relationship of monotone decreasing with forecast confidence, so as to for its forecast confidence qi(xk) bad sample
This classification gives higher penalty values weight, otherwise, gives lower penalty values weight.
Thus, for example, confidence weight uiIt can be defined as follows.
ui=(1-qi(xk))α (7)
Here, α is any real number greater than 0, it is preferable that α is positive integer (calculating simpler).
The present exemplary embodiment can solve the problem of difficult sample, and difficult sample refers to being difficult to accurately differentiate in network
Sample.For this kind of sample, the present invention passes through confidence weight uiTo increase the contribution of its loss function.In weighting loss value
Calculation formula in, qi(xk) refer to the confidence level of neural network forecast.For Mr. Yu's class sample, the confidence level q of neural network forecasti
(xk) higher, uiSmaller, the loss weight contribution of the sample is smaller, thus, it is possible to inhibit the weight contribution of the sample, solves difficult
The problem of sample is difficult to.
By adjusting the weight of penalty values based on the forecast confidence exported on the output node of classifier, so that right
In the bad sample class of forecast confidence, higher penalty values weight is given, lower weight is otherwise given, so that according to this
The classification method or classifier of invention can obtain higher nicety of grading.
The weight of further exemplary embodiment according to the present invention, penalty values (can be referred to as " comprehensive weight hi")
It can be adjusted according to both the number of Different categories of samples and forecast confidence of each output node output.That is, this implementation
Example can be regarded as the combination of above-mentioned two embodiment.
Specifically, comprehensive weight hiIt can be by wiAnd uiThe two is combined into.
For example, comprehensive weight hiIt can be made of following formula (8).
hi=(wi)s(ui)t(8),
Here, s and t all can be greater than 0 any real number, it is preferable that be positive integer.
As a result, by comprehensive weight hiThe weighting function of formation is as follows.
The forecast confidence exported on output node by number and classifier in conjunction with Different categories of samples loses to adjust
The weight of value, so that sample class unbalanced for sample size and that forecast confidence is bad, gives higher penalty values
Otherwise weight gives lower weight, classification method according to the present invention or classifier is enabled to obtain higher classification essence
Degree.
Fig. 4 gives the schematic block diagram of the data processing equipment of an exemplary embodiment according to the present invention.
Referring to fig. 4, which includes memory 10 and processor 20.
Processor 20 can be the processor of a multicore, also may include multiple processors.In some embodiments, locate
Reason device 20 may include a general primary processor and one or more special coprocessors, such as graphics processor
(GPU), digital signal processor (DSP) etc..In some embodiments, the circuit realization of customization can be used in processor 20,
Such as application-specific IC (ASIC, Application Specific Integrated Circuit) or scene can
Programmed logic gate array (FPGA, Field Programmable Gate Arrays).
It is stored with executable code on memory 10, when the executable code is executed by the processor 20, makes institute
It states processor 20 and executes one of data migration method described above.Wherein, memory 10 may include various types of storages
Unit, such as Installed System Memory, read-only memory (ROM) and permanent storage.Wherein, ROM can store processor 20 or
The static data or instruction that other modules of computer need.Permanent storage can be read-write storage device.Forever
Long storage device will not lose the non-volatile memory device of the instruction and data of storage can be after computer circuit breaking.
In some embodiments, permanent storage device is using mass storage device (such as magnetically or optically disk, flash memory) as permanent
Storage device.In other embodiment, permanent storage device can be removable storage equipment (such as floppy disk, light
It drives).Installed System Memory can be read-write storage equipment or the read-write storage equipment of volatibility, such as in dynamic random-access
It deposits.Installed System Memory can store the instruction and data that some or all processors need at runtime.In addition, memory 10 can
To include the combination of any computer readable storage medium, including various types of semiconductor memory chips (DRAM, SRAM,
SDRAM, flash memory, programmable read only memory), disk and/or CD can also use.In some embodiments, memory
10 may include removable storage equipment that is readable and/or writing, such as laser disc (CD), read-only digital versatile disc
(such as DVD-ROM, DVD-dual layer-ROM), read-only Blu-ray Disc, super disc density, flash card (such as SD card, min SD card,
Micro-SD card etc.), magnetic floppy disc etc..Computer readable storage medium does not include carrier wave and by wirelessly or non-wirelessly transmitting
Momentary electron signal.
In addition, being also implemented as a kind of computer program or computer program product, the meter according to the method for the present invention
Calculation machine program or computer program product include the calculating for executing the above steps limited in the above method of the invention
Machine program code instruction.
Alternatively, the present invention can also be embodied as a kind of (or the computer-readable storage of non-transitory machinable medium
Medium or machine readable storage medium), it is stored thereon with executable code (or computer program or computer instruction code),
When the executable code (or computer program or computer instruction code) by electronic equipment (or calculate equipment, server
Deng) processor execute when, so that the processor is executed each step according to the above method of the present invention.
Those skilled in the art will also understand is that, various illustrative logical blocks, mould in conjunction with described in disclosure herein
Block, circuit and algorithm steps may be implemented as the combination of electronic hardware, computer software or both.
What flow chart and block diagram in attached drawing etc. showed the system and method for multiple embodiments according to the present invention can
The architecture, function and operation being able to achieve.In this regard, each box in flowchart or block diagram can represent a mould
A part of block, program segment or code, a part of the module, section or code include one or more for realizing rule
The executable instruction of fixed logic function.It should also be noted that in some implementations as replacements, the function of being marked in box
It can also be occurred with being different from the sequence marked in attached drawing.For example, two continuous boxes can actually be substantially in parallel
It executes, they can also be executed in the opposite order sometimes, and this depends on the function involved.It is also noted that block diagram and/
Or the combination of each box in flow chart and the box in block diagram and or flow chart, can with execute as defined in function or
The dedicated hardware based system of operation is realized, or can be realized using a combination of dedicated hardware and computer instructions.
Various embodiments of the present invention are described above, above description is exemplary, and non-exclusive, and
It is not limited to disclosed each embodiment.Without departing from the scope and spirit of illustrated each embodiment, for this skill
Many modifications and changes are obvious for the those of ordinary skill in art field.The selection of term used herein, purport
In the principle, practical application or improvement to the technology in market for best explaining each embodiment, or make the art
Other those of ordinary skill can understand each embodiment disclosed herein.
Claims (14)
1. a kind of classifier based on neural network model characterized by comprising
Weighting loss value calculation apparatus, be configured for so that from the neural network model export, as neural network mould
The tentative prediction result of type, for indicating that training sample belongs to the forecast confidence of the probability of each class categories of sample, pass through
By the weighting loss value calculation apparatus, weighting loss value of the training sample in each class categories is obtained;And
Parameter adjustment controls are output to parameter adjustment dress by the weighting loss value that weighting loss value calculation apparatus exports
It sets, so that the parameter adjustment controls adjust the parameter of the neural network model based on the weighting loss value, thus completes
The training of the neural network model of the classifier;
Wherein, the weighting loss value calculation apparatus is in the forecast confidence and mark true value for being used to measure training sample
In the loss function for losing cost, weighted factor is added, weighting of the training sample in each class categories is thus calculated
Penalty values.
2. classifier as described in claim 1, which is characterized in that wherein, the weighted factor is related at least one of:
1) training data concentrates the number of the training sample of each class categories;And
2) as the tentative prediction result of the neural network model, for indicating that training sample belongs to each classification class of sample
The forecast confidence of other probability.
3. classifier as claimed in claim 2, which is characterized in that wherein, the weighted factor and training data concentrate each point
The number of the training sample of class classification is in inversely prroportional relationship.
4. classifier as claimed in claim 2, which is characterized in that wherein, the weighted factor with as neural network model
Tentative prediction result, for indicating that training sample belongs to the forecast confidence of the probability of each class categories of sample in dullness
Successively decrease relationship.
5. the classifier as described in any one in Claims 1 to 4, which is characterized in that the loss function includes intersecting
Entropy loss function.
6. the classifier as described in any one in Claims 1 to 4, which is characterized in that the parameter adjustment controls are based on
The weighting loss value, the parameter of neural network model is adjusted by the Back Propagation Algorithm of successive ignition.
7. a kind of neural network model training method for classifier characterized by comprising
Training sample is input to neural network model;
Neural network model export as tentative prediction result, for indicating that the training sample belongs to each classification class of sample
The forecast confidence of other probability;
By in the loss function of the loss cost of the forecast confidence and mark true value that are used to measure training sample
Weighted factor is added, and is added to the loss function of weighted factor using this, to calculate the training sample in each class categories
On weighting loss value;And
The weighting loss value based on the training sample in each class categories, adjusts the parameter of the neural network model,
Thus the training of neural network model is completed.
8. neural network model training method as claimed in claim 7, which is characterized in that wherein, the weighted factor with
It is at least one lower related:
1) training data concentrates the number of the training sample of each class categories;And
2) as the tentative prediction result of the neural network model, for indicating that training sample belongs to each classification class of sample
The forecast confidence of other probability.
9. neural network model training method as claimed in claim 8, which is characterized in that wherein, the weighted factor and instruction
The number for practicing the training sample of each class categories in data set is inversely proportional.
10. neural network model training method as claimed in claim 8, which is characterized in that wherein, the weighted factor and work
For neural network model tentative prediction result, for indicating that training sample belongs to the pre- of the probability of each class categories of sample
Surveying confidence level is in monotone decreasing relationship.
11. the neural network model training method as described in any one in claim 7~10, which is characterized in that described
Loss function includes cross entropy loss function.
12. the neural network model training method as described in any one in claim 7~10, which is characterized in that be based on
The weighting loss value, the parameter of neural network model is adjusted by the Back Propagation Algorithm of successive ignition.
13. a kind of data processing equipment, comprising:
Processor;And
Memory is stored thereon with executable code, when the executable code is executed by the processor, makes the processing
Device executes the method as described in any one of claim 7~12.
14. a kind of non-transitory machinable medium, is stored thereon with executable code, when the executable code is located
When managing device execution, the processor is made to execute the method as described in any one of claim 7~12.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910082386.8A CN109902722A (en) | 2019-01-28 | 2019-01-28 | Classifier, neural network model training method, data processing equipment and medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910082386.8A CN109902722A (en) | 2019-01-28 | 2019-01-28 | Classifier, neural network model training method, data processing equipment and medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109902722A true CN109902722A (en) | 2019-06-18 |
Family
ID=66944351
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910082386.8A Pending CN109902722A (en) | 2019-01-28 | 2019-01-28 | Classifier, neural network model training method, data processing equipment and medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109902722A (en) |
Cited By (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110567967A (en) * | 2019-08-20 | 2019-12-13 | 武汉精立电子技术有限公司 | Display panel detection method, system, terminal device and computer readable medium |
CN110717515A (en) * | 2019-09-06 | 2020-01-21 | 北京三快在线科技有限公司 | Model training method and device and electronic equipment |
CN110751197A (en) * | 2019-10-14 | 2020-02-04 | 上海眼控科技股份有限公司 | Picture classification method, picture model training method and equipment |
CN111127892A (en) * | 2019-12-27 | 2020-05-08 | 北京易华录信息技术股份有限公司 | Intersection timing parameter optimization model construction and intersection signal optimization method |
CN111260665A (en) * | 2020-01-17 | 2020-06-09 | 北京达佳互联信息技术有限公司 | Image segmentation model training method and device |
CN111291823A (en) * | 2020-02-24 | 2020-06-16 | 腾讯科技(深圳)有限公司 | Fusion method and device of classification models, electronic equipment and storage medium |
CN111310814A (en) * | 2020-02-07 | 2020-06-19 | 支付宝(杭州)信息技术有限公司 | Method and device for training business prediction model by utilizing unbalanced positive and negative samples |
CN111326148A (en) * | 2020-01-19 | 2020-06-23 | 北京世纪好未来教育科技有限公司 | Confidence correction and model training method, device, equipment and storage medium thereof |
CN111341459A (en) * | 2020-02-28 | 2020-06-26 | 上海交通大学医学院附属上海儿童医学中心 | Training method of classified deep neural network model and genetic disease detection method |
CN111507396A (en) * | 2020-04-15 | 2020-08-07 | 广州大学 | Method and device for relieving error classification of neural network on unknown samples |
CN111737429A (en) * | 2020-06-16 | 2020-10-02 | 平安科技(深圳)有限公司 | Training method, AI interview method and related equipment |
CN111930980A (en) * | 2020-08-21 | 2020-11-13 | 深圳市升幂科技有限公司 | Training method of image retrieval model, image retrieval method, device and medium |
CN112052900A (en) * | 2020-09-04 | 2020-12-08 | 京东数字科技控股股份有限公司 | Machine learning sample weight adjusting method and device and storage medium |
CN112182214A (en) * | 2020-09-27 | 2021-01-05 | 中国建设银行股份有限公司 | Data classification method, device, equipment and medium |
CN112232368A (en) * | 2020-09-10 | 2021-01-15 | 浙江大华技术股份有限公司 | Target recognition model training method, target recognition method and related device thereof |
CN112348161A (en) * | 2019-08-09 | 2021-02-09 | 北京地平线机器人技术研发有限公司 | Neural network training method, neural network training device and electronic equipment |
CN112465042A (en) * | 2020-12-02 | 2021-03-09 | 中国联合网络通信集团有限公司 | Generation method and device of classification network model |
CN112575513A (en) * | 2020-12-01 | 2021-03-30 | 珠海格力电器股份有限公司 | Defoaming control method and device for washing machine |
CN112784953A (en) * | 2019-11-07 | 2021-05-11 | 佳能株式会社 | Training method and device of object recognition model |
CN112801237A (en) * | 2021-04-15 | 2021-05-14 | 北京远鉴信息技术有限公司 | Training method and device for violence and terrorism content recognition model and readable storage medium |
CN112906434A (en) * | 2019-12-03 | 2021-06-04 | 富士通株式会社 | Information processing apparatus, information processing method, and computer program |
CN113011532A (en) * | 2021-04-30 | 2021-06-22 | 平安科技(深圳)有限公司 | Classification model training method and device, computing equipment and storage medium |
CN113066069A (en) * | 2021-03-31 | 2021-07-02 | 深圳中科飞测科技股份有限公司 | Adjusting method and device, adjusting equipment and storage medium |
CN113128539A (en) * | 2019-12-31 | 2021-07-16 | 财团法人工业技术研究院 | Object detection model training method and system based on adaptive labeling |
CN113378853A (en) * | 2021-05-25 | 2021-09-10 | 北京有竹居网络技术有限公司 | Image classification method, device, equipment and storage medium |
CN113392889A (en) * | 2021-06-07 | 2021-09-14 | 深圳市欢太数字科技有限公司 | Data processing method and device and electronic equipment |
CN113485848A (en) * | 2021-09-08 | 2021-10-08 | 深圳思谋信息科技有限公司 | Deep neural network deployment method and device, computer equipment and storage medium |
CN113535964A (en) * | 2021-09-15 | 2021-10-22 | 深圳前海环融联易信息科技服务有限公司 | Enterprise classification model intelligent construction method, device, equipment and medium |
CN113780473A (en) * | 2021-09-30 | 2021-12-10 | 平安科技(深圳)有限公司 | Data processing method and device based on depth model, electronic equipment and storage medium |
CN113901223A (en) * | 2021-11-19 | 2022-01-07 | 企查查科技有限公司 | Method and device for generating enterprise classification model, computer equipment and storage medium |
CN115496200A (en) * | 2022-09-05 | 2022-12-20 | 中国科学院半导体研究所 | Neural network quantitative model training method, device and equipment |
WO2023230748A1 (en) * | 2022-05-30 | 2023-12-07 | Nvidia Corporation | Dynamic class weighting for training one or more neural networks |
-
2019
- 2019-01-28 CN CN201910082386.8A patent/CN109902722A/en active Pending
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112348161A (en) * | 2019-08-09 | 2021-02-09 | 北京地平线机器人技术研发有限公司 | Neural network training method, neural network training device and electronic equipment |
CN110567967A (en) * | 2019-08-20 | 2019-12-13 | 武汉精立电子技术有限公司 | Display panel detection method, system, terminal device and computer readable medium |
CN110717515A (en) * | 2019-09-06 | 2020-01-21 | 北京三快在线科技有限公司 | Model training method and device and electronic equipment |
CN110751197A (en) * | 2019-10-14 | 2020-02-04 | 上海眼控科技股份有限公司 | Picture classification method, picture model training method and equipment |
CN112784953A (en) * | 2019-11-07 | 2021-05-11 | 佳能株式会社 | Training method and device of object recognition model |
CN112906434A (en) * | 2019-12-03 | 2021-06-04 | 富士通株式会社 | Information processing apparatus, information processing method, and computer program |
CN111127892A (en) * | 2019-12-27 | 2020-05-08 | 北京易华录信息技术股份有限公司 | Intersection timing parameter optimization model construction and intersection signal optimization method |
CN113128539A (en) * | 2019-12-31 | 2021-07-16 | 财团法人工业技术研究院 | Object detection model training method and system based on adaptive labeling |
CN113128539B (en) * | 2019-12-31 | 2024-06-18 | 财团法人工业技术研究院 | Object detection model training method and system based on adaptive annotation |
CN111260665A (en) * | 2020-01-17 | 2020-06-09 | 北京达佳互联信息技术有限公司 | Image segmentation model training method and device |
CN111260665B (en) * | 2020-01-17 | 2022-01-21 | 北京达佳互联信息技术有限公司 | Image segmentation model training method and device |
CN111326148A (en) * | 2020-01-19 | 2020-06-23 | 北京世纪好未来教育科技有限公司 | Confidence correction and model training method, device, equipment and storage medium thereof |
CN111326148B (en) * | 2020-01-19 | 2021-02-23 | 北京世纪好未来教育科技有限公司 | Confidence correction and model training method, device, equipment and storage medium thereof |
CN111310814A (en) * | 2020-02-07 | 2020-06-19 | 支付宝(杭州)信息技术有限公司 | Method and device for training business prediction model by utilizing unbalanced positive and negative samples |
CN111291823A (en) * | 2020-02-24 | 2020-06-16 | 腾讯科技(深圳)有限公司 | Fusion method and device of classification models, electronic equipment and storage medium |
CN111291823B (en) * | 2020-02-24 | 2023-08-18 | 腾讯科技(深圳)有限公司 | Fusion method and device of classification model, electronic equipment and storage medium |
CN111341459A (en) * | 2020-02-28 | 2020-06-26 | 上海交通大学医学院附属上海儿童医学中心 | Training method of classified deep neural network model and genetic disease detection method |
CN111507396B (en) * | 2020-04-15 | 2023-08-08 | 广州大学 | Method and device for relieving error classification of unknown class samples by neural network |
CN111507396A (en) * | 2020-04-15 | 2020-08-07 | 广州大学 | Method and device for relieving error classification of neural network on unknown samples |
CN111737429B (en) * | 2020-06-16 | 2023-11-03 | 平安科技(深圳)有限公司 | Training method, AI interview method and related equipment |
CN111737429A (en) * | 2020-06-16 | 2020-10-02 | 平安科技(深圳)有限公司 | Training method, AI interview method and related equipment |
CN111930980A (en) * | 2020-08-21 | 2020-11-13 | 深圳市升幂科技有限公司 | Training method of image retrieval model, image retrieval method, device and medium |
CN112052900B (en) * | 2020-09-04 | 2024-05-24 | 京东科技控股股份有限公司 | Machine learning sample weight adjustment method and device, and storage medium |
CN112052900A (en) * | 2020-09-04 | 2020-12-08 | 京东数字科技控股股份有限公司 | Machine learning sample weight adjusting method and device and storage medium |
CN112232368A (en) * | 2020-09-10 | 2021-01-15 | 浙江大华技术股份有限公司 | Target recognition model training method, target recognition method and related device thereof |
CN112232368B (en) * | 2020-09-10 | 2023-09-01 | 浙江大华技术股份有限公司 | Target recognition model training method, target recognition method and related devices thereof |
CN112182214A (en) * | 2020-09-27 | 2021-01-05 | 中国建设银行股份有限公司 | Data classification method, device, equipment and medium |
CN112182214B (en) * | 2020-09-27 | 2024-03-19 | 中国建设银行股份有限公司 | Data classification method, device, equipment and medium |
CN112575513A (en) * | 2020-12-01 | 2021-03-30 | 珠海格力电器股份有限公司 | Defoaming control method and device for washing machine |
CN112465042B (en) * | 2020-12-02 | 2023-10-24 | 中国联合网络通信集团有限公司 | Method and device for generating classified network model |
CN112465042A (en) * | 2020-12-02 | 2021-03-09 | 中国联合网络通信集团有限公司 | Generation method and device of classification network model |
CN113066069A (en) * | 2021-03-31 | 2021-07-02 | 深圳中科飞测科技股份有限公司 | Adjusting method and device, adjusting equipment and storage medium |
CN112801237A (en) * | 2021-04-15 | 2021-05-14 | 北京远鉴信息技术有限公司 | Training method and device for violence and terrorism content recognition model and readable storage medium |
CN112801237B (en) * | 2021-04-15 | 2021-07-23 | 北京远鉴信息技术有限公司 | Training method and device for violence and terrorism content recognition model and readable storage medium |
CN113011532A (en) * | 2021-04-30 | 2021-06-22 | 平安科技(深圳)有限公司 | Classification model training method and device, computing equipment and storage medium |
CN113378853B (en) * | 2021-05-25 | 2024-02-13 | 北京有竹居网络技术有限公司 | Image classification method, device, equipment and storage medium |
CN113378853A (en) * | 2021-05-25 | 2021-09-10 | 北京有竹居网络技术有限公司 | Image classification method, device, equipment and storage medium |
CN113392889A (en) * | 2021-06-07 | 2021-09-14 | 深圳市欢太数字科技有限公司 | Data processing method and device and electronic equipment |
CN113485848A (en) * | 2021-09-08 | 2021-10-08 | 深圳思谋信息科技有限公司 | Deep neural network deployment method and device, computer equipment and storage medium |
CN113485848B (en) * | 2021-09-08 | 2021-12-17 | 深圳思谋信息科技有限公司 | Deep neural network deployment method and device, computer equipment and storage medium |
WO2023039925A1 (en) * | 2021-09-15 | 2023-03-23 | 深圳前海环融联易信息科技服务有限公司 | Intelligent construction method and apparatus for enterprise classification model, and device and medium |
CN113535964A (en) * | 2021-09-15 | 2021-10-22 | 深圳前海环融联易信息科技服务有限公司 | Enterprise classification model intelligent construction method, device, equipment and medium |
CN113535964B (en) * | 2021-09-15 | 2021-12-24 | 深圳前海环融联易信息科技服务有限公司 | Enterprise classification model intelligent construction method, device, equipment and medium |
CN113780473A (en) * | 2021-09-30 | 2021-12-10 | 平安科技(深圳)有限公司 | Data processing method and device based on depth model, electronic equipment and storage medium |
CN113780473B (en) * | 2021-09-30 | 2023-07-14 | 平安科技(深圳)有限公司 | Depth model-based data processing method and device, electronic equipment and storage medium |
CN113901223B (en) * | 2021-11-19 | 2024-01-26 | 企查查科技股份有限公司 | Method, device, computer equipment and storage medium for generating enterprise classification model |
CN113901223A (en) * | 2021-11-19 | 2022-01-07 | 企查查科技有限公司 | Method and device for generating enterprise classification model, computer equipment and storage medium |
WO2023230748A1 (en) * | 2022-05-30 | 2023-12-07 | Nvidia Corporation | Dynamic class weighting for training one or more neural networks |
CN115496200B (en) * | 2022-09-05 | 2023-09-22 | 中国科学院半导体研究所 | Neural network quantization model training method, device and equipment |
CN115496200A (en) * | 2022-09-05 | 2022-12-20 | 中国科学院半导体研究所 | Neural network quantitative model training method, device and equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109902722A (en) | Classifier, neural network model training method, data processing equipment and medium | |
CN102521656B (en) | Integrated transfer learning method for classification of unbalance samples | |
WO2019179403A1 (en) | Fraud transaction detection method based on sequence width depth learning | |
CN109086799A (en) | A kind of crop leaf disease recognition method based on improvement convolutional neural networks model AlexNet | |
CN110298321A (en) | Route denial information extraction based on deep learning image classification | |
CN107644057B (en) | Absolute imbalance text classification method based on transfer learning | |
Sinha et al. | Class-wise difficulty-balanced loss for solving class-imbalance | |
CN111091201A (en) | Data partition mixed sampling-based unbalanced integrated classification method | |
CN109871443A (en) | A kind of short text classification method and device based on book keeping operation scene | |
CN109871855A (en) | A kind of adaptive depth Multiple Kernel Learning method | |
CN109491914A (en) | Defect report prediction technique is influenced based on uneven learning strategy height | |
CN110414780A (en) | A kind of financial transaction negative sample generation method based on generation confrontation network | |
CN109948680A (en) | The classification method and system of medical record data | |
CN107944460A (en) | One kind is applied to class imbalance sorting technique in bioinformatics | |
CN112966109B (en) | Multi-level Chinese text classification method and system | |
CN109376179A (en) | A kind of sample equilibrating method in data mining | |
CN110415260A (en) | Smog image segmentation and recognition methods based on dictionary and BP neural network | |
CN109493333A (en) | Ultrasonic Calcification in Thyroid Node point extraction algorithm based on convolutional neural networks | |
CN115148299A (en) | XGboost-based ore deposit type identification method and system | |
CN109543693A (en) | Weak labeling data noise reduction method based on regularization label propagation | |
CN115439718A (en) | Industrial detection method, system and storage medium combining supervised learning and feature matching technology | |
CN111144462A (en) | Unknown individual identification method and device for radar signals | |
Zhou et al. | Credit card fraud identification based on principal component analysis and improved AdaBoost algorithm | |
CN105335763A (en) | Fabric defect classification method based on improved extreme learning machine | |
CN108629680A (en) | A kind of Risk Identification Method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190618 |