CN108664924A - A kind of multi-tag object identification method based on convolutional neural networks - Google Patents
A kind of multi-tag object identification method based on convolutional neural networks Download PDFInfo
- Publication number
- CN108664924A CN108664924A CN201810443651.6A CN201810443651A CN108664924A CN 108664924 A CN108664924 A CN 108664924A CN 201810443651 A CN201810443651 A CN 201810443651A CN 108664924 A CN108664924 A CN 108664924A
- Authority
- CN
- China
- Prior art keywords
- convolutional neural
- neural networks
- tag
- network
- object identification
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a kind of multi-tag object identification method based on convolutional neural networks, this method utilizes the inclusion relation between label, and according to including the CNN characteristic extraction parts for sequentially building each label successively, pass through the constantly abstract each layer feature of extraction of convolution algorithm operation, and the grader of each label is respectively set in the different depth of network, the characteristic pattern that the CNN characteristic extraction parts of respective labels are extracted, it is input to corresponding grader, simultaneously error back propagation is carried out using multiple graders, the network weight parameter of the corresponding each layer of supervised training, each label classification is finally obtained to complete to identify.The multi-tag convolutional neural networks that the present invention uses, can well solve the fusion of information between multiple labels, solve the problems, such as that traditional multi-tag object identification accuracy rate is not high, while improve the efficiency of training and identification.
Description
Technical field
The invention belongs to technical field of image processing, more particularly to a kind of multi-tag object based on convolutional neural networks is known
Other method.
Background technology
With the rapid development of artificial intelligence technology, automatic object identification has become the hot spot studied both at home and abroad in recent years
Problem is with a wide range of applications in fields such as intelligent monitoring, telemetering remote sensing, robot, Medical Image Processings.Really giving birth to
In work, kind of object is various, and individual similarity is higher, the mankind can by visual perception shape, color and distance etc. information, and
And these comprehensive information, accurately to judge object generic, but this is often relatively difficult for computer.Cause
How this, make computer possess similar with the mankind, the even more than mankind recognition capabilities, it has also become current object identifies important
Direction and significant challenge.
The main approaches of object identification are that extraction object features realize Object representation, then utilize certain engineering
The study that algorithm carries out object type is practised, is finally classified to object, realizes object identification.But target to be identified exists
The high situation of similarity, the feature extracted in this case are generally difficult to reflect class inherited and class difference greatly, between class in class
Interior general character.Especially in the case where object has multiple labels, due to the limitation of traditional characteristic extracting mode, different labels are made
It is often no associated between feature, it is more difficult that this makes object identification get up.
Multi-tag object identification problem relies primarily on two kinds of method to handle at present.One is based on traditional images
The method of processing extracts the features such as some SIFT, HOG, SURF to subject image first, in conjunction with optimization BoW models and melt
Hop algorithm converts feature, and traditional grader (such as SVM) is finally used to obtain the prediction result of object category, the party
The recognition speed of method is than very fast, but recognition accuracy is relatively low.
Another kind is the method based on deep learning, and this method is mainly burning hoter using image processing field in recent years
Convolutional neural networks structure, for automatically extracting the required feature of classification, recognition accuracy is than traditional images processing method
It improves.But present most of convolutional neural networks structures are when handling multi-tag object identification problem, are according to each
Independent network is respectively trained in label, then by each independent network, carries out classification prediction to each label respectively.The party
Mainly there are two problems for method:On the one hand, since the network number of use is excessive, and the connection that is independent of each other, to increase network instruction
Experienced cost, causes information redundancy, so that time efficiency is low;On the other hand, intrinsic relationship between different labels is had ignored,
To have lost a part of priori, effective information is made not to be fully used, causes object identification accuracy rate not high.
Invention content
Goal of the invention:The repetition of characteristics of image is extracted for single label convolutional neural networks in the prior art, and
Between each label network lack contact, the problem for causing the accuracy of object recognition algorithm not high, provide it is a kind of using each label it
Between inclusion relation, carry out the feature extraction and classification of different levels, the object identification accuracy rate for solving traditional is not high
The multi-tag object identification method based on convolutional neural networks of technical problem.
Technical solution:In order to solve the above technical problems, the present invention provides a kind of multi-tag object based on convolutional neural networks
Body recognition methods, includes the following steps:
(1) data set used in object identification is read, the reference format of convolutional neural networks input is translated into;
(2) multi-tag convolutional neural networks model is built, and parameter initialization is carried out to the model;
(3) the multi-tag convolutional neural networks model put up is trained, continues to optimize network internal structural parameters;
(4) whether the multi-tag convolutional neural networks model in judgment step (3) after training meets training requirement, if full
Sufficient then enter step (5), return to step (3) re-starts training if being unsatisfactory for;
(5) test assessment is carried out to trained multi-tag convolutional neural networks model, obtains test accuracy rate;
(6) judge whether test accuracy rate can reach A grades, step (7) is if it is carried out, if otherwise updating model
Benefit soldier reenters step (2) and re-starts building, train and testing for model;
(7) final multi-tag convolutional neural networks model and parameter are exported, acquisition can be with the object identification of practical application
Method.
Further, the multi-tag convolutional neural networks model by data input unit, CNN characteristic extraction parts and divides
Class device part forms.
Further, CNN characteristic extraction parts need to carry out multiple convolution in the multi-tag convolutional neural networks model
Local shape factor is carried out to input feature vector figure.
Further, grader part includes similar point of several structures in the multi-tag convolutional neural networks model
Class device.
Further, the multi-tag convolutional neural networks model put up is trained in the step (3) specific
Steps are as follows:
(3.1) parameters of MLCNN models are initialized;
(3.2) training image of current step number is read to network layer;
(3.3) it allows image stream to carry out feedforward conduction in network model, obtains training error;According to current network parameter
Value, constantly carries out convolution sum pond arithmetic operation, until network successively since first convolutional layer to each image of reading
Export the training penalty values of each grader;
(3.4) whether the training penalty values exported in judgment step (3.3) reach trained penalty values requirement or reach setting
Step number enters step (3.5) if reaching, according to network losses value if not reaching, according to error back propagation
Method, obtain the variable quantity of each layer parameter, and carry out equivalent layer parameter update for next step number feed forward operation, finally
Return to step (3.2);
(3.5) network paramter models are exported.
Further, the method initialized to the parameters of MLCNN models in the step (3.1) includes constant
Initialization, Gaussian Profile initialize and are uniformly distributed initialization.
Further, test is carried out to trained multi-tag convolutional neural networks model in the step (5) to comment
Estimate and is as follows:
(5.1) network paramter models obtained in step (4) are loaded into multi-tag convolutional neural networks model;
(5.2) test image of current step number is read to network layer;
(5.3) parameter for obtaining the test image of current step number according to model structure and training successively, carries out convolution etc.
Feed forward operation operates, and exports the corresponding prediction classification of each image by grader part, and export prediction classification;
(5.4) judge whether current step number reaches the required minimum step number of the whole test set images of traversal, if reached
It then exports the label data of all images currently preserved and enters step (5.5);The return to step if not reaching
(5.2);
(5.5) the prediction classification for all test set images for recording front concrete class corresponding with each image carries out
Comparison, statistics obtain the test set classification accuracy under the model parameter.
Compared with the prior art, the advantages of the present invention are as follows:
Multi-tag object identification method provided by the invention based on convolutional neural networks, can solve object identification in object
In the case that body surface structure phase Sihe viewing angle is changeable, due to multi-tag relationship using not exclusively and acquisition of information it is not true
The low problem of qualitative caused correct recognition rata, and can shorten multi-tag identifying system training structure it is required when
Between.
Description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is the flow chart of multi-tag convolutional neural networks model training in Fig. 1;
Fig. 3 is the flow chart of multi-tag convolutional neural networks model measurement in Fig. 1;
Fig. 4 is the structural schematic diagram of multi-tag convolutional neural networks model in Fig. 1;
Fig. 5 is CNN characteristic extraction part structural schematic diagrams in embodiment;
Fig. 6 is grader structural schematic diagram in embodiment;
Fig. 7 is airplane data collection label hierarchical structure schematic diagram in embodiment.
Specific implementation mode
With reference to the accompanying drawings and detailed description, the present invention is furture elucidated.Embodiments described herein are only
A part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, ordinary skill people
Member's obtained other embodiment without making creative work belongs to the range that the present invention is protected.
The present invention proposes a kind of multi-tag object identification method based on convolutional neural networks.This method is directed to traditional convolution
Neural network handle multi-tag identification when there are the problem of, using the multi-tag convolutional neural networks (Multi- of such as Fig. 4
Label Convolutional Neural Network, MLCNN) structure, using the relationship between each label, by multiple labels
Feature extraction and fusion for classification to a complete network.
As shown in figure 4, MLCNN networks include mainly three parts:Data input unit, CNN characteristic extraction part Ci (i
=1,2 ..., n) and grader part Li (i=1,2 ..., n), the grader of each label Li has corresponding spy
Levy extractor Ci.Wherein n is number of tags possessed by the classification problem using the network, and is had as follows between each label
Inclusion relation:
Wherein,Show that any one class in label Li includes one or more of label Lj classes, that is,
Say that label Lj has classification that is more than label Li, more segmenting.It can be seen that similar between being subordinated to the inhomogeneity of label Lj
Property bigger, there is higher identification difficulty, it is therefore desirable to which the feature more abstract than label Li also just needs to build deeper
Convolutional network further carries out feature extraction.
Therefore, the characteristic extraction part C1 of lowermost layer is located in network MLCNN, is mainly responsible for the feature progress to image
Preliminary extraction, gained feature are used as the input of L1 label classifiers and deeper network C 2.Then, C2 pairs of characteristic extraction part
Characteristic pattern carries out the semantic abstraction of higher, and gained feature is used as the input of L2 label classifiers and deeper network C 3.According to
It is secondary downwards until the characteristic extraction part Cn of label Ln and the highest semanteme most segmented.
Since initial data input, data flow first passes around expansion and pretreatment, is then carried by corresponding CNN features
It takes part Ci to carry out convolution to be abstracted with after the dimensionality reduction of pond, respectively obtains and required feature of classifying is carried out to each label Li
Figure, then this feature figure is input in the corresponding graders of each label Li, by 1 × 1 convolution kernel dimensionality reduction and use dropout
Full articulamentum processing after, be input in softmax layers and carry out regression analysis, may finally obtain corresponding to each label Li
Classification.
Each section is described in detail as follows:
1.1 data input layers
Data input layer is mainly responsible for carries out expansion and data prediction to initial data set, to enhance the various of data
Property, prevent model over-fitting.This layer realize data extending method, mainly include flip horizontal, change of scale, rotation transformation and
Fancy PCA etc..
After expanding data set, need to carry out center type normalization operation to input picture, i.e., to each spy
Sign subtracts the average value of training set image, to highlight image individual difference.
1.2CNN feature extraction
As shown in figure 5, " CNN feature extractions " part mainly uses the core algorithm thought of CNN --- convolution is special to input
Sign figure carries out local shape factor, and in order to more efficiently control the depth and adjusting parameter of network, convolution kernel size is fixed
It is 3 × 3.Since the multi-tag network overall structure is deeper, to reduce the generation of gradient disperse, while avoiding ReLU functions common
" dead zone " problem, therefore the part use indexation linear unit (Exponential Linear Unit, ELU) function conduct
The activation primitive of characteristic value, function formula are as follows after convolution.
After multiple convolution, the characteristic pattern after convolution is carried out by the way of maximum value pond (max-pooling)
It is down-sampled, to reduce the size of characteristic pattern, and ensures a degree of feature invariance and prevent over-fitting.
1.3 grader
The grader part of MLCNN networks includes altogether the similar grader of n structure, corresponds respectively to the mark of n level
Label.The basic structure of the grader is as shown in Figure 6.
The characteristic pattern that characteristic extraction part obtains needs to first pass around the convolution operation dimensionality reduction that convolution kernel size is 1 × 1
Afterwards, then it is input to the full connection neuron for taking dropout strategies to be trained, the feature vector for then exporting neuron is defeated
Enter to softmax layers of progress regression analysis.
Softmax recurrence is popularization of the logistic regression in multi-class classification problem, class object loss function
For:
Wherein ω is network architecture parameters, xlFor the input of first of sample, N is training sample number, and C is single label
Including classification number, 1 { yl=i } meeting ylIt is 1 when=i, is otherwise 0.
Solution due to directly being acquired object function using mathematical method is difficult to realize, and is forced using gradient descent method
The minimum point of nearly cost function, and using the corresponding parameter of required minimum point as the systematic parameter estimated value of network, ladder
It is as follows to spend formula:
There is partial derivative formula above, it is brought into gradient descent algorithm, is carried out as follows in iteration each time
Weight updates, until target loss or train epochs reach requirement.
2 experiment tests
2.1 data set
The present invention carries out network training and test by taking aircraft object as an example, using the airplane data collection with multi-tag, comes
Detect the recognition performance of method provided by the present invention.It trains the reliability in order to ensure experimental result and is asked for multi-tag classification
The applicability of topic so that experiment can accurately test out the validity of algorithm, mainly selected when data set build Boeing with
Two manufacturers with Multiple Type aircraft of Airbus, while having carried out appropriate balance for the quantity of different label datas.
Image and label used by data set of the present invention, it is main to acquire from FGVC-Aircraft data sets, while basis
It needs to crawl picture from network by the way of reptile by aircraft type label, respective labels data is expanded, wrap altogether
Containing 10,000 aircraft brake discs.As shown in fig. 7, it includes mainly three-level label:" manufacturer ", " series ", " model ", and three has
Following inclusion relation:One manufacturer includes one or more series, and a series includes one or more models, and relationship is such as
Shown in figure.Such as figure, " manufacturer " includes 2 classes altogether, and " series " includes 14 classes altogether, and " model " includes 35 classes, and every aircraft brake disc all has
There are unique " manufacturer ", " series ", " model " mark.
In order to assess the performance that the present invention proposes algorithm, which has carried out multiple experiments, correct in the case of statistics is a variety of
Discrimination, correct recognition rata computational methods:
Wherein, P indicates correct recognition rata, nrIt indicates correctly to identify number in test sample, N indicates total test sample
Number.
2.2 experimental results and analysis
After the completion of dataset acquisition structure, 2/3rds of data set are split as training set, remaining three/
One is used as test set.Wherein training set is used as training neural network, makes network that can be updated under the supervision of object function certainly
Body weight, to realize self-teaching, and test set is used as the validity of verification algorithm.
Since the data set that the present invention uses has 3 labels, so n=3 corresponding in MLCNN structures in Fig. 4,
That is MLCNN used by experiment include altogether 3 CNN characteristic extraction parts (C1, C2, C3) and 3 corresponding labels (L1, L2,
L3 grader).Because the number of plies of convolutional layer directly determines the ability of CNN feature extractions, and influence the identification effect of grader
Fruit, so in the MLCNN networks, the number of plies of the respective convolutional layer of 3 CNN characteristic extraction parts (C1, C2, C3), is to influence it
The main structure parameters of recognition effect.
In order on data set of the present invention, select the preferable MLCNN networks of recognition effect.Not for C1, C2, C3 number of plies
It is distributed in proportion, different MLCNN networks is trained and is tested respectively.Simultaneously in order to which ensure between each group network can be into
Row compares, and total number of plies of fixed every group of MLCNN network is 15 layers.
The experiment chooses 7 groups of numbers of plies and is distributed different MLCNN networks altogether.Wherein, it includes 5 convolution that the 1st group of network, which corresponds to,
The C3 of the C1 of layer, C2 and 5 convolutional layer of 5 convolutional layers, the 2nd group of network correspond to C1,6 convolutional layers for including 6 convolutional layers
C2 and 3 convolutional layer C3.Equally, the number of the included convolutional layers of rear 5 groups of networks C1, C2, C3, is followed successively by:6 layers, 3 layers, 6
Layer, 3 layers, 6 layers, 6 layers, 3 layers, 3 layers, 9 layers, 3 layers, 9 layers, 3 layers, 9 layers, 3 layers, 3 layers.The experimental result of gained is as shown in table 1.
1 different characteristic of table extracts the test result of number of plies distribution
As it can be seen from table 1 when the discrimination of most subdivision label L3 is higher, the discrimination of label L1 and L2 are also higher;
The only discrimination of more final label L3 finds that the 2nd group and the 7th group of discrimination is higher than other each groups, and C1+C2=12 is at this time
Highest in several groups of experiments;Only compare the discrimination of label L2, finds to be also the 2nd group and the 7th group of discrimination highest, C1 layers at this time
Number is respectively 6 and 9, is not less than other each groups;Only compare the discrimination of label L1, be equally the 2nd group and the 7th group of discrimination most
It is high.
The number of plies that characteristic extraction part can be summarized as follows by the experiment selects experience:
(1) when total convolution number of plies is in unlimited time, the appropriate total number of plies for increasing MLCNN can improve the accurate of most subdivision label
Rate.
(2) when total convolution number of plies is limited, suitably reduce the number of plies of the last one feature extraction layer C3, increase C1's and C2
The number of plies can improve the discrimination of label L3;The appropriate number of plies for reducing C2, increases the number of plies of C1, can improve the knowledge of label L2
Not rate.It is summarized as the decrescence tactful of MLCNN feature extraction numbers of plies selection, it is as follows:
C1≥C2≥C3≥...≥Cn
Go out from the experimental results, when " CNN feature extractions " number of plies is using decrescence strategy, more can guarantee and accurately extract
Each required feature of layer label classifier, to generally improve the classification accuracy of each label.
The recognition effect for considering multiple labels selects the result of the 7th group of scheme of above-mentioned experiment to be directed to this as MLCNN
The optimal experimental result of invention data set.Meanwhile using other two kinds of traditional schemes, same data set is trained respectively and
Test.One is SIFT feature extractions, then the mode classified with SVM;Another kind is respectively trained solely for each label
Vertical CNN, for the ease of comparing, the number of plies of each CNN is corresponding with the feature extraction number of plies in MLCNN, respectively:9 layers, 12 layers
With 15 layers.Experimental result is as shown in table 2.
Test result of 2 distinct methods of table in data set of the present invention
From table 2 it can be seen that MLCNN proposed by the invention is above mutually independent list in the discrimination of three labels
Label C NN, especially in highest level label --- individual 15 layers of " model " CNN sorter networks are compared in " model ", are being tested
Discrimination on collection improves 7.54%.This shows that MLCNN proposed by the present invention can utilize the pass between different levels label
It is information, as the foundation of feature extraction, to reduce interference information, optimizes the recognition effect of network.
In order to test time efficiencies of the MLCNN in training, timing has been carried out while network training.Experiment is used
Allocation of computer be:GPU (TITANXP), CPU (E5-2650) and memory (64GB).Corresponding to the independent CNN of three labels
The training used time is respectively 5h, 6.5h and 7h, and the training used time of MLCNN is 7.5h.As can be seen that training one MLCNN compared with
Three independent CNN of training, it is more to have saved half in time.Therefore, there has also been larger to carry in time efficiency by MLCNN
It is high.
Claims (7)
1. a kind of multi-tag object identification method based on convolutional neural networks, which is characterized in that include the following steps:
(1) data set used in object identification is read, the reference format of convolutional neural networks input is translated into;
(2) multi-tag convolutional neural networks model is built, and parameter initialization is carried out to the model;
(3) the multi-tag convolutional neural networks model put up is trained, continues to optimize network internal structural parameters;
(4) whether the multi-tag convolutional neural networks model in judgment step (3) after training meets training requirement, if meeting
(5) are entered step, return to step (3) re-starts training if being unsatisfactory for;
(5) test assessment is carried out to trained multi-tag convolutional neural networks model, obtains test accuracy rate;
(6) judge whether test accuracy rate can reach A grades, if it is carry out step (7), if otherwise update model mends soldier
It reenters step (2) and re-starts building, train and testing for model;
(7) final multi-tag convolutional neural networks model and parameter are exported, acquisition can be with the object identification method of practical application.
2. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that
The multi-tag convolutional neural networks model is made of data input unit, CNN characteristic extraction parts and grader part.
3. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that
CNN characteristic extraction parts need to carry out multiple convolution to input feature vector figure carry out office in the multi-tag convolutional neural networks model
Portion's feature extraction.
4. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that
Grader part includes the similar grader of several structures in the multi-tag convolutional neural networks model.
5. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that
The multi-tag convolutional neural networks model put up is trained in the step (3) and is as follows:
(3.1) parameters of MLCNN models are initialized;
(3.2) training image of current step number is read to network layer;
(3.3) it allows image stream to carry out feedforward conduction in network model, obtains training error;According to current network parameter value, from
First convolutional layer starts each image constantly to reading and carries out convolution sum pond arithmetic operation successively, until network output is each
The training penalty values of grader;
(3.4) whether the training penalty values exported in judgment step (3.3) reach trained penalty values requirement or reach setting step number,
(3.5) are entered step if reaching, according to network losses value if not reaching, according to the side of error back propagation
Method obtains the variable quantity of each layer parameter, and carries out feed forward operation of the parameter update of equivalent layer for next step number, finally returns to
Step (3.2);
(3.5) network paramter models are exported.
6. a kind of multi-tag object identification method based on convolutional neural networks according to claim 5, which is characterized in that
At the beginning of the method initialized to the parameters of MLCNN models in the step (3.1) includes constant initialization, Gaussian Profile
Beginningization and it is uniformly distributed initialization.
7. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that
Test assessment is carried out in the step (5) to trained multi-tag convolutional neural networks model to be as follows:
(5.1) network paramter models obtained in step (4) are loaded into multi-tag convolutional neural networks model;
(5.2) test image of current step number is read to network layer;
(5.3) parameter for obtaining the test image of current step number according to model structure and training successively, carries out the feedforwards such as convolution
Arithmetic operation exports the corresponding prediction classification of each image by grader part, and exports prediction classification;
(5.4) judge whether current step number reaches the required minimum step number of the whole test set images of traversal, it is defeated if reaching
Go out the label data of all images currently preserved and enters step (5.5);Return to step (5.2) if not reaching;
(5.5) the prediction classification concrete class corresponding with each image for all test set images that front records is carried out pair
Than statistics obtains the test set classification accuracy under the model parameter.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810443651.6A CN108664924B (en) | 2018-05-10 | 2018-05-10 | Multi-label object identification method based on convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810443651.6A CN108664924B (en) | 2018-05-10 | 2018-05-10 | Multi-label object identification method based on convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108664924A true CN108664924A (en) | 2018-10-16 |
CN108664924B CN108664924B (en) | 2022-07-08 |
Family
ID=63778964
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810443651.6A Active CN108664924B (en) | 2018-05-10 | 2018-05-10 | Multi-label object identification method based on convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108664924B (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109886335A (en) * | 2019-02-21 | 2019-06-14 | 厦门美图之家科技有限公司 | Disaggregated model training method and device |
CN109977902A (en) * | 2019-04-03 | 2019-07-05 | 刘西 | A kind of construction vehicle identification method based on deep learning |
CN109993058A (en) * | 2019-02-27 | 2019-07-09 | 北京大学 | The recognition methods of road signs based on multi-tag classification |
CN110070113A (en) * | 2019-03-29 | 2019-07-30 | 广州思德医疗科技有限公司 | A kind of training method and device of training set |
CN110163161A (en) * | 2019-05-24 | 2019-08-23 | 西安电子科技大学 | Multiple features fusion pedestrian detection method based on Scale invariant |
CN110598849A (en) * | 2019-08-26 | 2019-12-20 | 广州大学 | HMET scattering parameter extraction method and system based on neural network and storage medium |
CN110648655A (en) * | 2019-09-11 | 2020-01-03 | 北京探境科技有限公司 | Voice recognition method, device, system and storage medium |
CN110689081A (en) * | 2019-09-30 | 2020-01-14 | 中国科学院大学 | Weak supervision target classification and positioning method based on bifurcation learning |
CN111353515A (en) * | 2018-12-21 | 2020-06-30 | 湖南工业大学 | Multi-scale grading-based classification and identification method for damage of train wheel set tread |
CN111507403A (en) * | 2020-04-17 | 2020-08-07 | 腾讯科技(深圳)有限公司 | Image classification method and device, computer equipment and storage medium |
CN111626357A (en) * | 2020-05-27 | 2020-09-04 | 北京微智信业科技有限公司 | Image identification method based on neural network model |
CN111898467A (en) * | 2020-07-08 | 2020-11-06 | 浙江大华技术股份有限公司 | Attribute identification method and device, storage medium and electronic device |
CN111967364A (en) * | 2020-08-11 | 2020-11-20 | 西交利物浦大学 | Composite fault diagnosis method, device, electronic equipment and storage medium |
CN112580750A (en) * | 2020-12-30 | 2021-03-30 | Oppo广东移动通信有限公司 | Image recognition method and device, electronic equipment and storage medium |
CN113449781A (en) * | 2021-06-17 | 2021-09-28 | 上海深至信息科技有限公司 | Generation method and system of thyroid nodule classification model |
CN113673482A (en) * | 2021-09-03 | 2021-11-19 | 四川大学 | Cell antinuclear antibody fluorescence recognition method and system based on dynamic label distribution |
CN115406852A (en) * | 2021-12-28 | 2022-11-29 | 中山小池科技有限公司 | Fabric fiber component qualitative method based on multi-label convolutional neural network |
US11978280B2 (en) | 2021-07-30 | 2024-05-07 | Lemon Inc. | Method and device for evaluating effect of classifying fuzzy attribute |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060147126A1 (en) * | 2005-01-06 | 2006-07-06 | Leo Grady | System and method for multilabel random walker segmentation using prior models |
CN102184422A (en) * | 2011-05-15 | 2011-09-14 | 中科院成都信息技术有限公司 | Average error classification cost minimized classifier integrating method |
CN105740906A (en) * | 2016-01-29 | 2016-07-06 | 中国科学院重庆绿色智能技术研究院 | Depth learning based vehicle multi-attribute federation analysis method |
CN106504255A (en) * | 2016-11-02 | 2017-03-15 | 南京大学 | A kind of multi-Target Image joint dividing method based on multi-tag multi-instance learning |
CN107330444A (en) * | 2017-05-27 | 2017-11-07 | 苏州科技大学 | A kind of image autotext mask method based on generation confrontation network |
CN107577785A (en) * | 2017-09-15 | 2018-01-12 | 南京大学 | A kind of level multi-tag sorting technique suitable for law identification |
-
2018
- 2018-05-10 CN CN201810443651.6A patent/CN108664924B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060147126A1 (en) * | 2005-01-06 | 2006-07-06 | Leo Grady | System and method for multilabel random walker segmentation using prior models |
CN102184422A (en) * | 2011-05-15 | 2011-09-14 | 中科院成都信息技术有限公司 | Average error classification cost minimized classifier integrating method |
CN105740906A (en) * | 2016-01-29 | 2016-07-06 | 中国科学院重庆绿色智能技术研究院 | Depth learning based vehicle multi-attribute federation analysis method |
CN106504255A (en) * | 2016-11-02 | 2017-03-15 | 南京大学 | A kind of multi-Target Image joint dividing method based on multi-tag multi-instance learning |
CN107330444A (en) * | 2017-05-27 | 2017-11-07 | 苏州科技大学 | A kind of image autotext mask method based on generation confrontation network |
CN107577785A (en) * | 2017-09-15 | 2018-01-12 | 南京大学 | A kind of level multi-tag sorting technique suitable for law identification |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111353515B (en) * | 2018-12-21 | 2024-01-26 | 湖南工业大学 | Multi-scale classification-based train wheel set tread damage classification and identification method |
CN111353515A (en) * | 2018-12-21 | 2020-06-30 | 湖南工业大学 | Multi-scale grading-based classification and identification method for damage of train wheel set tread |
CN109886335A (en) * | 2019-02-21 | 2019-06-14 | 厦门美图之家科技有限公司 | Disaggregated model training method and device |
CN109886335B (en) * | 2019-02-21 | 2021-11-26 | 厦门美图之家科技有限公司 | Classification model training method and device |
CN109993058A (en) * | 2019-02-27 | 2019-07-09 | 北京大学 | The recognition methods of road signs based on multi-tag classification |
CN110070113B (en) * | 2019-03-29 | 2021-03-30 | 广州思德医疗科技有限公司 | Training method and device for training set |
CN110070113A (en) * | 2019-03-29 | 2019-07-30 | 广州思德医疗科技有限公司 | A kind of training method and device of training set |
CN109977902A (en) * | 2019-04-03 | 2019-07-05 | 刘西 | A kind of construction vehicle identification method based on deep learning |
CN110163161A (en) * | 2019-05-24 | 2019-08-23 | 西安电子科技大学 | Multiple features fusion pedestrian detection method based on Scale invariant |
CN110598849A (en) * | 2019-08-26 | 2019-12-20 | 广州大学 | HMET scattering parameter extraction method and system based on neural network and storage medium |
CN110648655A (en) * | 2019-09-11 | 2020-01-03 | 北京探境科技有限公司 | Voice recognition method, device, system and storage medium |
CN110648655B (en) * | 2019-09-11 | 2022-03-01 | 北京探境科技有限公司 | Voice recognition method, device, system and storage medium |
CN110689081A (en) * | 2019-09-30 | 2020-01-14 | 中国科学院大学 | Weak supervision target classification and positioning method based on bifurcation learning |
CN111507403A (en) * | 2020-04-17 | 2020-08-07 | 腾讯科技(深圳)有限公司 | Image classification method and device, computer equipment and storage medium |
CN111626357A (en) * | 2020-05-27 | 2020-09-04 | 北京微智信业科技有限公司 | Image identification method based on neural network model |
CN111626357B (en) * | 2020-05-27 | 2021-11-23 | 北京东方通网信科技有限公司 | Image identification method based on neural network model |
CN111898467B (en) * | 2020-07-08 | 2023-02-28 | 浙江大华技术股份有限公司 | Attribute identification method and device, storage medium and electronic device |
CN111898467A (en) * | 2020-07-08 | 2020-11-06 | 浙江大华技术股份有限公司 | Attribute identification method and device, storage medium and electronic device |
CN111967364A (en) * | 2020-08-11 | 2020-11-20 | 西交利物浦大学 | Composite fault diagnosis method, device, electronic equipment and storage medium |
CN111967364B (en) * | 2020-08-11 | 2023-11-21 | 西交利物浦大学 | Composite fault diagnosis method, device, electronic equipment and storage medium |
CN112580750A (en) * | 2020-12-30 | 2021-03-30 | Oppo广东移动通信有限公司 | Image recognition method and device, electronic equipment and storage medium |
CN113449781A (en) * | 2021-06-17 | 2021-09-28 | 上海深至信息科技有限公司 | Generation method and system of thyroid nodule classification model |
CN113449781B (en) * | 2021-06-17 | 2023-04-07 | 上海深至信息科技有限公司 | Generation method and system of thyroid nodule classification model |
US11978280B2 (en) | 2021-07-30 | 2024-05-07 | Lemon Inc. | Method and device for evaluating effect of classifying fuzzy attribute |
CN113673482B (en) * | 2021-09-03 | 2023-04-18 | 四川大学 | Cell antinuclear antibody fluorescence recognition method and system based on dynamic label distribution |
CN113673482A (en) * | 2021-09-03 | 2021-11-19 | 四川大学 | Cell antinuclear antibody fluorescence recognition method and system based on dynamic label distribution |
CN115406852A (en) * | 2021-12-28 | 2022-11-29 | 中山小池科技有限公司 | Fabric fiber component qualitative method based on multi-label convolutional neural network |
Also Published As
Publication number | Publication date |
---|---|
CN108664924B (en) | 2022-07-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108664924A (en) | A kind of multi-tag object identification method based on convolutional neural networks | |
US11195051B2 (en) | Method for person re-identification based on deep model with multi-loss fusion training strategy | |
CN109961034B (en) | Video target detection method based on convolution gating cyclic neural unit | |
CN111553193B (en) | Visual SLAM closed-loop detection method based on lightweight deep neural network | |
CN113378632B (en) | Pseudo-label optimization-based unsupervised domain adaptive pedestrian re-identification method | |
CN109325547A (en) | Non-motor vehicle image multi-tag classification method, system, equipment and storage medium | |
CN101097564B (en) | Parameter learning method, parameter learning apparatus, pattern classification method, and pattern classification apparatus | |
CN110276248B (en) | Facial expression recognition method based on sample weight distribution and deep learning | |
CN109002755B (en) | Age estimation model construction method and estimation method based on face image | |
CN112883839B (en) | Remote sensing image interpretation method based on adaptive sample set construction and deep learning | |
CN105389326B (en) | Image labeling method based on weak matching probability typical relevancy models | |
CN111461038A (en) | Pedestrian re-identification method based on layered multi-mode attention mechanism | |
CN106845525A (en) | A kind of depth confidence network image bracket protocol based on bottom fusion feature | |
CN110765960B (en) | Pedestrian re-identification method for adaptive multi-task deep learning | |
CN117149944B (en) | Multi-mode situation emotion recognition method and system based on wide time range | |
CN108879732A (en) | Transient stability evaluation in power system method and device | |
KR102049331B1 (en) | Apparatus and method for classifying images, and apparatus for training images for classification of images | |
Chen et al. | Military image scene recognition based on CNN and semantic information | |
CN109948498A (en) | A kind of dynamic gesture identification method based on 3D convolutional neural networks algorithm | |
CN113920472A (en) | Unsupervised target re-identification method and system based on attention mechanism | |
CN103136540A (en) | Behavior recognition method based on concealed structure reasoning | |
CN116310647A (en) | Labor insurance object target detection method and system based on incremental learning | |
Yao | Application of higher education management in colleges and universities by deep learning | |
CN116363712B (en) | Palmprint palm vein recognition method based on modal informativity evaluation strategy | |
Banik et al. | Multi-label object attribute classification using a convolutional neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |