CN108664924A - A kind of multi-tag object identification method based on convolutional neural networks - Google Patents

A kind of multi-tag object identification method based on convolutional neural networks Download PDF

Info

Publication number
CN108664924A
CN108664924A CN201810443651.6A CN201810443651A CN108664924A CN 108664924 A CN108664924 A CN 108664924A CN 201810443651 A CN201810443651 A CN 201810443651A CN 108664924 A CN108664924 A CN 108664924A
Authority
CN
China
Prior art keywords
convolutional neural
neural networks
tag
network
object identification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810443651.6A
Other languages
Chinese (zh)
Other versions
CN108664924B (en
Inventor
李新德
孙振华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN201810443651.6A priority Critical patent/CN108664924B/en
Publication of CN108664924A publication Critical patent/CN108664924A/en
Application granted granted Critical
Publication of CN108664924B publication Critical patent/CN108664924B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of multi-tag object identification method based on convolutional neural networks, this method utilizes the inclusion relation between label, and according to including the CNN characteristic extraction parts for sequentially building each label successively, pass through the constantly abstract each layer feature of extraction of convolution algorithm operation, and the grader of each label is respectively set in the different depth of network, the characteristic pattern that the CNN characteristic extraction parts of respective labels are extracted, it is input to corresponding grader, simultaneously error back propagation is carried out using multiple graders, the network weight parameter of the corresponding each layer of supervised training, each label classification is finally obtained to complete to identify.The multi-tag convolutional neural networks that the present invention uses, can well solve the fusion of information between multiple labels, solve the problems, such as that traditional multi-tag object identification accuracy rate is not high, while improve the efficiency of training and identification.

Description

A kind of multi-tag object identification method based on convolutional neural networks
Technical field
The invention belongs to technical field of image processing, more particularly to a kind of multi-tag object based on convolutional neural networks is known Other method.
Background technology
With the rapid development of artificial intelligence technology, automatic object identification has become the hot spot studied both at home and abroad in recent years Problem is with a wide range of applications in fields such as intelligent monitoring, telemetering remote sensing, robot, Medical Image Processings.Really giving birth to In work, kind of object is various, and individual similarity is higher, the mankind can by visual perception shape, color and distance etc. information, and And these comprehensive information, accurately to judge object generic, but this is often relatively difficult for computer.Cause How this, make computer possess similar with the mankind, the even more than mankind recognition capabilities, it has also become current object identifies important Direction and significant challenge.
The main approaches of object identification are that extraction object features realize Object representation, then utilize certain engineering The study that algorithm carries out object type is practised, is finally classified to object, realizes object identification.But target to be identified exists The high situation of similarity, the feature extracted in this case are generally difficult to reflect class inherited and class difference greatly, between class in class Interior general character.Especially in the case where object has multiple labels, due to the limitation of traditional characteristic extracting mode, different labels are made It is often no associated between feature, it is more difficult that this makes object identification get up.
Multi-tag object identification problem relies primarily on two kinds of method to handle at present.One is based on traditional images The method of processing extracts the features such as some SIFT, HOG, SURF to subject image first, in conjunction with optimization BoW models and melt Hop algorithm converts feature, and traditional grader (such as SVM) is finally used to obtain the prediction result of object category, the party The recognition speed of method is than very fast, but recognition accuracy is relatively low.
Another kind is the method based on deep learning, and this method is mainly burning hoter using image processing field in recent years Convolutional neural networks structure, for automatically extracting the required feature of classification, recognition accuracy is than traditional images processing method It improves.But present most of convolutional neural networks structures are when handling multi-tag object identification problem, are according to each Independent network is respectively trained in label, then by each independent network, carries out classification prediction to each label respectively.The party Mainly there are two problems for method:On the one hand, since the network number of use is excessive, and the connection that is independent of each other, to increase network instruction Experienced cost, causes information redundancy, so that time efficiency is low;On the other hand, intrinsic relationship between different labels is had ignored, To have lost a part of priori, effective information is made not to be fully used, causes object identification accuracy rate not high.
Invention content
Goal of the invention:The repetition of characteristics of image is extracted for single label convolutional neural networks in the prior art, and Between each label network lack contact, the problem for causing the accuracy of object recognition algorithm not high, provide it is a kind of using each label it Between inclusion relation, carry out the feature extraction and classification of different levels, the object identification accuracy rate for solving traditional is not high The multi-tag object identification method based on convolutional neural networks of technical problem.
Technical solution:In order to solve the above technical problems, the present invention provides a kind of multi-tag object based on convolutional neural networks Body recognition methods, includes the following steps:
(1) data set used in object identification is read, the reference format of convolutional neural networks input is translated into;
(2) multi-tag convolutional neural networks model is built, and parameter initialization is carried out to the model;
(3) the multi-tag convolutional neural networks model put up is trained, continues to optimize network internal structural parameters;
(4) whether the multi-tag convolutional neural networks model in judgment step (3) after training meets training requirement, if full Sufficient then enter step (5), return to step (3) re-starts training if being unsatisfactory for;
(5) test assessment is carried out to trained multi-tag convolutional neural networks model, obtains test accuracy rate;
(6) judge whether test accuracy rate can reach A grades, step (7) is if it is carried out, if otherwise updating model Benefit soldier reenters step (2) and re-starts building, train and testing for model;
(7) final multi-tag convolutional neural networks model and parameter are exported, acquisition can be with the object identification of practical application Method.
Further, the multi-tag convolutional neural networks model by data input unit, CNN characteristic extraction parts and divides Class device part forms.
Further, CNN characteristic extraction parts need to carry out multiple convolution in the multi-tag convolutional neural networks model Local shape factor is carried out to input feature vector figure.
Further, grader part includes similar point of several structures in the multi-tag convolutional neural networks model Class device.
Further, the multi-tag convolutional neural networks model put up is trained in the step (3) specific Steps are as follows:
(3.1) parameters of MLCNN models are initialized;
(3.2) training image of current step number is read to network layer;
(3.3) it allows image stream to carry out feedforward conduction in network model, obtains training error;According to current network parameter Value, constantly carries out convolution sum pond arithmetic operation, until network successively since first convolutional layer to each image of reading Export the training penalty values of each grader;
(3.4) whether the training penalty values exported in judgment step (3.3) reach trained penalty values requirement or reach setting Step number enters step (3.5) if reaching, according to network losses value if not reaching, according to error back propagation Method, obtain the variable quantity of each layer parameter, and carry out equivalent layer parameter update for next step number feed forward operation, finally Return to step (3.2);
(3.5) network paramter models are exported.
Further, the method initialized to the parameters of MLCNN models in the step (3.1) includes constant Initialization, Gaussian Profile initialize and are uniformly distributed initialization.
Further, test is carried out to trained multi-tag convolutional neural networks model in the step (5) to comment Estimate and is as follows:
(5.1) network paramter models obtained in step (4) are loaded into multi-tag convolutional neural networks model;
(5.2) test image of current step number is read to network layer;
(5.3) parameter for obtaining the test image of current step number according to model structure and training successively, carries out convolution etc. Feed forward operation operates, and exports the corresponding prediction classification of each image by grader part, and export prediction classification;
(5.4) judge whether current step number reaches the required minimum step number of the whole test set images of traversal, if reached It then exports the label data of all images currently preserved and enters step (5.5);The return to step if not reaching (5.2);
(5.5) the prediction classification for all test set images for recording front concrete class corresponding with each image carries out Comparison, statistics obtain the test set classification accuracy under the model parameter.
Compared with the prior art, the advantages of the present invention are as follows:
Multi-tag object identification method provided by the invention based on convolutional neural networks, can solve object identification in object In the case that body surface structure phase Sihe viewing angle is changeable, due to multi-tag relationship using not exclusively and acquisition of information it is not true The low problem of qualitative caused correct recognition rata, and can shorten multi-tag identifying system training structure it is required when Between.
Description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is the flow chart of multi-tag convolutional neural networks model training in Fig. 1;
Fig. 3 is the flow chart of multi-tag convolutional neural networks model measurement in Fig. 1;
Fig. 4 is the structural schematic diagram of multi-tag convolutional neural networks model in Fig. 1;
Fig. 5 is CNN characteristic extraction part structural schematic diagrams in embodiment;
Fig. 6 is grader structural schematic diagram in embodiment;
Fig. 7 is airplane data collection label hierarchical structure schematic diagram in embodiment.
Specific implementation mode
With reference to the accompanying drawings and detailed description, the present invention is furture elucidated.Embodiments described herein are only A part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, ordinary skill people Member's obtained other embodiment without making creative work belongs to the range that the present invention is protected.
The present invention proposes a kind of multi-tag object identification method based on convolutional neural networks.This method is directed to traditional convolution Neural network handle multi-tag identification when there are the problem of, using the multi-tag convolutional neural networks (Multi- of such as Fig. 4 Label Convolutional Neural Network, MLCNN) structure, using the relationship between each label, by multiple labels Feature extraction and fusion for classification to a complete network.
As shown in figure 4, MLCNN networks include mainly three parts:Data input unit, CNN characteristic extraction part Ci (i =1,2 ..., n) and grader part Li (i=1,2 ..., n), the grader of each label Li has corresponding spy Levy extractor Ci.Wherein n is number of tags possessed by the classification problem using the network, and is had as follows between each label Inclusion relation:
Wherein,Show that any one class in label Li includes one or more of label Lj classes, that is, Say that label Lj has classification that is more than label Li, more segmenting.It can be seen that similar between being subordinated to the inhomogeneity of label Lj Property bigger, there is higher identification difficulty, it is therefore desirable to which the feature more abstract than label Li also just needs to build deeper Convolutional network further carries out feature extraction.
Therefore, the characteristic extraction part C1 of lowermost layer is located in network MLCNN, is mainly responsible for the feature progress to image Preliminary extraction, gained feature are used as the input of L1 label classifiers and deeper network C 2.Then, C2 pairs of characteristic extraction part Characteristic pattern carries out the semantic abstraction of higher, and gained feature is used as the input of L2 label classifiers and deeper network C 3.According to It is secondary downwards until the characteristic extraction part Cn of label Ln and the highest semanteme most segmented.
Since initial data input, data flow first passes around expansion and pretreatment, is then carried by corresponding CNN features It takes part Ci to carry out convolution to be abstracted with after the dimensionality reduction of pond, respectively obtains and required feature of classifying is carried out to each label Li Figure, then this feature figure is input in the corresponding graders of each label Li, by 1 × 1 convolution kernel dimensionality reduction and use dropout Full articulamentum processing after, be input in softmax layers and carry out regression analysis, may finally obtain corresponding to each label Li Classification.
Each section is described in detail as follows:
1.1 data input layers
Data input layer is mainly responsible for carries out expansion and data prediction to initial data set, to enhance the various of data Property, prevent model over-fitting.This layer realize data extending method, mainly include flip horizontal, change of scale, rotation transformation and Fancy PCA etc..
After expanding data set, need to carry out center type normalization operation to input picture, i.e., to each spy Sign subtracts the average value of training set image, to highlight image individual difference.
1.2CNN feature extraction
As shown in figure 5, " CNN feature extractions " part mainly uses the core algorithm thought of CNN --- convolution is special to input Sign figure carries out local shape factor, and in order to more efficiently control the depth and adjusting parameter of network, convolution kernel size is fixed It is 3 × 3.Since the multi-tag network overall structure is deeper, to reduce the generation of gradient disperse, while avoiding ReLU functions common " dead zone " problem, therefore the part use indexation linear unit (Exponential Linear Unit, ELU) function conduct The activation primitive of characteristic value, function formula are as follows after convolution.
After multiple convolution, the characteristic pattern after convolution is carried out by the way of maximum value pond (max-pooling) It is down-sampled, to reduce the size of characteristic pattern, and ensures a degree of feature invariance and prevent over-fitting.
1.3 grader
The grader part of MLCNN networks includes altogether the similar grader of n structure, corresponds respectively to the mark of n level Label.The basic structure of the grader is as shown in Figure 6.
The characteristic pattern that characteristic extraction part obtains needs to first pass around the convolution operation dimensionality reduction that convolution kernel size is 1 × 1 Afterwards, then it is input to the full connection neuron for taking dropout strategies to be trained, the feature vector for then exporting neuron is defeated Enter to softmax layers of progress regression analysis.
Softmax recurrence is popularization of the logistic regression in multi-class classification problem, class object loss function For:
Wherein ω is network architecture parameters, xlFor the input of first of sample, N is training sample number, and C is single label Including classification number, 1 { yl=i } meeting ylIt is 1 when=i, is otherwise 0.
Solution due to directly being acquired object function using mathematical method is difficult to realize, and is forced using gradient descent method The minimum point of nearly cost function, and using the corresponding parameter of required minimum point as the systematic parameter estimated value of network, ladder It is as follows to spend formula:
There is partial derivative formula above, it is brought into gradient descent algorithm, is carried out as follows in iteration each time Weight updates, until target loss or train epochs reach requirement.
2 experiment tests
2.1 data set
The present invention carries out network training and test by taking aircraft object as an example, using the airplane data collection with multi-tag, comes Detect the recognition performance of method provided by the present invention.It trains the reliability in order to ensure experimental result and is asked for multi-tag classification The applicability of topic so that experiment can accurately test out the validity of algorithm, mainly selected when data set build Boeing with Two manufacturers with Multiple Type aircraft of Airbus, while having carried out appropriate balance for the quantity of different label datas.
Image and label used by data set of the present invention, it is main to acquire from FGVC-Aircraft data sets, while basis It needs to crawl picture from network by the way of reptile by aircraft type label, respective labels data is expanded, wrap altogether Containing 10,000 aircraft brake discs.As shown in fig. 7, it includes mainly three-level label:" manufacturer ", " series ", " model ", and three has Following inclusion relation:One manufacturer includes one or more series, and a series includes one or more models, and relationship is such as Shown in figure.Such as figure, " manufacturer " includes 2 classes altogether, and " series " includes 14 classes altogether, and " model " includes 35 classes, and every aircraft brake disc all has There are unique " manufacturer ", " series ", " model " mark.
In order to assess the performance that the present invention proposes algorithm, which has carried out multiple experiments, correct in the case of statistics is a variety of Discrimination, correct recognition rata computational methods:
Wherein, P indicates correct recognition rata, nrIt indicates correctly to identify number in test sample, N indicates total test sample Number.
2.2 experimental results and analysis
After the completion of dataset acquisition structure, 2/3rds of data set are split as training set, remaining three/ One is used as test set.Wherein training set is used as training neural network, makes network that can be updated under the supervision of object function certainly Body weight, to realize self-teaching, and test set is used as the validity of verification algorithm.
Since the data set that the present invention uses has 3 labels, so n=3 corresponding in MLCNN structures in Fig. 4, That is MLCNN used by experiment include altogether 3 CNN characteristic extraction parts (C1, C2, C3) and 3 corresponding labels (L1, L2, L3 grader).Because the number of plies of convolutional layer directly determines the ability of CNN feature extractions, and influence the identification effect of grader Fruit, so in the MLCNN networks, the number of plies of the respective convolutional layer of 3 CNN characteristic extraction parts (C1, C2, C3), is to influence it The main structure parameters of recognition effect.
In order on data set of the present invention, select the preferable MLCNN networks of recognition effect.Not for C1, C2, C3 number of plies It is distributed in proportion, different MLCNN networks is trained and is tested respectively.Simultaneously in order to which ensure between each group network can be into Row compares, and total number of plies of fixed every group of MLCNN network is 15 layers.
The experiment chooses 7 groups of numbers of plies and is distributed different MLCNN networks altogether.Wherein, it includes 5 convolution that the 1st group of network, which corresponds to, The C3 of the C1 of layer, C2 and 5 convolutional layer of 5 convolutional layers, the 2nd group of network correspond to C1,6 convolutional layers for including 6 convolutional layers C2 and 3 convolutional layer C3.Equally, the number of the included convolutional layers of rear 5 groups of networks C1, C2, C3, is followed successively by:6 layers, 3 layers, 6 Layer, 3 layers, 6 layers, 6 layers, 3 layers, 3 layers, 9 layers, 3 layers, 9 layers, 3 layers, 9 layers, 3 layers, 3 layers.The experimental result of gained is as shown in table 1.
1 different characteristic of table extracts the test result of number of plies distribution
As it can be seen from table 1 when the discrimination of most subdivision label L3 is higher, the discrimination of label L1 and L2 are also higher; The only discrimination of more final label L3 finds that the 2nd group and the 7th group of discrimination is higher than other each groups, and C1+C2=12 is at this time Highest in several groups of experiments;Only compare the discrimination of label L2, finds to be also the 2nd group and the 7th group of discrimination highest, C1 layers at this time Number is respectively 6 and 9, is not less than other each groups;Only compare the discrimination of label L1, be equally the 2nd group and the 7th group of discrimination most It is high.
The number of plies that characteristic extraction part can be summarized as follows by the experiment selects experience:
(1) when total convolution number of plies is in unlimited time, the appropriate total number of plies for increasing MLCNN can improve the accurate of most subdivision label Rate.
(2) when total convolution number of plies is limited, suitably reduce the number of plies of the last one feature extraction layer C3, increase C1's and C2 The number of plies can improve the discrimination of label L3;The appropriate number of plies for reducing C2, increases the number of plies of C1, can improve the knowledge of label L2 Not rate.It is summarized as the decrescence tactful of MLCNN feature extraction numbers of plies selection, it is as follows:
C1≥C2≥C3≥...≥Cn
Go out from the experimental results, when " CNN feature extractions " number of plies is using decrescence strategy, more can guarantee and accurately extract Each required feature of layer label classifier, to generally improve the classification accuracy of each label.
The recognition effect for considering multiple labels selects the result of the 7th group of scheme of above-mentioned experiment to be directed to this as MLCNN The optimal experimental result of invention data set.Meanwhile using other two kinds of traditional schemes, same data set is trained respectively and Test.One is SIFT feature extractions, then the mode classified with SVM;Another kind is respectively trained solely for each label Vertical CNN, for the ease of comparing, the number of plies of each CNN is corresponding with the feature extraction number of plies in MLCNN, respectively:9 layers, 12 layers With 15 layers.Experimental result is as shown in table 2.
Test result of 2 distinct methods of table in data set of the present invention
From table 2 it can be seen that MLCNN proposed by the invention is above mutually independent list in the discrimination of three labels Label C NN, especially in highest level label --- individual 15 layers of " model " CNN sorter networks are compared in " model ", are being tested Discrimination on collection improves 7.54%.This shows that MLCNN proposed by the present invention can utilize the pass between different levels label It is information, as the foundation of feature extraction, to reduce interference information, optimizes the recognition effect of network.
In order to test time efficiencies of the MLCNN in training, timing has been carried out while network training.Experiment is used Allocation of computer be:GPU (TITANXP), CPU (E5-2650) and memory (64GB).Corresponding to the independent CNN of three labels The training used time is respectively 5h, 6.5h and 7h, and the training used time of MLCNN is 7.5h.As can be seen that training one MLCNN compared with Three independent CNN of training, it is more to have saved half in time.Therefore, there has also been larger to carry in time efficiency by MLCNN It is high.

Claims (7)

1. a kind of multi-tag object identification method based on convolutional neural networks, which is characterized in that include the following steps:
(1) data set used in object identification is read, the reference format of convolutional neural networks input is translated into;
(2) multi-tag convolutional neural networks model is built, and parameter initialization is carried out to the model;
(3) the multi-tag convolutional neural networks model put up is trained, continues to optimize network internal structural parameters;
(4) whether the multi-tag convolutional neural networks model in judgment step (3) after training meets training requirement, if meeting (5) are entered step, return to step (3) re-starts training if being unsatisfactory for;
(5) test assessment is carried out to trained multi-tag convolutional neural networks model, obtains test accuracy rate;
(6) judge whether test accuracy rate can reach A grades, if it is carry out step (7), if otherwise update model mends soldier It reenters step (2) and re-starts building, train and testing for model;
(7) final multi-tag convolutional neural networks model and parameter are exported, acquisition can be with the object identification method of practical application.
2. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that The multi-tag convolutional neural networks model is made of data input unit, CNN characteristic extraction parts and grader part.
3. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that CNN characteristic extraction parts need to carry out multiple convolution to input feature vector figure carry out office in the multi-tag convolutional neural networks model Portion's feature extraction.
4. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that Grader part includes the similar grader of several structures in the multi-tag convolutional neural networks model.
5. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that The multi-tag convolutional neural networks model put up is trained in the step (3) and is as follows:
(3.1) parameters of MLCNN models are initialized;
(3.2) training image of current step number is read to network layer;
(3.3) it allows image stream to carry out feedforward conduction in network model, obtains training error;According to current network parameter value, from First convolutional layer starts each image constantly to reading and carries out convolution sum pond arithmetic operation successively, until network output is each The training penalty values of grader;
(3.4) whether the training penalty values exported in judgment step (3.3) reach trained penalty values requirement or reach setting step number, (3.5) are entered step if reaching, according to network losses value if not reaching, according to the side of error back propagation Method obtains the variable quantity of each layer parameter, and carries out feed forward operation of the parameter update of equivalent layer for next step number, finally returns to Step (3.2);
(3.5) network paramter models are exported.
6. a kind of multi-tag object identification method based on convolutional neural networks according to claim 5, which is characterized in that At the beginning of the method initialized to the parameters of MLCNN models in the step (3.1) includes constant initialization, Gaussian Profile Beginningization and it is uniformly distributed initialization.
7. a kind of multi-tag object identification method based on convolutional neural networks according to claim 1, which is characterized in that Test assessment is carried out in the step (5) to trained multi-tag convolutional neural networks model to be as follows:
(5.1) network paramter models obtained in step (4) are loaded into multi-tag convolutional neural networks model;
(5.2) test image of current step number is read to network layer;
(5.3) parameter for obtaining the test image of current step number according to model structure and training successively, carries out the feedforwards such as convolution Arithmetic operation exports the corresponding prediction classification of each image by grader part, and exports prediction classification;
(5.4) judge whether current step number reaches the required minimum step number of the whole test set images of traversal, it is defeated if reaching Go out the label data of all images currently preserved and enters step (5.5);Return to step (5.2) if not reaching;
(5.5) the prediction classification concrete class corresponding with each image for all test set images that front records is carried out pair Than statistics obtains the test set classification accuracy under the model parameter.
CN201810443651.6A 2018-05-10 2018-05-10 Multi-label object identification method based on convolutional neural network Active CN108664924B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810443651.6A CN108664924B (en) 2018-05-10 2018-05-10 Multi-label object identification method based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810443651.6A CN108664924B (en) 2018-05-10 2018-05-10 Multi-label object identification method based on convolutional neural network

Publications (2)

Publication Number Publication Date
CN108664924A true CN108664924A (en) 2018-10-16
CN108664924B CN108664924B (en) 2022-07-08

Family

ID=63778964

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810443651.6A Active CN108664924B (en) 2018-05-10 2018-05-10 Multi-label object identification method based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN108664924B (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109886335A (en) * 2019-02-21 2019-06-14 厦门美图之家科技有限公司 Disaggregated model training method and device
CN109977902A (en) * 2019-04-03 2019-07-05 刘西 A kind of construction vehicle identification method based on deep learning
CN109993058A (en) * 2019-02-27 2019-07-09 北京大学 The recognition methods of road signs based on multi-tag classification
CN110070113A (en) * 2019-03-29 2019-07-30 广州思德医疗科技有限公司 A kind of training method and device of training set
CN110163161A (en) * 2019-05-24 2019-08-23 西安电子科技大学 Multiple features fusion pedestrian detection method based on Scale invariant
CN110598849A (en) * 2019-08-26 2019-12-20 广州大学 HMET scattering parameter extraction method and system based on neural network and storage medium
CN110648655A (en) * 2019-09-11 2020-01-03 北京探境科技有限公司 Voice recognition method, device, system and storage medium
CN110689081A (en) * 2019-09-30 2020-01-14 中国科学院大学 Weak supervision target classification and positioning method based on bifurcation learning
CN111353515A (en) * 2018-12-21 2020-06-30 湖南工业大学 Multi-scale grading-based classification and identification method for damage of train wheel set tread
CN111507403A (en) * 2020-04-17 2020-08-07 腾讯科技(深圳)有限公司 Image classification method and device, computer equipment and storage medium
CN111626357A (en) * 2020-05-27 2020-09-04 北京微智信业科技有限公司 Image identification method based on neural network model
CN111898467A (en) * 2020-07-08 2020-11-06 浙江大华技术股份有限公司 Attribute identification method and device, storage medium and electronic device
CN111967364A (en) * 2020-08-11 2020-11-20 西交利物浦大学 Composite fault diagnosis method, device, electronic equipment and storage medium
CN112580750A (en) * 2020-12-30 2021-03-30 Oppo广东移动通信有限公司 Image recognition method and device, electronic equipment and storage medium
CN113449781A (en) * 2021-06-17 2021-09-28 上海深至信息科技有限公司 Generation method and system of thyroid nodule classification model
CN113673482A (en) * 2021-09-03 2021-11-19 四川大学 Cell antinuclear antibody fluorescence recognition method and system based on dynamic label distribution
CN115406852A (en) * 2021-12-28 2022-11-29 中山小池科技有限公司 Fabric fiber component qualitative method based on multi-label convolutional neural network
US11978280B2 (en) 2021-07-30 2024-05-07 Lemon Inc. Method and device for evaluating effect of classifying fuzzy attribute

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060147126A1 (en) * 2005-01-06 2006-07-06 Leo Grady System and method for multilabel random walker segmentation using prior models
CN102184422A (en) * 2011-05-15 2011-09-14 中科院成都信息技术有限公司 Average error classification cost minimized classifier integrating method
CN105740906A (en) * 2016-01-29 2016-07-06 中国科学院重庆绿色智能技术研究院 Depth learning based vehicle multi-attribute federation analysis method
CN106504255A (en) * 2016-11-02 2017-03-15 南京大学 A kind of multi-Target Image joint dividing method based on multi-tag multi-instance learning
CN107330444A (en) * 2017-05-27 2017-11-07 苏州科技大学 A kind of image autotext mask method based on generation confrontation network
CN107577785A (en) * 2017-09-15 2018-01-12 南京大学 A kind of level multi-tag sorting technique suitable for law identification

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060147126A1 (en) * 2005-01-06 2006-07-06 Leo Grady System and method for multilabel random walker segmentation using prior models
CN102184422A (en) * 2011-05-15 2011-09-14 中科院成都信息技术有限公司 Average error classification cost minimized classifier integrating method
CN105740906A (en) * 2016-01-29 2016-07-06 中国科学院重庆绿色智能技术研究院 Depth learning based vehicle multi-attribute federation analysis method
CN106504255A (en) * 2016-11-02 2017-03-15 南京大学 A kind of multi-Target Image joint dividing method based on multi-tag multi-instance learning
CN107330444A (en) * 2017-05-27 2017-11-07 苏州科技大学 A kind of image autotext mask method based on generation confrontation network
CN107577785A (en) * 2017-09-15 2018-01-12 南京大学 A kind of level multi-tag sorting technique suitable for law identification

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111353515B (en) * 2018-12-21 2024-01-26 湖南工业大学 Multi-scale classification-based train wheel set tread damage classification and identification method
CN111353515A (en) * 2018-12-21 2020-06-30 湖南工业大学 Multi-scale grading-based classification and identification method for damage of train wheel set tread
CN109886335A (en) * 2019-02-21 2019-06-14 厦门美图之家科技有限公司 Disaggregated model training method and device
CN109886335B (en) * 2019-02-21 2021-11-26 厦门美图之家科技有限公司 Classification model training method and device
CN109993058A (en) * 2019-02-27 2019-07-09 北京大学 The recognition methods of road signs based on multi-tag classification
CN110070113B (en) * 2019-03-29 2021-03-30 广州思德医疗科技有限公司 Training method and device for training set
CN110070113A (en) * 2019-03-29 2019-07-30 广州思德医疗科技有限公司 A kind of training method and device of training set
CN109977902A (en) * 2019-04-03 2019-07-05 刘西 A kind of construction vehicle identification method based on deep learning
CN110163161A (en) * 2019-05-24 2019-08-23 西安电子科技大学 Multiple features fusion pedestrian detection method based on Scale invariant
CN110598849A (en) * 2019-08-26 2019-12-20 广州大学 HMET scattering parameter extraction method and system based on neural network and storage medium
CN110648655A (en) * 2019-09-11 2020-01-03 北京探境科技有限公司 Voice recognition method, device, system and storage medium
CN110648655B (en) * 2019-09-11 2022-03-01 北京探境科技有限公司 Voice recognition method, device, system and storage medium
CN110689081A (en) * 2019-09-30 2020-01-14 中国科学院大学 Weak supervision target classification and positioning method based on bifurcation learning
CN111507403A (en) * 2020-04-17 2020-08-07 腾讯科技(深圳)有限公司 Image classification method and device, computer equipment and storage medium
CN111626357A (en) * 2020-05-27 2020-09-04 北京微智信业科技有限公司 Image identification method based on neural network model
CN111626357B (en) * 2020-05-27 2021-11-23 北京东方通网信科技有限公司 Image identification method based on neural network model
CN111898467B (en) * 2020-07-08 2023-02-28 浙江大华技术股份有限公司 Attribute identification method and device, storage medium and electronic device
CN111898467A (en) * 2020-07-08 2020-11-06 浙江大华技术股份有限公司 Attribute identification method and device, storage medium and electronic device
CN111967364A (en) * 2020-08-11 2020-11-20 西交利物浦大学 Composite fault diagnosis method, device, electronic equipment and storage medium
CN111967364B (en) * 2020-08-11 2023-11-21 西交利物浦大学 Composite fault diagnosis method, device, electronic equipment and storage medium
CN112580750A (en) * 2020-12-30 2021-03-30 Oppo广东移动通信有限公司 Image recognition method and device, electronic equipment and storage medium
CN113449781A (en) * 2021-06-17 2021-09-28 上海深至信息科技有限公司 Generation method and system of thyroid nodule classification model
CN113449781B (en) * 2021-06-17 2023-04-07 上海深至信息科技有限公司 Generation method and system of thyroid nodule classification model
US11978280B2 (en) 2021-07-30 2024-05-07 Lemon Inc. Method and device for evaluating effect of classifying fuzzy attribute
CN113673482B (en) * 2021-09-03 2023-04-18 四川大学 Cell antinuclear antibody fluorescence recognition method and system based on dynamic label distribution
CN113673482A (en) * 2021-09-03 2021-11-19 四川大学 Cell antinuclear antibody fluorescence recognition method and system based on dynamic label distribution
CN115406852A (en) * 2021-12-28 2022-11-29 中山小池科技有限公司 Fabric fiber component qualitative method based on multi-label convolutional neural network

Also Published As

Publication number Publication date
CN108664924B (en) 2022-07-08

Similar Documents

Publication Publication Date Title
CN108664924A (en) A kind of multi-tag object identification method based on convolutional neural networks
US11195051B2 (en) Method for person re-identification based on deep model with multi-loss fusion training strategy
CN109961034B (en) Video target detection method based on convolution gating cyclic neural unit
CN111553193B (en) Visual SLAM closed-loop detection method based on lightweight deep neural network
CN113378632B (en) Pseudo-label optimization-based unsupervised domain adaptive pedestrian re-identification method
CN109325547A (en) Non-motor vehicle image multi-tag classification method, system, equipment and storage medium
CN101097564B (en) Parameter learning method, parameter learning apparatus, pattern classification method, and pattern classification apparatus
CN110276248B (en) Facial expression recognition method based on sample weight distribution and deep learning
CN109002755B (en) Age estimation model construction method and estimation method based on face image
CN112883839B (en) Remote sensing image interpretation method based on adaptive sample set construction and deep learning
CN105389326B (en) Image labeling method based on weak matching probability typical relevancy models
CN111461038A (en) Pedestrian re-identification method based on layered multi-mode attention mechanism
CN106845525A (en) A kind of depth confidence network image bracket protocol based on bottom fusion feature
CN110765960B (en) Pedestrian re-identification method for adaptive multi-task deep learning
CN117149944B (en) Multi-mode situation emotion recognition method and system based on wide time range
CN108879732A (en) Transient stability evaluation in power system method and device
KR102049331B1 (en) Apparatus and method for classifying images, and apparatus for training images for classification of images
Chen et al. Military image scene recognition based on CNN and semantic information
CN109948498A (en) A kind of dynamic gesture identification method based on 3D convolutional neural networks algorithm
CN113920472A (en) Unsupervised target re-identification method and system based on attention mechanism
CN103136540A (en) Behavior recognition method based on concealed structure reasoning
CN116310647A (en) Labor insurance object target detection method and system based on incremental learning
Yao Application of higher education management in colleges and universities by deep learning
CN116363712B (en) Palmprint palm vein recognition method based on modal informativity evaluation strategy
Banik et al. Multi-label object attribute classification using a convolutional neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant