CN109840530A - The method and apparatus of training multi-tag disaggregated model - Google Patents

The method and apparatus of training multi-tag disaggregated model Download PDF

Info

Publication number
CN109840530A
CN109840530A CN201711187395.0A CN201711187395A CN109840530A CN 109840530 A CN109840530 A CN 109840530A CN 201711187395 A CN201711187395 A CN 201711187395A CN 109840530 A CN109840530 A CN 109840530A
Authority
CN
China
Prior art keywords
matrix
sample
label
network
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711187395.0A
Other languages
Chinese (zh)
Inventor
刘晓阳
胡晓林
王月红
曹忆南
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Huawei Technologies Co Ltd
Original Assignee
Tsinghua University
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University, Huawei Technologies Co Ltd filed Critical Tsinghua University
Priority to CN201711187395.0A priority Critical patent/CN109840530A/en
Priority to PCT/CN2018/094400 priority patent/WO2019100724A1/en
Publication of CN109840530A publication Critical patent/CN109840530A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

This application provides training multi-tag disaggregated model method and apparatus, can dynamic learning characteristics of image, so that feature extraction network is more adapted to mission requirements, and multi-tag good classification effect.N sample and label matrix Y corresponding with the n sample are determined this method comprises: concentrating in training datac*n, the label matrix Yc*nIn element yi*jIndicate i-th of sample whether include the instruction of j-th of label object, c indicates the number of relevant to sample label;The eigenmatrix X of the n sample is extracted using feature extraction networkd*n;The eigenmatrix X is obtained using Feature Mapping networkd*nPrediction label matrixThe prediction label matrixIn elementIndicate that i-th of sample includes the confidence level of the object of j-th of label instruction;According to the label matrix Yc*nWith the prediction label matrix, to the weighting parameter Z, the Feature Mapping matrix Mc*dIt is updated, the training multi-tag disaggregated model.

Description

The method and apparatus of training multi-tag disaggregated model
Technical field
This application involves computer fields, and the mould more particularly to the training multi-tag in computer field is classified The method and apparatus of type.
Background technique
With the promotion of the process performance of smart phone, more and more applications propose requirement to the identification of image.Than It such as,, can be to it if smart phone can accurately identify the object in coverage during with mobile phone photograph Color, shape carries out targetedly operation, to improve shooting effect.And in the machine learning of intelligence system, in image The training that is identified of object also just at a very important aspect.Usually, machine learning is for largely There is image that label is set for object wherein included, identification parameter is constantly then adjusted by self evolution of computer, by Gradually improve the recognition accuracy to object.
Due to the complexity and ambiguity of objective object itself, real-life many objects may simultaneously with multiple classes Distinguishing label is related.In order to preferably embody multi-semantic meaning possessed by practical object, it is often used a sub-set of tags appropriate (including multiple relevant semantic labels) describes the object, and which forms so-called multi-tag classification problems.At this moment, each sample This all corresponds to the respective labels subclass being made of multiple labels, the target of study be exactly be that unknown sample predicts that it is corresponding Sub-set of tags.
In the classification of actual multi-tag, a series of training datas can be given first, here this series of training data group At set be properly termed as training dataset.But since the label that training data is concentrated is different people mark, or mark Some objects are had ignored when label, leading to label may be to have missing, therefore can be by the label concentrated to training data Completion is carried out to improve the accuracy of multi-tag classification.Completion is carried out there are many method to known label in multi-tag classification, In one is by the order of nuclear norm constrained forecast label matrix, and the loss function by minimizing multi-tag classification calculates This feature mapping matrix obtains the prediction label matrix of low-rank, realizes label completion, and then improve the performance of multi-tag classification. But this method needs first to extract the feature of image, then according to the feature calculation Feature Mapping matrix of image.It is being extracted After the feature of image, the feature of the image is fixed, because being unable to dynamically learn input picture according to label Characteristic information.
Summary of the invention
The application provides a kind of method and apparatus of trained multi-tag disaggregated model, can dynamic learning characteristics of image, make Feature extraction network more adapts to mission requirements, and multi-tag good classification effect.
In a first aspect, providing a kind of method of trained multi-tag disaggregated model, comprising:
It is concentrated in training data and determines n sample and label matrix Y corresponding with the n samplec*n, the label square Battle array Yc*nIn element yi*jIndicate i-th of sample whether include j-th of label instruction object, c indicate with the training data The number of the relevant label of the sample of concentration.
The eigenmatrix X of the n sample is extracted using feature extraction networkd*n, wherein the feature extraction network tool There are weighting parameter Z, d to indicate the eigenmatrix Xd*nCharacteristic dimension.
Here, feature extraction network, which can be any one, can extract the neural network of characteristics of image, such as can be Convolutional neural networks or multi-layer perception (MLP) etc., the embodiment of the present application does not limit this.Wherein, the weight of feature extraction network can be with It is expressed as Z, specifically, Z may include multiple weight matrixs.The parameter of weight matrix can be generated with random initializtion, Huo Zheke Using the model parameter of pre-training.Here, the model parameter of pre-training refers to the parameter of trained model, such as Vgg16 network trained model parameter on ImageNet data set.
The eigenmatrix X is obtained using Feature Mapping networkd*nPrediction label matrixThe prediction label matrixIn elementIndicate that i-th of sample includes the confidence level of the object of j-th of label instruction, wherein the Feature Mapping net The weight matrix of network is the Feature Mapping matrix M of low-rankc*d, Mc*dIt can indicate the characteristic attribute and class in multi-tag disaggregated model Associated weight between distinguishing label, initial value can generate at random.As an example, Feature Mapping network can be weight matrix For the Feature Mapping matrix M of low-rankc*dMapping network, such as can be full articulamentum.
Specifically, Feature Mapping network can be expressed as FCM.The eigenmatrix X of feature extraction network outputd*nIt can be defeated Enter to FCM, then by FCM by the eigenmatrix X of inputd*nIt is mapped to prediction label space, obtains prediction label matrixI.e. Have:
According to the label matrix Yc*nWith the prediction label matrixTo the weighting parameter Z, the Feature Mapping Matrix Mc*dIt is updated, the training multi-tag disaggregated model.
Wherein, n, c, i, j and d are positive integer, and the value range of i is 1 to n, and the value range of j is 1 to c.
Therefore, the nerve network system provided by the embodiment of the present application can directly train model from input data, Without additional intermediate steps, i.e., the nerve network system is a nervous system end to end.Here, end to end Advantage is that feature extraction, Feature Mapping matrix and low-rank label correlation matrix can optimize simultaneously, that is to say, that the application is real Applying example can make feature extraction network more adapt to mission requirements, and multi-tag good classification effect with dynamic learning characteristics of image.
Optionally, the Feature Mapping network of the low-rank includes the first sub- mapping network and the second sub- mapping network, described The Feature Mapping network of low-rank, the first sub- mapping network and the second sub- mapping network have following relationship:
Wherein, the weight matrix of the described first sub- mapping network isThe weight matrix of the second sub- mapping network is Hc*r, here, in order to guarantee Mc*dAnd Hc*rLow-rank, can be set r be positive integer and r≤min (d, c).
In a specific embodiment, the first sub- mapping network can be for weight matrixFull articulamentum, Two sub- mapping networks can be H with weight matrixc*rFull articulamentum,And Hc*rInitial value can generate at random.
In other words, in the embodiment of the present application, completion can be carried out to label matrix by way of matrix low rank decomposition, I.e. by prediction label matrixCarry out low-rank decomposition, it may be assumed that
Here, made by the way that r≤min (d, c) is arrangedAnd Hc*rLow-rank, due to the square obtained after two matrix multiples Rank of matrix can make less than any one rank of matrix in two matrixes(i.e. Mc*d) low-rank, and then make ?I.e.Low-rank.Here, r can take optimal value by repeatedly training.
That is, the embodiment of the present application can pass through preset Feature Mapping matrix M (i.e. Mc*d) by X (i.e. Xd*n) mapping Obtain prediction label matrix(i.e.), i.e.,BecauseOrder be less than or equal to the order of M or X, so doing low-rank point to M Guarantee while solution can make M low-rankLow-rank, therefore low-rank decomposition can also be done to M, i.e., above-mentioned (2) formula in this way may be used Be equivalent to byThe form of two low dimensional matrix multiples has been resolved into, and then has been guaranteedLow-rank.
Optionally, according to the label matrix Yc*nWith the prediction label matrixTo the weighting parameter Z, described Feature Mapping matrix Mc*dIt is updated, comprising:
Determine the prediction label matrixWith the label matrix Yc*nBetween Euclidean distance loss function, the loss Following (3) formula of the expression formula of function:
Alternatively, following (4) formula of the expression formula of the loss function:
Then, according to the Euclidean distance loss function, to the weighting parameter Z, the weight matrixAnd Hc*rInto Row updates.
Optionally, described according to the Euclidean distance loss function, to the weighting parameter Z, the weight matrixWith Hc*rIt is updated, comprising:
By the sum of the Euclidean distance loss function and regular terms, it is determined as the majorized function L of the n samplen, In, the regular terms is for constraining the weighting parameter Z, the weight matrixAnd Hc*r, LnExpression formula such as (7) formula or (8) shown in formula:
Wherein, majorized function LnFirst item be above-mentioned loss function εn, Section 2 is regular terms, and the regular terms is for about Shu Suoshu weighting parameter Z, the weight matrixAnd Hc*r, prevent over-fitting.
Here it is possible to minimize loss function L using error backpropagation algorithmn, by the value of the majorized function Corresponding weighting parameter Z is as updated weighting parameter Z when minimum, corresponding when by the value minimum of the majorized function Weight matrixAs updated weight matrixCorresponding weight when by the value minimum of the majorized function Matrix Hc*rAs updated weight matrix Hc*r
Next, it is determined whether reaching stop condition.
Here, stop condition are as follows: LnNo longer decline or fall is less than preset threshold value, or reaches maximum training time Number.The repetition training if not reaching, until reaching stop condition.In the embodiment of the present application, all pictures are all inputted a calculation Make one wheel of training, it usually needs several wheels of training.
Optionally, described that the label matrix Y for determining n sample and the n sample is concentrated in training datac*n, comprising:
Determine training dataset, the training data concentrate include D sample and with each sample in the D sample Label vector, wherein the element y in the label vector of each samplejIndicate whether each sample includes j-th of mark Sign the object of instruction, wherein D is the positive integer not less than n;
N sample is randomly selected from training data concentration, and generates the label matrix Y of the n samplec*n, described Label matrix Yc*nIncluding the corresponding label vector of each sample in the n sample.
Therefore, in the embodiment of the present application, it is not necessary to disposably input entire training dataset and be calculated, and only need in batches Secondary input picture is calculated, therefore the embodiment of the present application can input entire data set in batch and be trained.Also It is to say, in the embodiment of the present application, model can be trained by the partial data that multiple batches of ground input data is concentrated, In, the data inputted every time, which can be, to be randomly selected from the picture sample not inputted in data set.Due to training dataset A large amount of sample is generally included, therefore the embodiment of the present application can reduce training pattern by inputting training dataset in batches In the process to the occupancy of resource, the demand during training pattern to memory source is greatly reduced, can effectively be solved big The computational problem of low-rank label correlation matrix under scale data.
Optionally, further includes: the fisrt feature matrix of first sample is extracted using the feature extraction network, wherein institute It states first sample and is not belonging to the n sample;
The first prediction label matrix of the fisrt feature matrix is obtained using first mapping network, described first is pre- Survey the confidence level that first sample described in the element representation in label matrix includes the object of j-th of label instruction.
Specifically, in test phase, only test picture need to be input to the spy in the neural network model after the completion of training Sign extracts network, extracts the fisrt feature matrix of the test picture using the feature extraction network, and by the fisrt feature square Battle array is input to Feature Mapping network (can specifically include FCW and FCH), is obtained using Feature Mapping network and exports described first The prediction label matrix of eigenmatrix, test described in the element representation in the prediction label matrix include j-th of label instruction Object confidence level.Here, test picture can be one or more pictures, and can be not belonging to training dataset.
Second aspect provides a kind of device of trained multi-tag disaggregated model, and described device is for executing above-mentioned first party Method in any possible implementation of face or first aspect.Specifically, the apparatus may include for executing first The module of method in any possible implementation of aspect or first aspect.
The third aspect, provides a kind of device of trained multi-tag disaggregated model, and described device includes memory and processor, For storing instruction, the processor is used to execute the instruction of the memory storage to the memory, and to the storage The execution of the instruction stored in device is so that the processor executes any possible implementation of first aspect or first aspect In method.
Fourth aspect provides a kind of computer readable storage medium, finger is stored in the computer readable storage medium It enables, when described instruction is run on computers, so that computer executes any possible reality of first aspect or first aspect Method in existing mode.
5th aspect, provides a kind of computer program product comprising instruction, when the computer program product is in computer When upper operation, so that computer executes the method in any possible implementation of first aspect or first aspect.
Detailed description of the invention
Fig. 1 shows the schematic diagram of single labeling and multi-tag classification problem.
Fig. 2 shows a kind of schematic flows of the method for trained multi-tag disaggregated model provided by the embodiments of the present application Figure.
Fig. 3 shows a kind of schematic diagram of multi-tag disaggregated model provided by the embodiments of the present application.
Fig. 4 shows a kind of schematic diagram of multi-tag disaggregated model provided by the embodiments of the present application.
Fig. 5 shows a kind of schematic block diagram of the device of trained multi-tag disaggregated model provided by the embodiments of the present application.
Fig. 6 shows the schematic frame of the device of another training multi-tag disaggregated model provided by the embodiments of the present application Figure.
Specific embodiment
Below in conjunction with attached drawing, the technical solution in the application is described.
Fig. 1 shows the schematic diagram of single labeling and multi-tag classification problem.As shown in figure 1 shown in (a), single labeling Often assume that sample corresponds only to a class label, that is, has unique semantic meaning.Then this hypothesis is in many reality In the case of may and it is invalid, particularly contemplate the existing semantic diversity of objective objects itself, object be likely to simultaneously with Multiple and different class labels is related.Therefore in multi-tag problem, as shown in figure 1 shown in (b), multiple relevant classifications are often used Label describes semantic information corresponding to each object, for example, each image may correspond to multiple semantic labels simultaneously, such as " meadow ", " sky " and " sea ", per song segment may contain there are many mood, such as " pleasure " and " light ".
In multi-tag classification problem, a series of training datas can be given first, and this series of training data forms here Set is properly termed as training dataset.By learning given training data, its corresponding label can be predicted for unknown sample Subset.Here, training dataset can correspond to a tag set, may include related to the training data in the tag set The different classes of labels of c, c is positive integer.Training dataset may include D sample and corresponding label of each sample Collection, wherein D is positive integer.It is understood that sub-set of tags here is a subset of the tag set.That is, passing through The multiple samples and the corresponding sub-set of tags of each sample that the given training data of study is concentrated, can predict the mark of unknown sample Bamboo slips used for divination or drawing lots collection.
In the embodiment of the present application, sub-set of tags can be expressed as label vector.In other words, the label vector of sample can be with Indicate which label sample has or belong to which type.For example, the label vector of piece image is [0 1001 0], Then show to share 6 kinds of classifications, wherein each element in the label vector represents a kind of classification or a label, and 0 indicates image In without this kind of or this label, have this kind of or this label in 1 expression image.Since there are two 1 marks for the label vector Label, then it represents that there are two types of objects in the image, are belonging respectively to the second class and the 5th class.In this way, the D sample that training data is concentrated In each sample can correspond to a label vector yj, indicate the sample whether include j-th of label instruction object, here The value range of j is 1 to c.It should be understood that the embodiment of the present application in, sample whether include j-th of label instruction object i.e. sample It whether include j-th of label.
In this way, the label vector for all or part of sample that training data is concentrated will form a label matrix Y:
In addition, prediction label vector is the output of multi-tag classifier, dimension is identical as label vector, represents multi-tag Prediction of the classifier to the image generic.The value of the element of prediction label vector is real value, if the real value is more than given As soon as threshold value be otherwise not belonging to the category then the corresponding position of the element belongs to respective classes.For example, prediction label Vector is [0.7 0.2 0.1 0.8 1.0 0.0], and the number on each is compared by threshold value 0.5 with threshold value, is greater than threshold Value, which is then equivalent to, belongs to the category.The classification predicted in this way is the first kind, the 4th class and the 5th class.If the prediction label to Measuring corresponding label vector is [1 00101 0], then the prediction label vector is completely correct.
It is each sample in data in practical problem especially in the case where a large amount of class labels involved in data It is often extremely difficult all to provide its corresponding complete tag information.Therefore, label letter corresponding to the sample that training data is concentrated Breath is likely to incomplete.That is, sample does not include certain label and does not represent practical feelings in the label matrix of data Sample is uncorrelated to the label under condition.Therefore, it is necessary to concentrate existing data by training data, label matrix is mended It entirely, include then richer label information by this to obtain the prediction label matrix comprising richer label information Prediction label matrix can more accurately predict the label information of unknown sample.
The prior art needs first to extract the feature of image, then according to the spy of image when carrying out completion to label matrix Sign calculates the Feature Mapping matrix of low-rank.After being extracted the feature of image, the feature of the image is fixed, because without The characteristic information of input picture can dynamically be learnt according to label.Based on this, the embodiment of the present application devises a kind of for more The neural network of labeling can realize multi-tag point by learning characteristic mapping matrix and optimization feature extraction network Class algorithm.
Nerve network system is a kind of intelligentized identifying system, adds up training result by way of repetition training, To improve the recognition capability to various target objects or sound.Convolutional neural networks be neural network development main flow direction it One.Convolutional neural networks generally comprise convolutional layer (Convolutional Layer), correct linear unit (Rectified Linear Units, ReLU) layer, pond (Pooling) layer and full connection (Fully Connect, FC) layer.Wherein, convolution Layer, ReLU layers and Pooling layers may alternately be repeated several times setting.
Convolutional layer can be considered as the core of convolutional neural networks, and when being used for image recognition, input terminal receives image Data, for being identified by filter image.Here image data can be the image conversion knot that video camera photographed Fruit is also possible to the processing result of the front layer of convolutional layer.Usual image data is three-dimensional pattern matrix, such as 32x32x3, In, 32x32 is the two-dimensional for the image that image data represents, i.e., wide and high, depth value 3 here is then because image is usual It is divided into green, red, blue three data channel.Multiple filters are equipped in convolutional layer, it is special that different filters corresponds to different images Sign (boundary, color, shape etc.) is scanned the image data of input according to certain step-length.It is arranged in different filters There is different weight matrix, the weight matrix is that neural network generates in learning process for specific image features.Often The each region for clapping scan image of one filter, obtaining a three-dimensional input matrix, (MxNx3, M and N are determined The size of scanning area), input matrix and weight matrix are made dot product by convolutional network, obtain an end value, then can be with spy Fixed step size scans next region, for example, traversing two lattice.After a filter scans through all areas according to particular step size, End value can constitute a two-dimensional matrix;And after all filters are completed to scan, end value will constitute a three-dimensional matrice As the output of current convolutional layer, the different depth layer of this three-dimensional matrice respectively corresponds the scanning result of a filter (i.e. The two-dimensional matrix constituted after each filter scan).
The output of convolutional layer can be sent to ReLU layers again and process (the numberical range progress by max (0, x) function to output Limit), and Pooling layers are sent to by down-sampling size reduction, before being sent to FC layers, image data can may also pass through Multiple convolutional layers, with to characteristics of image carry out profound identification (such as first time convolutional layer only to the contour feature of image into Row identification, second of convolutional layer start identification icon etc.), finally enter FC layers.It is similar but slightly different with convolutional layer, FC layers It is weight processing to be made to input data by multiple filters, but FC layers obtain each filter unlike the filter of convolutional layer Different zones are scanned by the displacement of each bat like that, but the disposably all areas of the image data of scanning input, Then operation is carried out with weight matrix obtain an end value.As soon as final FC layer export be a 1x1xN matrix, in fact It is a data sequence, each of this data sequence corresponds to different target objects, and value thereon can be considered these Score value existing for object target.In convolutional layer and FC layers, weight matrix can be all used, neural network can be tieed up by self-training Protect a variety of weight matrix.
The method of the training multi-tag disaggregated model of the embodiment of the present application is discussed in detail below in conjunction with Fig. 2 and Fig. 3.
Fig. 2 shows a kind of schematic flows of the method for trained multi-tag disaggregated model provided by the embodiments of the present application Figure.It should be understood that the step of Fig. 2 shows the methods of trained multi-tag disaggregated model or operation, but these steps or operation are only The deformation of other operations or each operation in Fig. 2 can also be performed in example, the embodiment of the present application.In addition, each in Fig. 2 Step can be executed according to the different sequences presented from Fig. 2, and it is possible to not really want to execute all operationss in Fig. 2.
Fig. 3 shows a kind of schematic diagram of multi-tag disaggregated model 300 provided by the embodiments of the present application.Multi-tag classification Model 300 is specially nerve network system.The multi-tag disaggregated model 300 includes feature extraction network 301, Feature Mapping network 302 and processing unit 305, wherein Feature Mapping network 302 may include FCW 303 and FCH 304.It should be understood that Fig. 3 is shown Multi-tag disaggregated model 300 be only example, the embodiment of the present application can also include each in other modules or unit or Fig. 3 The deformation of a module or unit.
It should be noted that multi-tag classification method can be applied to image labeling, image recognition, sound knowledge in the embodiment of the present application Not, the multiple fields such as text classification, specifically, the sample concentrated of corresponding training data can for image, sound, document etc., The embodiment of the present application does not limit this.For convenience, it will hereafter be carried out for using image pattern to carry out image recognition Description, but this can't limit the forecast scheme configuration of the embodiment of the present application.
210, initialize the weight of multi-tag disaggregated model 200.
Feature extraction network 301, Feature Mapping in weight, that is, initialization system of initialization multi-tag disaggregated model 200 The weight of network (i.e. FCW303 and FCH 304).
Here, feature extraction network 301, which can be any one, can extract the neural network of characteristics of image, such as can be with For convolutional neural networks or multi-layer perception (MLP) etc., the embodiment of the present application does not limit this.Wherein, the power of feature extraction network 301 Value can be expressed as Z, specifically, Z may include multiple weight matrixs.The parameter of weight matrix can be generated with random initializtion, Or the model parameter of pre-training can be used.Here, the model parameter of pre-training refers to the ginseng of trained model Number, as vgg16 network on ImageNet data set trained model parameter.
In addition, Feature Mapping network 302 can be the Feature Mapping matrix M that weight matrix is low-rankc*dMapping network, It such as can be full articulamentum, wherein Mc*dIt can indicate between the characteristic attribute and class label in multi-tag disaggregated model Associated weight, initial value can generate at random.In a specific embodiment, Feature Mapping network 302 may include FCW303 and FCH 304, wherein FCW 303 indicates that weight matrix isFull articulamentum, FCH 304 indicate weight matrix For Hc*rFull articulamentum,And Hc*rInitial value can generate at random.Here, in order to guarantee Mc*dAnd Hc*rLow-rank Property, r≤min (d, c) can be set.
220, input n width picture.
Due to the characteristic of neural network, it is not necessary to disposably input entire training dataset and be calculated, and only need in batches Secondary input picture is calculated, therefore the embodiment of the present application can input entire data set in batch and be trained.Also It is to say, in the embodiment of the present application, model can be trained by the partial data that multiple batches of ground input data is concentrated, In, the data inputted every time, which can be, to be randomly selected from the picture sample not inputted in data set.Due to training dataset A large amount of sample is generally included, therefore the embodiment of the present application can reduce training pattern by inputting training dataset in batches In the process to the occupancy of resource.
At this moment, the number that a batch is input to the sample of multi-tag disaggregated model 300 can be n.When sample is figure When piece, which can be expressed as image_n, and more specifically, and image_n can be the D sample from training dataset The n picture randomly selected in this, also, the value of n can be much smaller than D.Specifically, the size of n can be according to more marks The ability for signing disaggregated model 300 determines.For example, n can if the data-handling capacity of the multi-tag disaggregated model 300 is stronger With the bigger of setting, to shorten the time of training pattern.In another example if the data processing of the multi-tag disaggregated model 300 Ability is weaker, then n can be set smaller, to reduce resource consumed by training pattern.In this way, the embodiment of the present application can The value of n is neatly set according to the data-handling capacity of multi-tag disaggregated model 300.
Also, label matrix corresponding to the n sample can be expressed as Yc*n, label matrix Yc*nIn element yi*jTable Show i-th of sample whether include the instruction of j-th of label object, the value range of i is 1 to n here, the value range of j be 1 to c.Specifically, the description of label matrix may refer to description above, to avoid repeating, which is not described herein again.
In the embodiment of the present application, training data can be input to multi-tag disaggregated model 300 shown in Fig. 3.Specifically , the label matrix Y of the n picture and the n picture that training data can be concentratedc*nIt is separately input into multi-tag classification Model 300.
230, extract the feature of picture.
Specifically, n picture can be input to feature extraction network 301, feature extraction network 301 passes through convolution Layer, activation primitive layer, Pooling layers, full articulamentum, Batchnorm layers of effect, can extract the feature of the n picture, and Export eigenmatrix Xd*n.Wherein, d is positive integer and indicates the eigenmatrix Xd*nCharacteristic dimension.
240, according to the prediction label matrix of the feature calculation picture of picture.
Specifically, the eigenmatrix X that feature extraction network 301 exportsd*nFeature Mapping network 302 can be input to.Due to The weight matrix of Feature Mapping network is the Feature Mapping matrix M of low-rankc*d, Mc*dIt can indicate the spy in multi-tag disaggregated model The associated weight between attribute and class label is levied, therefore Feature Mapping network 302 can be by the eigenmatrix X of inputd*nMapping To prediction label space, prediction label matrix is obtainedHave:
Here, prediction label matrixIt can be the label matrix comprising richer label information, each element thereinIndicate that i-th of sample includes the confidence level of the object of j-th of label instruction.Therefore, prediction label matrix can be claimedTo mend Full label matrix claims label matrix Yc*nTo there is the label matrix of missing.
It should be noted that the label in label matrix is not independent from each other in practical problem, but semantic phase It closes.Such as sheep and grass a possibility that occurring in a width picture it is very big, a possibility that mountain and sky occur together, is also very big, and A possibility that sheep and office occur together very little, and this correlation can be used to improve the accuracy of multi-tag classification. It follows that completion label matrixIn label between there is correlation, i.e.,It is low-rank, therefore can basis The low-rank structure of matrix is by Yc*nIt obtainsAnd this process is properly termed as the completion of label matrix.
In the embodiment of the present application, completion can be carried out to label matrix by way of matrix low rank decomposition, i.e., it will prediction Label matrixCarry out low-rank decomposition, it may be assumed that
That is, the Feature Mapping network of the low-rank may include the first sub- mapping network in the embodiment of the present application With the second sub- mapping network, the Feature Mapping network of the low-rank, the first sub- mapping network and the second son mapping net Network has following relationship:
Wherein, the weight matrix of the described first sub- mapping network isThe weight matrix of the second sub- mapping network is Hc*r, here, in order to guarantee Mc*dAnd Hc*rLow-rank, can be set r be positive integer and r≤min (d, c).
In a specific embodiment, the first sub- mapping network can be for weight matrixFull articulamentum, indicate For FCW, the second sub- mapping network can be that weight matrix is Hc*rFull articulamentum, be expressed as FCH,And Hc*rInitial value It can generate at random.
Here, made by the way that r≤min (d, c) is arrangedAnd Hc*rLow-rank, due to the matrix obtained after two matrix multiples Order less than any one rank of matrix in two matrixes, therefore can make(i.e. Mc*d) low-rank, so thatI.e.Low-rank.Here, r can take optimal value by repeatedly training.
In other words, the embodiment of the present application can pass through preset Feature Mapping matrix M (i.e. Mc*d) by X (i.e. Xd*n) mapping Obtain prediction label matrix(i.e.), i.e.,BecauseOrder be less than or equal to the order of M or X, so doing low-rank point to M Guarantee while solution can make M low-rankLow-rank, therefore low-rank decomposition can also be done to M, i.e., above-mentioned (2) formula in this way may be used Be equivalent to byThe form of two low dimensional matrix multiples has been resolved into, and then has been guaranteedLow-rank.
250, calculation optimization function.
Specifically, in the embodiment of the present application, preset matrix H can be passed throughc*rWithTo replace default matrix M. But due toIt is using ready-made matrix Hc*rWithIt obtains, therefore prediction label matrix in this wayIt is inaccurate, at this time It needs to be compareed in learning process with the label matrix Y of reference, to learn to update matrix Hc*rWith
At this point, processing unit 305 can be according to the label matrix Yc*nWith the prediction label matrixTo the power Value parameter Z, the Feature Mapping matrix Mc*dIt is updated, with the training multi-tag disaggregated model 300.
Specifically, processing unit 305 can determine the prediction label matrixWith the label matrix Yc*nBetween Euclidean distance loss function, effect are constraintsIt is allowed to and Yc*nIt is close, following (3) formula of the expression formula of the loss function:
Alternatively, following (4) formula of the expression formula of the loss function:
Here, for ease of description, it is omittedHc*rSubscript and subscript.Wherein,It is the Frobenius model of matrix Number, matrix Am*nFrobenius norm is defined as:
Wherein, AijFor the element of matrix A, i.e. Euclidean distance loss function.
In addition, the P in formula (4)ΩFor projection operator, that is, the element observed remains unchanged, unobservable element value It is 0, effect is exactly that the element observed only is allowed to participate in calculating.Concrete form are as follows:
For example assumeY=[1 00? 0 1],? the element being missing from, that ?I.e.? position exists 0 is set to when calculating.In this way,It is not involved in calculating (being considered as 0) with the element on the position in Y, it can be to avoid scarce The element of mistake causes the value of loss function bigger than normal, and then improves the accuracy calculated.
It is possible to further be determined as the loss function of the n sample for the sum of above-mentioned loss function and regular terms Ln.Here, loss function LnIt is referred to as majorized function Ln, specifically, LnExpression formula as shown in (7) formula or (8) formula:
Wherein, majorized function LnFirst item be above-mentioned loss function εn, Section 2 is regular terms, and the regular terms is for about Shu Suoshu weighting parameter Z, the weight matrixAnd Hc*r, prevent over-fitting.
260, weighting parameter is updated using error inverse algorithms.
Error backpropagation algorithm is a kind of method for multilayer neural network training, using gradient descent method as base Plinth carries out study update to every layer of neural network of weight by optimizing loss function.
Specifically, can use error backpropagation algorithm, loss function L is minimizedn, by taking for the majorized function Value weighting parameter Z corresponding when minimum is right by the value minimum when institute of the majorized function as updated weighting parameter Z The Feature Mapping matrix M answeredc*dAs updated Feature Mapping matrix Mc*d
WhenWhen, then have: corresponding weight matrix when by the value minimum of the majorized function As updated weight matrixCorresponding weight matrix H when by the value minimum of the majorized functionc*rAs update Weight matrix H afterwardsc*r
To use error backpropagation algorithm, derivation is carried out to the variable in (7) formula below.With one width picture of input, just Then item uses l2For norm.
Remember L1For the majorized function of a width picture, then have:
Wherein, the l of square corresponding vector of the Frobenius norm of matrix2Square of norm.
It is right belowHc*rEach element derivation obtain:
Wherein, hkjFor matrix Hc*rElement, wjiFor matrixElement, xiFor vector xdVector, pjFor vectorElement,For vectorElement, yj/ykFor vector ycElement, xd、prycRespectively matrix Xd*n、Pr*nYc*nColumn vector.The reversed derivation of error of feature extraction network weight Z can be passed throughTransmitting obtains. Then Hc*rWithElement update are as follows:
It is that this updates obtained value,It is the value updated last time, wjiIt is similar therewith, η1、η2It is H respectivelyc*r WithLearning rate, for controlling renewal rate.The update of feature extraction network portion weight Z is similar.
It thus may learn weight Z, the Feature Mapping matrix of feature extraction networkIt supplies scarce Lose-submission label promote the ability of multi-tag classification.
270, judge whether to reach stop condition.
Here, stop condition are as follows: LnNo longer decline or fall is less than preset threshold value, or reaches maximum training time Number.If do not reached, step 220 is repeated to 260, until reaching stop condition.It is in the embodiment of the present application, all pictures are all defeated Entering one time can be regarded as one wheel of training, it usually needs several wheels of training.
After the completion of training, in test phase, 220 and 230 only need to be executed, i.e., test picture are input to the neural network mould Feature extraction network in type, extracts the fisrt feature matrix of the test picture using the feature extraction network, and by this One eigenmatrix is input to FCM, and the prediction label matrix of the fisrt feature matrix, the prediction are obtained and exported using FCM The confidence level of object of the test described in element representation in label matrix comprising the instruction of j-th of label.Here, test picture can be with For one or more pictures, and training dataset can be not belonging to.
And specifically, to the single predicted vector of prediction label matrixFrom the point of view of, by rightIt processes and this can be obtained One or more classifications belonging to picture, such asSome or some element values be greater than preset threshold value and indicate the picture It is equipped with class label in an element or multiple element corresponding positions, which belongs to this kind of or several classes.Here, preset Threshold value can be that 0.5 or other numerical value, the embodiment of the present application do not limit this.
Therefore, the nerve network system provided by the embodiment of the present application can directly train model from input data, Without additional intermediate steps, i.e., the nerve network system is a nervous system end to end.Here, end to end Advantage is that the weighting parameter of feature extraction network and Feature Mapping matrix can optimize simultaneously, that is to say, that the embodiment of the present application Feature extraction network can be made more to adapt to mission requirements, multi-tag good classification effect with dynamic learning characteristics of image.
In addition, the embodiment of the present application can utilize the box counting algorithm Feature Mapping matrix of picture sample in batch, It, i.e., need not be disposably with the figure of whole samples without disposably using the characteristics of image of entire data set to be calculated as input As feature is trained, the demand during training pattern to memory source is greatly reduced, can effectively be solved extensive The computational problem that multi-tag is classified under data.
Fig. 4 shows a kind of schematic diagram of multi-tag disaggregated model 500 provided by the embodiments of the present application.The model 500 Feature extraction network portion uses VGG16 network, and by the Dropout after the full articulamentum of the penultimate of VGG16 network The output of layer is as eigenmatrix X.In addition, the weighting parameter Z of feature extraction network uses the training on ImageNet data set Good weighting parameter, then finely tunes that (fine tuning refers to the several layers of weight in fixed front or only carries out the adjustment of very little, instructs completely to it Practice the last layer or two-tier network).The initial value of weight matrix H and W can be initialized using Gaussian Profile, and H and W Value will train completely.Regular terms can use Frobenius norm.
Specifically, the weight of feature extraction network VGG16 (removing the last one full articulamentum) uses in training The weight of pre-training on ImageNet data set.
The RGB triple channel picture image_n that n width pixel size is 224*224 is input in VGG16 network, here 1≤ N≤N, N are the quantity of picture in training set, and picture size can be expressed as the four-matrixes such as n*C*h*w or h*w*C*n, wherein C is port number (RGB image 3), and h is picture height (224 pixel), and w is picture width (224 pixel).Picture is by multiple After the operation such as convolution, activation, Pooling, image characteristic matrix is obtained using two full articulamentums and Dropout layers X4096*n
X4096*nIt is respectively by weight matrixAnd Hc*rFull articulamentum (FCW 503 and FCH 504), predicted Label matrix
Processing unit 505 is according to label matrix Yc*nWith prediction label matrixObtain majorized function:
Then, using error backpropagation algorithm, above-mentioned majorized function is minimized, updates weighting parameter Z, weight matrixAnd Hc*r.Specifically optimization process may refer to described above, and to avoid repeating, which is not described herein again.
Having updated weighting parameter Z, weight matrixAnd Hc*rLater, judge whether to reach stop condition, not reach such as Step is then repeated, until reaching stop condition.Specifically, stop condition may refer to it is described above, to avoid repeating, this In repeat no more.
After the completion of training, test picture can be input to feature extraction network 501, and feature extraction network is extracted The feature of picture be input to FCW 503 and FCH 504, prediction label matrix is obtained by FCW 503 and FCH 504.
It should be noted that the structure of feature extraction network can be replaced using other networks, such as in the embodiment of the present application AlexNet, GoogleNet, ResNet and customized network etc., the embodiment of the present application does not limit this.The layer of feature output Several convolutional layers or full articulamentum can also be increased and decreased using the output of a certain layer of above-mentioned network on the basis of the above.Separately Outside, the embodiment of the present application can also use different regularization terms.
Therefore, the nerve network system provided by the embodiment of the present application can directly train model from input data, Without additional intermediate steps, i.e., the nerve network system is a nervous system end to end.Here, end to end Advantage is that the weighting parameter of feature extraction network and Feature Mapping matrix can optimize simultaneously, that is to say, that the embodiment of the present application Feature extraction network can be made more to adapt to mission requirements, multi-tag good classification effect with dynamic learning characteristics of image.
In addition, the embodiment of the present application can utilize the box counting algorithm Feature Mapping matrix of picture sample in batch, It, i.e., need not be disposably with the figure of whole samples without disposably using the characteristics of image of entire data set to be calculated as input As feature is trained, the demand during training pattern to memory source is greatly reduced, can effectively be solved extensive The computational problem that multi-tag is classified under data.
It should be noted that the embodiment of the present application does not limit special product form, the side of the multi-tag classification of the embodiment of the present application Method can be deployed on general computer node.The multi-tag disaggregated model of Primary Construction can be stored in harddisk memory In, algorithm is run by processor and memory, existing training dataset is learnt, the multi-tag disaggregated model is obtained. The label of unknown sample can be predicted by leading to the multi-tag disaggregated model, and prediction result is stored in harddisk memory, be realized Completion is carried out to existing tally set, and the corresponding label of unknown sample is predicted.
Fig. 5 shows a kind of schematic frame of the device 600 of trained multi-tag disaggregated model provided by the embodiments of the present application Figure.Device 600 includes determination unit 610, extraction unit 620, acquiring unit 630 and updating unit 640.
Determination unit 610 determines n sample and label square corresponding with the n sample for concentrating in training data Battle array Yc*n, the label matrix Yc*nIn element yi*jIndicate i-th of sample whether include j-th of label instruction object, c table Show the number of label relevant to the sample that the training data is concentrated.
Extraction unit 620, for extracting the eigenmatrix X of the n sample using feature extraction networkd*n, wherein institute Stating feature extraction network indicates the eigenmatrix X with weighting parameter Z, dd*nCharacteristic dimension.
Acquiring unit 630, for obtaining the eigenmatrix X using Feature Mapping networkd*nPrediction label matrix The prediction label matrixIn elementIndicate that i-th of sample includes the confidence level of the object of j-th of label instruction, In, the weight matrix of the Feature Mapping network is the Feature Mapping matrix M of low-rankc*d
Updating unit 640, for according to the label matrix Yc*nWith the prediction label matrixThe weight is joined Number Z, the Feature Mapping matrix Mc*dIt is updated, the training multi-tag disaggregated model.
Wherein, n, c, i, j and d are positive integer, and the value range of i is 1 to n, and the value range of j is 1 to c.
Therefore, the nerve network system provided by the embodiment of the present application can directly train model from input data, Without additional intermediate steps, i.e., the nerve network system is a nervous system end to end.Here, end to end Advantage is that feature extraction, Feature Mapping matrix and low-rank label correlation matrix can optimize simultaneously, that is to say, that the application is real Applying example can make feature extraction network more adapt to mission requirements, and multi-tag good classification effect with dynamic learning characteristics of image.
Optionally, the Feature Mapping network of the low-rank includes the first sub- mapping network and the second sub- mapping network, described The Feature Mapping network of low-rank, the first sub- mapping network and the second sub- mapping network have following relationship:
Wherein, the weight matrix of the described first sub- mapping network isThe weight matrix of the second sub- mapping network is Hc*r, r is positive integer and r≤min (d, c).
Optionally, the updating unit is specifically used for:
Determine the prediction label matrixWith the label matrix Yc*nBetween Euclidean distance loss function;
According to the Euclidean distance loss function, to the weighting parameter Z, the weight matrixAnd Hc*rIt carries out more Newly.
Optionally, the updating unit is specifically also used to:
By the sum of the Euclidean distance loss function and regular terms, it is determined as the majorized function of the n sample, wherein The regular terms is for constraining the weighting parameter Z, the weight matrixAnd Hc*r
Corresponding weighting parameter Z is as updated weighting parameter Z when using the value minimum of the majorized function, by institute State weight matrix corresponding when the value minimum of majorized functionAs updated weight matrixBy the optimization Corresponding weight matrix H when the value minimum of functionc*rAs updated weight matrix Hc*r
Optionally, the determination unit is specifically used for:
Determine training dataset, the training data concentrate include D sample and with each sample in the D sample Label vector, wherein the element y in the label vector of each samplejIndicate whether each sample includes j-th of mark Sign the object of instruction, wherein D is the positive integer not less than n;
N sample is randomly selected from training data concentration, and generates the label matrix Y of the n samplec*n, described Label matrix Yc*nIncluding the corresponding label vector of each sample in the n sample.
Therefore, in the embodiment of the present application, it is not necessary to disposably input entire training dataset and be calculated, and only need in batches Secondary input picture is calculated, therefore the embodiment of the present application can input entire data set in batch and be trained.Due to Training dataset generally includes a large amount of sample, therefore the embodiment of the present application can reduce by inputting training dataset in batches To the occupancy of resource during training pattern, the demand during training pattern to memory source is greatly reduced, it can be with Effectively solve the computational problem of low-rank label correlation matrix under large-scale data.
Optionally, further includes: the extraction unit is also used to extract the of first sample using the feature extraction network One eigenmatrix, wherein the first sample is not belonging to the n sample;
The acquiring unit is also used to obtain the first prediction of the fisrt feature matrix using first mapping network Label matrix, first sample described in the element representation in the first prediction label matrix include the object of j-th of label instruction Confidence level.
It should be noted that in the embodiment of the present invention, determination unit 610, extraction unit 620, acquiring unit 630 and updating unit 640 can be realized by processor.As shown in fig. 6, the device 700 of training multi-tag disaggregated model may include processor 710, deposit Reservoir 720 and communication interface 730.Wherein, memory 720 can be used for instruction or code of the execution of storage processor 710 etc..When The instruction or code are performed, the method which is used to execute above method embodiment offer, and processor 710 is also It is communicated for controlling communication interface 730 with the external world.
During realization, each step of the above method can pass through the integrated logic circuit of the hardware in processor 710 Or the instruction of software form is completed.The step of method in conjunction with disclosed in the embodiment of the present invention, can be embodied directly at hardware Reason device executes completion, or in processor hardware and software module combine and execute completion.Software module can be located at random Memory, flash memory, read-only memory, the abilities such as programmable read only memory or electrically erasable programmable memory, register In the storage medium of domain maturation.The storage medium is located at memory 720, and processor 710 reads the information in memory 720, knot Close the step of its hardware completes the above method.To avoid repeating, it is not detailed herein.
The dress of the device 600 of trained multi-tag disaggregated model shown in fig. 5 or trained multi-tag disaggregated model shown in fig. 6 Setting 700 can be realized the corresponding each process of preceding method embodiment, specifically, the device of the training multi-tag disaggregated model 600 or the device 700 of training multi-tag disaggregated model may refer to described above, to avoid repeating, which is not described herein again.
It should be understood that magnitude of the sequence numbers of the above procedures are not meant to execute suitable in the various embodiments of the application Sequence it is successive, the execution of each process sequence should be determined by its function and internal logic, the implementation without coping with the embodiment of the present application Process constitutes any restriction.
The embodiment of the present application also provides a kind of computer readable storage mediums, which is characterized in that including computer program, When the computer program is run on computers, so that the computer executes the method that above method embodiment provides.
The embodiment of the present application also provides a kind of computer program products comprising instruction, which is characterized in that when the meter When calculation machine program product is run on computers, so that the computer executes the method that above method embodiment provides.
It should be understood that processor as mentioned in the embodiments of the present invention can be central processing unit (Central Processing Unit, CPU), it can also be other general processors, digital signal processor (Digital Signal Processor, DSP), it is specific integrated circuit (Application Specific Integrated Circuit, ASIC), existing At programmable gate array (Field Programmable Gate Array, FPGA) or other programmable logic device, discrete Door or transistor logic, discrete hardware components etc..General processor can be microprocessor or the processor can also To be any conventional processor etc..
It should also be understood that memory as mentioned in the embodiments of the present invention can be volatile memory or non-volatile memories Device, or may include both volatile and non-volatile memories.Wherein, nonvolatile memory can be read-only memory (Read-Only Memory, ROM), programmable read only memory (Programmable ROM, PROM), erasable programmable are only Read memory (Erasable PROM, EPROM), electrically erasable programmable read-only memory (Electrically EPROM, ) or flash memory EEPROM.Volatile memory can be random access memory (Random Access Memory, RAM), use Make External Cache.By exemplary but be not restricted explanation, the RAM of many forms is available, such as static random-access Memory (Static RAM, SRAM), dynamic random access memory (Dynamic RAM, DRAM), synchronous dynamic random-access Memory (Synchronous DRAM, SDRAM), double data speed synchronous dynamic RAM (Double Data Rate SDRAM, DDR SDRAM), it is enhanced Synchronous Dynamic Random Access Memory (Enhanced SDRAM, ESDRAM), same Step connection dynamic random access memory (Synchlink DRAM, SLDRAM) and direct rambus random access memory (Direct Rambus RAM, DR RAM).
It should be noted that when processor is general processor, DSP, ASIC, FPGA or other programmable logic devices When part, discrete gate or transistor logic, discrete hardware components, memory (memory module) is integrated in the processor.
It should be noted that memory described herein is intended to include but is not limited to the storage of these and any other suitable type Device.
It should also be understood that the differentiation that first, second and various digital numbers that are referred to herein only carry out for convenience of description, It is not intended to limit scope of the present application.
Those of ordinary skill in the art may be aware that list described in conjunction with the examples disclosed in the embodiments of the present disclosure Member and algorithm steps can be realized with the combination of electronic hardware or computer software and electronic hardware.These functions are actually It is implemented in hardware or software, the specific application and design constraint depending on technical solution.Professional technician Each specific application can be used different methods to achieve the described function, but this realization is it is not considered that exceed Scope of the present application.
It is apparent to those skilled in the art that for convenience and simplicity of description, the system of foregoing description, The specific work process of device and unit, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In several embodiments provided herein, it should be understood that disclosed systems, devices and methods, it can be with It realizes by another way.For example, the apparatus embodiments described above are merely exemplary, for example, the unit It divides, only a kind of logical function partition, there may be another division manner in actual implementation, such as multiple units or components It can be combined or can be integrated into another system, or some features can be ignored or not executed.Another point, it is shown or The mutual coupling, direct-coupling or communication connection discussed can be through some interfaces, the indirect coupling of device or unit It closes or communicates to connect, can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme 's.
It, can also be in addition, each functional unit in each embodiment of the application can integrate in one processing unit It is that each unit physically exists alone, can also be integrated in one unit with two or more units.
It, can be with if the function is realized in the form of SFU software functional unit and when sold or used as an independent product It is stored in a computer readable storage medium.Based on this understanding, the technical solution of the application is substantially in other words The part of the part that contributes to existing technology or the technical solution can be embodied in the form of software products, the meter Calculation machine software product is stored in a storage medium, including some instructions are used so that a computer equipment (can be a People's computer, server or network equipment etc.) execute each embodiment the method for the application all or part of the steps. And storage medium above-mentioned includes: that USB flash disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), arbitrary access are deposited The various media that can store program code such as reservoir (Random Access Memory, RAM), magnetic or disk.
The above, the only specific embodiment of the application, but the protection scope of the application is not limited thereto, it is any Those familiar with the art within the technical scope of the present application, can easily think of the change or the replacement, and should all contain Lid is within the scope of protection of this application.Therefore, the protection scope of the application should be based on the protection scope of the described claims.

Claims (14)

1. a kind of method of trained multi-tag disaggregated model characterized by comprising
It is concentrated in training data and determines n sample and label matrix Y corresponding with the n samplec*n, the label matrix Yc*n In element yi*jIndicate i-th of sample whether include j-th of label instruction object, c indicate with the training data concentrate The number of the relevant label of sample;
The eigenmatrix X of the n sample is extracted using feature extraction networkd*n, wherein the feature extraction network has power Value parameter Z, d indicate the eigenmatrix Xd*nCharacteristic dimension;
The eigenmatrix X is obtained using Feature Mapping networkd*nPrediction label matrixThe prediction label matrixIn ElementIndicate that i-th of sample includes the confidence level of the object of j-th of label instruction, wherein the Feature Mapping network Weight matrix is the Feature Mapping matrix M of low-rankc*d
According to the label matrix Yc*nWith the prediction label matrixTo the weighting parameter Z, the Feature Mapping matrix Mc*dIt is updated, the training multi-tag disaggregated model;
Wherein, n, c, i, j and d are positive integer, and the value range of i is 1 to n, and the value range of j is 1 to c.
2. the method according to claim 1, wherein the Feature Mapping network of the low-rank includes the first son mapping Network and the second sub- mapping network, the Feature Mapping network of the low-rank, the first sub- mapping network and second son reflect Network is penetrated with following relationship:
Wherein, the weight matrix of the described first sub- mapping network isThe weight matrix of the second sub- mapping network is Hc*r, R is positive integer and r≤min (d, c).
3. according to the method described in claim 2, it is characterized in that, according to the label matrix Yc*nWith the prediction label square Battle arrayTo the weighting parameter Z, the Feature Mapping matrix Mc*dIt is updated, comprising:
Determine the prediction label matrixWith the label matrix Yc*nBetween Euclidean distance loss function;
According to the Euclidean distance loss function, to the weighting parameter Z, the weight matrixAnd Hc*rIt is updated.
4. according to the method described in claim 3, it is characterized in that, described according to the Euclidean distance loss function, to described Weighting parameter Z, the weight matrixAnd Hc*rIt is updated, comprising:
By the sum of the Euclidean distance loss function and regular terms, it is determined as the majorized function of the n sample, wherein described Regular terms is for constraining the weighting parameter Z, the weight matrixAnd Hc*r
Corresponding weighting parameter Z, will be described excellent as updated weighting parameter Z when using the value minimum of the majorized function Change weight matrix corresponding when the value minimum of functionAs updated weight matrixBy the majorized function Value minimum when corresponding weight matrix Hc*rAs updated weight matrix Hc*r
5. method according to claim 1-4, which is characterized in that described concentrate in training data determines n sample The label matrix Y of this and the n samplec*n, comprising:
Determine training dataset, it includes D sample and the label with each sample in the D sample that the training data, which is concentrated, Vector, wherein the element y in the label vector of each samplejIndicate whether each sample includes that j-th of label refers to The object shown, wherein D is the positive integer not less than n;
N sample is randomly selected from training data concentration, and generates the label matrix Y of the n samplec*n, the label Matrix Yc*nIncluding the corresponding label vector of each sample in the n sample.
6. method according to claim 1-5, which is characterized in that further include:
The fisrt feature matrix of first sample is extracted using the feature extraction network, wherein the first sample is not belonging to institute State n sample;
The first prediction label matrix of the fisrt feature matrix, the first pre- mark are obtained using first mapping network Sign the confidence level that first sample described in the element representation in matrix includes the object of j-th of label instruction.
7. a kind of device of trained multi-tag disaggregated model characterized by comprising
Determination unit determines n sample and label matrix Y corresponding with the n sample for concentrating in training datac*n, institute State label matrix Yc*nIn element yi*jIndicate i-th of sample whether include j-th of label instruction object, c indicate with it is described The number for the relevant label of sample that training data is concentrated;
Extraction unit, for extracting the eigenmatrix X of the n sample using feature extraction networkd*n, wherein the feature mentions Take network that there is weighting parameter Z, d to indicate the eigenmatrix Xd*nCharacteristic dimension;
Acquiring unit, for obtaining the eigenmatrix X using Feature Mapping networkd*nPrediction label matrixThe prediction Label matrixIn elementIndicate that i-th of sample includes the confidence level of the object of j-th of label instruction, wherein described The weight matrix of Feature Mapping network is the Feature Mapping matrix M of low-rankc*d
Updating unit, for according to the label matrix Yc*nWith the prediction label matrixTo the weighting parameter Z, described Feature Mapping matrix Mc*dIt is updated, the training multi-tag disaggregated model;
Wherein, n, c, i, j and d are positive integer, and the value range of i is 1 to n, and the value range of j is 1 to c.
8. device according to claim 7, which is characterized in that the Feature Mapping network of the low-rank includes the first son mapping Network and the second sub- mapping network, the Feature Mapping network of the low-rank, the first sub- mapping network and second son reflect Network is penetrated with following relationship:
Wherein, the weight matrix of the described first sub- mapping network isThe weight matrix of the second sub- mapping network is Hc*r, R is positive integer and r≤min (d, c).
9. device according to claim 8, which is characterized in that the updating unit is specifically used for:
Determine the prediction label matrixWith the label matrix Yc*nBetween Euclidean distance loss function;
According to the Euclidean distance loss function, to the weighting parameter Z, the weight matrixAnd Hc*rIt is updated.
10. device according to claim 9, which is characterized in that the updating unit is specifically also used to:
By the sum of the Euclidean distance loss function and regular terms, it is determined as the majorized function of the n sample, wherein described Regular terms is for constraining the weighting parameter Z, the weight matrixAnd Hc*r
Corresponding weighting parameter Z, will be described excellent as updated weighting parameter Z when using the value minimum of the majorized function Change weight matrix corresponding when the value minimum of functionAs updated weight matrixBy the majorized function Value minimum when corresponding weight matrix Hc*rAs updated weight matrix Hc*r
11. according to the described in any item devices of claim 7-10, which is characterized in that the determination unit is specifically used for:
Determine training dataset, it includes D sample and the label with each sample in the D sample that the training data, which is concentrated, Vector, wherein the element y in the label vector of each samplejIndicate whether each sample includes that j-th of label refers to The object shown, wherein D is the positive integer not less than n;
N sample is randomly selected from training data concentration, and generates the label matrix Y of the n samplec*n, the label Matrix Yc*nIncluding the corresponding label vector of each sample in the n sample.
12. according to the described in any item devices of claim 7-11, which is characterized in that further include:
The extraction unit is also used to extract the fisrt feature matrix of first sample using the feature extraction network, wherein institute It states first sample and is not belonging to the n sample;
The acquiring unit is also used to obtain the first prediction label of the fisrt feature matrix using first mapping network Matrix, first sample described in the element representation in the first prediction label matrix include setting for the object of j-th of label instruction Reliability.
13. a kind of computer readable storage medium, which is characterized in that including computer program, when the computer program is being counted When being run on calculation machine, so that the computer executes such as method of any of claims 1-6.
14. a kind of computer program product comprising instruction, which is characterized in that when the computer program product on computers When operation, so that the computer executes such as method of any of claims 1-6.
CN201711187395.0A 2017-11-24 2017-11-24 The method and apparatus of training multi-tag disaggregated model Pending CN109840530A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201711187395.0A CN109840530A (en) 2017-11-24 2017-11-24 The method and apparatus of training multi-tag disaggregated model
PCT/CN2018/094400 WO2019100724A1 (en) 2017-11-24 2018-07-04 Method and device for training multi-label classification model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711187395.0A CN109840530A (en) 2017-11-24 2017-11-24 The method and apparatus of training multi-tag disaggregated model

Publications (1)

Publication Number Publication Date
CN109840530A true CN109840530A (en) 2019-06-04

Family

ID=66630474

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711187395.0A Pending CN109840530A (en) 2017-11-24 2017-11-24 The method and apparatus of training multi-tag disaggregated model

Country Status (2)

Country Link
CN (1) CN109840530A (en)
WO (1) WO2019100724A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110929785A (en) * 2019-11-21 2020-03-27 中国科学院深圳先进技术研究院 Data classification method and device, terminal equipment and readable storage medium
CN111353076A (en) * 2020-02-21 2020-06-30 华为技术有限公司 Method for training cross-modal retrieval model, cross-modal retrieval method and related device
CN111368976A (en) * 2020-02-27 2020-07-03 杭州国芯科技股份有限公司 Data compression method based on neural network feature recognition
CN111652315A (en) * 2020-06-04 2020-09-11 广州虎牙科技有限公司 Model training method, object classification method, model training device, object classification device, electronic equipment and storage medium
CN111737520A (en) * 2020-06-22 2020-10-02 Oppo广东移动通信有限公司 Video classification method, video classification device, electronic equipment and storage medium
CN111898703A (en) * 2020-08-14 2020-11-06 腾讯科技(深圳)有限公司 Multi-label video classification method, model training method, device and medium
CN112115997A (en) * 2020-09-11 2020-12-22 苏州浪潮智能科技有限公司 Training method, system and device of object recognition model
CN112308299A (en) * 2020-10-19 2021-02-02 新奥数能科技有限公司 Sample data extraction method and device for power system load prediction model
CN112633419A (en) * 2021-03-09 2021-04-09 浙江宇视科技有限公司 Small sample learning method and device, electronic equipment and storage medium
CN113139433A (en) * 2021-03-29 2021-07-20 西安天和防务技术股份有限公司 Method and device for determining direction of arrival
WO2022227777A1 (en) * 2021-04-26 2022-11-03 华为技术有限公司 Model processing method and apparatus
CN115550014A (en) * 2022-09-22 2022-12-30 中国电信股份有限公司 Application program protection method and related equipment
CN117078359A (en) * 2023-10-16 2023-11-17 山东大学 Product recommendation method, system, equipment and medium based on user group classification
CN117217288A (en) * 2023-09-21 2023-12-12 摩尔线程智能科技(北京)有限责任公司 Fine tuning method and device for large model, electronic equipment and storage medium
CN117274726A (en) * 2023-11-23 2023-12-22 南京信息工程大学 Picture classification method and system based on multi-view supplementary tag

Families Citing this family (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10878296B2 (en) * 2018-04-12 2020-12-29 Discovery Communications, Llc Feature extraction and machine learning for automated metadata analysis
CN112149705A (en) * 2019-06-28 2020-12-29 京东数字科技控股有限公司 Method and system for training classification model, computer equipment and storage medium
CN112183757B (en) * 2019-07-04 2023-10-27 创新先进技术有限公司 Model training method, device and system
CN111797881A (en) * 2019-07-30 2020-10-20 华为技术有限公司 Image classification method and device
CN110659667A (en) * 2019-08-14 2020-01-07 平安科技(深圳)有限公司 Picture classification model training method and system and computer equipment
CN110688893A (en) * 2019-08-22 2020-01-14 成都通甲优博科技有限责任公司 Detection method for wearing safety helmet, model training method and related device
CN110569764B (en) * 2019-08-28 2023-12-22 北京工业大学 Mobile phone model identification method based on convolutional neural network
CN112529029A (en) * 2019-09-18 2021-03-19 华为技术有限公司 Information processing method, neural network training method, device and storage medium
CN111027582B (en) * 2019-09-20 2023-06-27 哈尔滨理工大学 Semi-supervised feature subspace learning method and device based on low-rank graph learning
CN112579746A (en) * 2019-09-29 2021-03-30 京东数字科技控股有限公司 Method and device for acquiring behavior information corresponding to text
CN110765935A (en) * 2019-10-22 2020-02-07 上海眼控科技股份有限公司 Image processing method, image processing device, computer equipment and readable storage medium
CN112825144A (en) * 2019-11-20 2021-05-21 深圳云天励飞技术有限公司 Picture labeling method and device, electronic equipment and storage medium
CN112884159A (en) * 2019-11-30 2021-06-01 华为技术有限公司 Model updating system, model updating method and related equipment
CN112994701B (en) * 2019-12-02 2024-05-03 阿里巴巴集团控股有限公司 Data compression method, device, electronic equipment and computer readable medium
CN113010500A (en) * 2019-12-18 2021-06-22 中国电信股份有限公司 Processing method and processing system for DPI data
CN111291618B (en) * 2020-01-13 2024-01-09 腾讯科技(深圳)有限公司 Labeling method, labeling device, server and storage medium
CN111275089B (en) * 2020-01-16 2024-03-05 北京小米松果电子有限公司 Classification model training method and device and storage medium
CN111339362B (en) * 2020-02-05 2023-07-18 天津大学 Short video multi-label classification method based on deep collaborative matrix decomposition
CN112016040A (en) * 2020-02-06 2020-12-01 李迅 Weight matrix construction method, device, equipment and storage medium
CN111340131B (en) * 2020-03-09 2023-07-14 北京字节跳动网络技术有限公司 Image labeling method and device, readable medium and electronic equipment
CN111461191B (en) * 2020-03-25 2024-01-23 杭州跨视科技有限公司 Method and device for determining image sample set for model training and electronic equipment
GB2606792A (en) 2020-03-26 2022-11-23 Shenzhen Inst Adv Tech Time series data generation method and device based on multi-condition constraints, and medium
CN111475496B (en) * 2020-03-26 2023-07-21 深圳先进技术研究院 Time sequence data generation method, device and medium based on multi-condition constraint
CN111159773B (en) * 2020-04-01 2020-11-03 支付宝(杭州)信息技术有限公司 Picture classification method and device for protecting data privacy
CN111581488B (en) * 2020-05-14 2023-08-04 上海商汤智能科技有限公司 Data processing method and device, electronic equipment and storage medium
CN111611386B (en) * 2020-05-28 2024-03-29 北京明略昭辉科技有限公司 Text classification method and device
CN111860572B (en) * 2020-06-04 2024-01-26 北京百度网讯科技有限公司 Data set distillation method, device, electronic equipment and storage medium
CN111709475B (en) * 2020-06-16 2024-03-15 全球能源互联网研究院有限公司 N-gram-based multi-label classification method and device
CN111914885B (en) * 2020-06-19 2024-04-26 合肥工业大学 Multi-task personality prediction method and system based on deep learning
CN111931809A (en) * 2020-06-29 2020-11-13 北京大米科技有限公司 Data processing method and device, storage medium and electronic equipment
CN111916144B (en) * 2020-07-27 2024-02-09 西安电子科技大学 Protein classification method based on self-attention neural network and coarsening algorithm
CN111898707A (en) * 2020-08-24 2020-11-06 鼎富智能科技有限公司 Model training method, text classification method, electronic device and storage medium
CN112215795B (en) * 2020-09-02 2024-04-09 苏州超集信息科技有限公司 Intelligent detection method for server component based on deep learning
CN112069319B (en) * 2020-09-10 2024-03-22 杭州中奥科技有限公司 Text extraction method, text extraction device, computer equipment and readable storage medium
CN112182214B (en) * 2020-09-27 2024-03-19 中国建设银行股份有限公司 Data classification method, device, equipment and medium
CN112149692B (en) * 2020-10-16 2024-03-05 腾讯科技(深圳)有限公司 Visual relationship identification method and device based on artificial intelligence and electronic equipment
CN112434722B (en) * 2020-10-23 2024-03-19 浙江智慧视频安防创新中心有限公司 Label smooth calculation method and device based on category similarity, electronic equipment and medium
CN112307133A (en) * 2020-10-29 2021-02-02 平安普惠企业管理有限公司 Security protection method and device, computer equipment and storage medium
CN112560966B (en) * 2020-12-18 2023-09-15 西安电子科技大学 Polarized SAR image classification method, medium and equipment based on scattering map convolution network
CN112668509B (en) * 2020-12-31 2024-04-02 深圳云天励飞技术股份有限公司 Training method and recognition method of social relation recognition model and related equipment
CN113033318B (en) * 2021-03-01 2023-09-26 深圳大学 Human body motion detection method, device and computer readable storage medium
CN113095364B (en) * 2021-03-12 2023-12-19 西安交通大学 High-speed rail seismic event extraction method, medium and equipment using convolutional neural network
CN112948937B (en) * 2021-03-12 2024-03-01 中建西部建设贵州有限公司 Intelligent pre-judging method and device for concrete strength
CN113095210A (en) * 2021-04-08 2021-07-09 北京一起教育科技有限责任公司 Method and device for detecting pages of exercise book and electronic equipment
CN113157788B (en) * 2021-04-13 2024-02-13 福州外语外贸学院 Big data mining method and system
CN113516239A (en) * 2021-04-16 2021-10-19 Oppo广东移动通信有限公司 Model training method and device, storage medium and electronic equipment
CN115481746B (en) * 2021-06-15 2023-09-01 华为技术有限公司 Model training method, related system and storage medium
CN113469249B (en) * 2021-06-30 2024-04-09 阿波罗智联(北京)科技有限公司 Image classification model training method, classification method, road side equipment and cloud control platform
CN113657087B (en) * 2021-08-25 2023-12-15 平安科技(深圳)有限公司 Information matching method and device
CN113821664A (en) * 2021-08-30 2021-12-21 湖南军芃科技股份有限公司 Image classification method, system, terminal and readable storage medium based on histogram statistical frequency
CN113837394A (en) * 2021-09-03 2021-12-24 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Multi-feature view data label prediction method, system and readable storage medium
CN114648635A (en) * 2022-03-15 2022-06-21 安徽工业大学 Multi-label image classification method fusing strong correlation among labels
CN116070120B (en) * 2023-04-06 2023-06-27 湖南归途信息科技有限公司 Automatic identification method and system for multi-tag time sequence electrophysiological signals
CN117076994B (en) * 2023-10-18 2024-01-26 清华大学深圳国际研究生院 Multi-channel physiological time sequence classification method
CN117312865B (en) * 2023-11-30 2024-02-27 山东理工职业学院 Nonlinear dynamic optimization-based data classification model construction method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150039613A1 (en) * 2013-07-31 2015-02-05 Linkedln Corporation Framework for large-scale multi-label classification
CN104899596A (en) * 2015-03-16 2015-09-09 景德镇陶瓷学院 Multi-label classification method and apparatus thereof
CN105320967A (en) * 2015-11-04 2016-02-10 中科院成都信息技术股份有限公司 Multi-label AdaBoost integration method based on label correlation
US20160140451A1 (en) * 2014-11-17 2016-05-19 Yahoo! Inc. System and method for large-scale multi-label learning using incomplete label assignments

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8379994B2 (en) * 2010-10-13 2013-02-19 Sony Corporation Digital image analysis utilizing multiple human labels
CN105825502B (en) * 2016-03-12 2018-06-15 浙江大学 A kind of Weakly supervised method for analyzing image of the dictionary study based on conspicuousness guidance
CN107292322B (en) * 2016-03-31 2020-12-04 华为技术有限公司 Image classification method, deep learning model and computer system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150039613A1 (en) * 2013-07-31 2015-02-05 Linkedln Corporation Framework for large-scale multi-label classification
US20160140451A1 (en) * 2014-11-17 2016-05-19 Yahoo! Inc. System and method for large-scale multi-label learning using incomplete label assignments
CN104899596A (en) * 2015-03-16 2015-09-09 景德镇陶瓷学院 Multi-label classification method and apparatus thereof
CN105320967A (en) * 2015-11-04 2016-02-10 中科院成都信息技术股份有限公司 Multi-label AdaBoost integration method based on label correlation

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
姚红革等: "基于小波分析和BP神经网络的图像特征提取", 《西安工业大学学报》 *
王臻: "基于学习标签相关性的多标签分类算法", 《中国科技大学 硕士论文》 *

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110929785A (en) * 2019-11-21 2020-03-27 中国科学院深圳先进技术研究院 Data classification method and device, terminal equipment and readable storage medium
WO2021098618A1 (en) * 2019-11-21 2021-05-27 中国科学院深圳先进技术研究院 Data classification method and apparatus, terminal device and readable storage medium
CN110929785B (en) * 2019-11-21 2023-12-05 中国科学院深圳先进技术研究院 Data classification method, device, terminal equipment and readable storage medium
CN111353076A (en) * 2020-02-21 2020-06-30 华为技术有限公司 Method for training cross-modal retrieval model, cross-modal retrieval method and related device
CN111353076B (en) * 2020-02-21 2023-10-10 华为云计算技术有限公司 Method for training cross-modal retrieval model, cross-modal retrieval method and related device
CN111368976A (en) * 2020-02-27 2020-07-03 杭州国芯科技股份有限公司 Data compression method based on neural network feature recognition
CN111368976B (en) * 2020-02-27 2022-09-02 杭州国芯科技股份有限公司 Data compression method based on neural network feature recognition
CN111652315A (en) * 2020-06-04 2020-09-11 广州虎牙科技有限公司 Model training method, object classification method, model training device, object classification device, electronic equipment and storage medium
CN111737520A (en) * 2020-06-22 2020-10-02 Oppo广东移动通信有限公司 Video classification method, video classification device, electronic equipment and storage medium
CN111737520B (en) * 2020-06-22 2023-07-25 Oppo广东移动通信有限公司 Video classification method, video classification device, electronic equipment and storage medium
CN111898703A (en) * 2020-08-14 2020-11-06 腾讯科技(深圳)有限公司 Multi-label video classification method, model training method, device and medium
CN111898703B (en) * 2020-08-14 2023-11-10 腾讯科技(深圳)有限公司 Multi-label video classification method, model training method, device and medium
CN112115997A (en) * 2020-09-11 2020-12-22 苏州浪潮智能科技有限公司 Training method, system and device of object recognition model
CN112115997B (en) * 2020-09-11 2022-12-02 苏州浪潮智能科技有限公司 Training method, system and device of object recognition model
CN112308299A (en) * 2020-10-19 2021-02-02 新奥数能科技有限公司 Sample data extraction method and device for power system load prediction model
CN112308299B (en) * 2020-10-19 2024-04-19 新奥数能科技有限公司 Sample data extraction method and device for power system load prediction model
CN112633419B (en) * 2021-03-09 2021-07-06 浙江宇视科技有限公司 Small sample learning method and device, electronic equipment and storage medium
CN112633419A (en) * 2021-03-09 2021-04-09 浙江宇视科技有限公司 Small sample learning method and device, electronic equipment and storage medium
CN113139433A (en) * 2021-03-29 2021-07-20 西安天和防务技术股份有限公司 Method and device for determining direction of arrival
WO2022227777A1 (en) * 2021-04-26 2022-11-03 华为技术有限公司 Model processing method and apparatus
CN115550014B (en) * 2022-09-22 2024-03-19 中国电信股份有限公司 Application program protection method and related equipment
CN115550014A (en) * 2022-09-22 2022-12-30 中国电信股份有限公司 Application program protection method and related equipment
CN117217288A (en) * 2023-09-21 2023-12-12 摩尔线程智能科技(北京)有限责任公司 Fine tuning method and device for large model, electronic equipment and storage medium
CN117217288B (en) * 2023-09-21 2024-04-05 摩尔线程智能科技(北京)有限责任公司 Fine tuning method and device for large model, electronic equipment and storage medium
CN117078359B (en) * 2023-10-16 2024-01-12 山东大学 Product recommendation method, system, equipment and medium based on user group classification
CN117078359A (en) * 2023-10-16 2023-11-17 山东大学 Product recommendation method, system, equipment and medium based on user group classification
CN117274726B (en) * 2023-11-23 2024-02-23 南京信息工程大学 Picture classification method and system based on multi-view supplementary tag
CN117274726A (en) * 2023-11-23 2023-12-22 南京信息工程大学 Picture classification method and system based on multi-view supplementary tag

Also Published As

Publication number Publication date
WO2019100724A1 (en) 2019-05-31

Similar Documents

Publication Publication Date Title
CN109840530A (en) The method and apparatus of training multi-tag disaggregated model
CN109840531B (en) Method and device for training multi-label classification model
CN110188795A (en) Image classification method, data processing method and device
CN110309856A (en) Image classification method, the training method of neural network and device
US10002313B2 (en) Deeply learned convolutional neural networks (CNNS) for object localization and classification
Kao et al. Visual aesthetic quality assessment with a regression model
CN112446270B (en) Training method of pedestrian re-recognition network, pedestrian re-recognition method and device
CN110378381A (en) Object detecting method, device and computer storage medium
CN108765278A (en) A kind of image processing method, mobile terminal and computer readable storage medium
CN107844784A (en) Face identification method, device, computer equipment and readable storage medium storing program for executing
CN109583483A (en) A kind of object detection method and system based on convolutional neural networks
CN106845418A (en) A kind of hyperspectral image classification method based on deep learning
CN108921061A (en) A kind of expression recognition method, device and equipment
CN107563280A (en) Face identification method and device based on multi-model
CN110619059B (en) Building marking method based on transfer learning
CN109271991A (en) A kind of detection method of license plate based on deep learning
CN105320945A (en) Image classification method and apparatus
CN110023989B (en) Sketch image generation method and device
CN110222717A (en) Image processing method and device
CN110222718B (en) Image processing method and device
CN112561027A (en) Neural network architecture searching method, image processing method, device and storage medium
CN113505768A (en) Model training method, face recognition method, electronic device and storage medium
CN109492589A (en) The recognition of face working method and intelligent chip merged by binary features with joint stepped construction
CN108647571A (en) Video actions disaggregated model training method, device and video actions sorting technique
EP3867808A1 (en) Method and device for automatic identification of labels of image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20190604

WD01 Invention patent application deemed withdrawn after publication