CN106909905A - A kind of multi-modal face identification method based on deep learning - Google Patents

A kind of multi-modal face identification method based on deep learning Download PDF

Info

Publication number
CN106909905A
CN106909905A CN201710122193.1A CN201710122193A CN106909905A CN 106909905 A CN106909905 A CN 106909905A CN 201710122193 A CN201710122193 A CN 201710122193A CN 106909905 A CN106909905 A CN 106909905A
Authority
CN
China
Prior art keywords
mode
modal
similarity
face
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710122193.1A
Other languages
Chinese (zh)
Other versions
CN106909905B (en
Inventor
张�浩
韩琥
山世光
陈熙霖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
In Extension (beijing) Technology Co Ltd
Original Assignee
In Extension (beijing) Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by In Extension (beijing) Technology Co Ltd filed Critical In Extension (beijing) Technology Co Ltd
Priority to CN201710122193.1A priority Critical patent/CN106909905B/en
Publication of CN106909905A publication Critical patent/CN106909905A/en
Application granted granted Critical
Publication of CN106909905B publication Critical patent/CN106909905B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a kind of multi-modal face identification method based on deep learning, Face datection and alignment are carried out to RGB facial images including (1), human face data the collection S0, S1, S2 ... for making RGB mode and other mode are cut according to the mapping relations between mode;(2) a depth convolutional neural networks structure N1 for multi-modal fusion is designed, and trains N1 networks;(3) a multi-modal shared depth convolutional neural networks structure N2 is designed, and trains N2 networks;(4) feature stage, (5) Similarity Measure stage and (6) similarity fusing stage are extracted.The present invention uses multimodal systems, by the collection for carrying out various face modal datas, using the respective advantage of multiple modalities information, overcome by convergence strategy in some of single mode system in weakness, multiple modalities information is sufficiently utilized simultaneously, effectively improve the performance of face identification system so that recognition of face is more quick accurate.

Description

A kind of multi-modal face identification method based on deep learning
Technical field
The present invention relates to a kind of face identification method, more particularly to a kind of multi-modal recognition of face side based on deep learning Method.
Background technology
Three-dimensional face identification has it to illumination robust, by factors such as attitude and expressions relative to two-dimension human face identification The advantages of influence is smaller, therefore developed rapidly in three dimensional data collection technology and the quality and precision of three-dimensional data are greatly promoted Afterwards, during their research is put into the field by many scholars.
The image of face different modalities is easily influenceed by different factors etc., and these factors have impact on to a certain extent The stability and accuracy of single mode face identification system.CN104778441A proposes a kind of fusion half-tone information and depth letter The multi-modal face identification device and method of breath, its core methed are the (institutes in invention after extracting multi-modal face characteristic The feature for using is the feature of hand-designed), it is stitched together to form a feature pool, it is characterized each feature construction one in pond Individual Weak Classifier, then using Adaboost algorithm, picks out for the maximally efficient feature, last base of classifying in feature pool In the feature that multi-modal Feature-level fusion is obtained, matching fraction is calculated using nearest neighbor classifier, multi-modal people is realized with this Face is recognized.But the feature that the invention is used is the feature of engineer, ability to express is not strong enough;And the feature of the invention Fusion Features and feature selecting are carried out using Adaboost algorithm, it is less efficient;And the invention is for specific both modalities which is designed , with limitation.
The content of the invention
For the weak point in solving the above problems, the invention provides a kind of multi-modal face based on deep learning Recognition methods.
In order to solve the above technical problems, the technical solution adopted by the present invention is:It is a kind of based on the multi-modal of deep learning Face identification method, comprises the following steps:
(1) Face datection is carried out to RGB facial images, positioning feature point, alignment cuts, and makes the RGB mode after cutting Human face data collection S0;According to the coordinate mapping relations between RGB mode and other mode, the feature of other mode faces is found Point, and cut human face data the collection S1, S2 ... for making other mode;
(2) a depth convolutional neural networks structure N1 for multi-modal fusion is designed, in this structure, first half is Several independent neutral net branches, input one mode of correspondence of each branch, then with specific network structure multiple Bimodal branch is fused into a branch, reconnects a series of neural network structure units;Then by S0, S1, S2 ... are sent to In the different branches of N1, N1 networks, the model for training is trained to be represented with M1;
Above-mentioned neural network structure unit includes but is not limited to convolutional layer, normalization layer, non-linear layer, pond layer and connects entirely Connect layer and distribution normalization layer;Mode includes but is not limited to RGB mode, depth mode and near-infrared mode;Specific network knot In structure, there is respective Classification Loss in each branch as the respective supervisory signals of each mode, and structure fusion method is included but not It is limited to simple merging features;
(3) a multi-modal shared depth convolutional neural networks structure N2 is designed, by S0, S1, S2 ... are not added with distinguishing The N2 of feeding together in, train N2 networks, the model for training represents with M2;
(4) feature stage is extracted, for registered set and the image of query set, its modality range is in training set modality range It is interior.The different modalities of certain image can be expressed as I0, I1, I2 ..., and then I0, I1, I2 ... are respectively in model M 1 and M2 Upper extraction feature, feature can use F0, F0C, F1, F1C, F2, F2C ... to represent, it is the feature extracted from M2 that C is represented;
(5) the similarity S11's, GF2 and PF2 of similarity S00, GF1 and PF1 respectively between calculating GF0 and PF0 is similar Degree S22;The similarity S01 between GF0C and PF1C is calculated, similar calculates cross-module state similarity S02, S03, S12, S13, S23……;
Above-mentioned GF0 represents that the F0 of registered set image, PF0 represent that the F0 of query set image, GF0C represent registered set image F0C, PF1C represent the F1C of query set image;
(6) summation fusion is weighted to all of registered set and query set similarity, final fusion similarity is obtained S, recognition of face and confirming face are carried out on the similarity matrix that fusion similarity S is constituted.
In step (2), when training N1 networks, loss layer using softmax with loss or can use it His loss layer.
In step (3), by S0, S1, S2 ... be not added with distinguish feeding N2 in when, if their port number is not Together, then can take and all be changed into single pass mode and be normalized to same channels, or to repeat single channel identical to port number, Then it is re-fed into being trained in network structure.
In step (4), for registered set and the image of query set, its modality range is within training set modality range.
The present invention uses multimodal systems, by carrying out the collection of various face modal datas, using multiple modalities information Respective advantage, is overcome in some of single mode system in weakness by convergence strategy, while sufficiently utilizing multiple modalities Information, effectively improves the performance of face identification system so that recognition of face is more quick accurate.
Brief description of the drawings
The present invention is further detailed explanation with reference to the accompanying drawings and detailed description.
Fig. 1 is the flow chart of face recognition algorithms of the present invention.
Fig. 2 is the structured flowchart of the depth convolutional neural networks of multi-modal fusion of the present invention.
Fig. 3 is the concrete structure block diagram of Loss4 in Fig. 2.
Specific embodiment
As shown in figure 1, the present invention specifically includes following steps:
(1) Face datection is carried out to RGB facial images, positioning feature point, alignment cuts, and makes the RGB mode after cutting Human face data collection S0;Then according to the coordinate mapping between RGB mode and other mode (such as depth information, Near Infrared Information etc.) Relation, finds the characteristic point of other mode faces, and cuts human face data the collection S1, S2 ... for making other mode;
(2) a deep neural network structure N1 for multi-modal fusion is designed, in this structure, first half is several Independent neutral net branch, one mode of input correspondence (such as the RGB mode, depth mode, near-infrared mode of each branch Deng), multiple bimodal branch are then permeated a neutral net branch of synthesis (such as these with specific network structure Feature is coupled together, or is stacked up by passage, or other attachment structures, such as structure etc. as accompanying drawing 3), reconnect A series of neural network structure units (such as convolutional layer, normalization layer, non-linear layer, pond layer, full articulamentum);Then will S0, S1, S2 ... are sent in the different branches of N1, train N1 networks, the model for training to be represented with M1;
(3) a multi-modal shared deep neural network structure N2 is designed, by S0, S1, S2 ... are not added with for distinguishing Rise in feeding N2, train N2 networks, the model for training to be represented with M2;
(4) feature stage is extracted, for registered set and the image of query set, its modality range is in training set modality range Interior, the different modalities of certain image can be expressed as I0, I1, I2 ..., and then I0, I1, I2 ... are respectively in model M 1 and M2 Upper extraction feature, feature can use F0, F0C, F1, F1C, F2, F2C ... to represent, it is the feature extracted from M2 that C is represented;
(5) similarity between GF0 and PF0 (representing the F0 of the registered set image and F0 of query set image respectively) is calculated The similarity S22 of the similarity S11, GF2 and PF2 of S00, GF1 and PF1;Calculate GF0C and PF1C (represents registered set image respectively F0C and query set image F1C) between similarity S01, similar calculates cross-module state similarity S02, S03, S12, S13, S23 ...;
(6) similarity to all of registered set and query set is weighted summation fusion, obtains final fusion similar Degree S, recognition of face and confirming face are carried out on the similarity matrix that fusion similarity S is constituted.
The formula of above-mentioned weighted sum fusion can be expressed as:S=p1*S1+p2*S2+p3*S3 ..., wherein p1:p2: P3 can be S1, and S2, S3 individually carry out the inverse ratio (1/r1 of the recognition correct rate (r1, r2, r3) that face recognition experiment is obtained:1/ r2:1/r3)。
In deep neural network structure N1 in step (2), network structure elements can include but is not limited to convolutional layer, Pond layer, nonlinear function layer, full articulamentum, distribution normalization layer etc., and include but is not limited to any combination of these layers; The simple combination of network structure elements brand-new design, or can be changed based on existing open network structure, The simple combination of network structure elements is not in itself the scope of protection, as long as meet the present invention being described, first half is Several independent neutral net branches, a branch for the individual synthesis that permeated by ad hoc structure, and each mode in ad hoc structure There are respective independent supervisory signals in branch, connects a series of structure shape of neural network structure units after the branch after synthesis again Formula, then belong to the scope of protection.
One is given as shown in Figure 2 based on Google Inception-v2 (a kind of depth network structure that Google is proposed) The network structure come according to above-mentioned rules modification, in the network structure, if removing Loss4 loss structures, removes depth mould State and near-infrared bimodal branch, a remaining RGB branch is directly connected to 3*Inception structures, and (* represents multiple similar structure strings Connection, is 3 Inception structures in series such as herein, and Inception structures are a sub-network structure of Google's definition, [2] have the expression to Inception structures in the Figure 3 in) on, then completely it is exactly Google Inception-v2 network knots In itself, in the Figure 5 of [1], the 2* (convolution+pond) of accompanying drawing 2 corresponds to the convolution (convolutional layer) of Figure 5 to structure, Max pool (maximum pond layer), this four layers of convolution (convolutional layer), max pool (maximum pond layer), and from it is lower to On first 3*Inception then correspond to [1] in inception (3a), inception (3b), inception (3c), 4*Inception thereon then corresponds to inception (4a), inception (4b), inception (4c) in [1], Inception (4d), then 3*Inception thereon then corresponds to inception (4e), the inception (5a) in [1], Inception (5b), Loss3 ponds layer then corresponds to the avg pool (average pond layer) in [1].
In our network structure, we first allow the data (RGB, depth information, Near Infrared Information etc.) of different modalities By a series of neural network structures, (each respective supervisory signals of mode can in fusion structure for study to the distinctive feature of mode To ensure and promote the study of mode characteristic feature), then fusion connection is got up by a series of neural network structures, to help Network carries out the complementary feature learning of mode, and can be used when fusion connection will in the way of passage (channel) is stacked Characteristic pattern is stacked up.For Loss4 structures, it is the structure of particular design, except that can promote network convergence, also may be used To promote e-learning mode complementary characteristic, a kind of structure of Loss4 is shown in accompanying drawing 3, three branches correspond to three respectively Mode, picks out from three branches of accompanying drawing 2, and FC is to represent full articulamentum in structure, after FC or under numeral be to represent The node quantity of the full articulamentum;It is 5*5 that average pond Hua Ceng pondizations are interval, and interval steps are 3;Convolutional layer convolution kernel size is 1* 1, interval steps are 1;The total number of persons of training set, is the categorical measure of label (label);+ expression is input to three 512 therein The full articulamentum of node is averaging (fc3 [i]=(fcc3 [i]+fcd3 [i]+fcn3 [i])/3, i=1,2,3 ... by node 512), it is clear that obtain an or full articulamentum for 512 nodes, namely fc3;FC2048 is three the 512 of mode and connects entirely Connect layer and three modality fusions 512 full articulamentums be connected in series after the full articulamentums of 2048 dimensions that obtain;
When present invention training N1 networks, loss layer can use softmax with loss (classical Classification Loss Layer), it is also possible to use unknown losses layer.The used algorithm of training is that Back Propagation reversely return algorithm, is passed through Return the error of loss layer to update each layer of parameter so that network parameter is updated, and finally gives convergence.Its is specific Training step, as a example by the data set that is used during with present invention experiment (non-public, to be gathered with partner cooperation), the data set Training set scale be for about 500,000 samples, about 500 people.Use 32 batch (data block) scale, basic learning rate 0.045, every 6400 iterative learning rate is multiplied by 0.9, weight decay (weights decay), and for 0.0002, momentum, (momentum is joined Number) it is 0.9. about 400,000 iteration of training.
The present invention in deep neural network structure N2, likewise, network structure elements can include but is not limited to convolution Layer, pond layer, nonlinear function layer, full articulamentum, distribution normalization layer etc., and include but is not limited to any group of these layers Close;In this step, network structure is not the scope of protection, trains multi-modal public characteristic empty by way of not differentiating between mode Between method be protection scope, network structure can be with brand-new design, it is also possible to using network structure disclosed in academia such as The networks such as AlexNet, GoogleNet, ResNet.By S0, S1, S2 etc. be not added with distinguish the N2 of feeding together in when, if Their port number is different, then can take all be changed into single pass mode to be normalized to same channels (such as RGB triple channels change It is gray-scale map single channel), or to repeat single channel identical (such as repeat near-infrared single channel figure three times to triple channel) to port number, Then it is re-fed into being trained in network structure.
The present embodiment is with RGB, depth information, as a example by Near Infrared Information, because depth information and Near Infrared Information are list The RGB figures of all samples, can be switched to single channel gray-scale map by passage, be re-fed into being trained in N2 networks.With Google As a example by Inception-v2 [1], by all samples of 500,000 training sets regardless of mode regard as 1,500,000 samples (RGB, Depth, three kinds of mode of near-infrared, wherein RGB switchs to gray-scale map herein), study is trained in input Inception-v2, make With 32 batch block sizes, basic learning rate 0.045, every 19200 iterative learning rates are multiplied by 0.9, weight decay (weights decay) is that 0.0002, momentum (momentum parameter) is 0.9, trains about 1,200,000 iteration.The model for obtaining It is set to M2.Then for the feature of model, it is believed that be the feature of cross-module state, the RGB of sample (having switched to a gray-scale map) feature Can be compared with its depth information feature, the RGB feature of query set can carry out similarity-rough set with the depth information feature of registered set, This is the rudimentary algorithm of cross-module state identification division of the present invention.
Feature stage is being extracted, can be training set modality range for the modality range of registered set and query set image Complete or collected works, or proper subclass, but can not be beyond training set modality range;The position for extracting feature can be in the He of model M 1 The top layer of M2, it is also possible at non-top layer (such as middle a certain layer).
In above-mentioned step (5), the distance metric method for calculating similarity can be COS distance, or European Distance, or other distance metrics such as mahalanobis distance.The present invention by taking COS distance as an example, for two characteristic vectors x1, x2, its COS distance d=x1 ' * x2/ (| x1 | * | x2 |).Where it is assumed that x1, x2 are column vector, x1 ' is the transposition of x1, and x1 ' * x2 are The dot product of x1 ' and x2, | x1 |, | x2 | are x1, and the mould of x2 is long, wherein mould satisfaction | x1 | 2=x1 ' * x1 long.
Face recognition experiment of the present invention refers to giving a registered set, and registered set has a series of different face figures Picture, when a query image is given, which individual that the query image is registered set is found, and method can think registration Collection image and that people of query image similarity highest;Confirming face experiment refers to given two query images, determines this Whether two images are same persons, and method can be to determine a threshold value, if two image similarities are higher than threshold value, then it is assumed that It is same person, otherwise it is assumed that not being same person.
The present invention has several key technology points:1) multi-modal fusion is carried out using network structure, fusion position is in network Middle part, neither bottom input position also non-top layer loss layer position, the mode quantity of fusion can be two, or many It is individual, for each mode provides single supervisory signals in fusion structure, to allow each mode feature of itself to be filled Divide and excavate and be unlikely to be fallen into oblivion by other modal characteristics.Therefore the most abundant Generalization Capability for recognition of face can be obtained Feature representation and syncretizing effect, and be not limited to two mode, be adapted to multiple mode, more flexibly.2) for across The mode of learning in mode public characteristic space, the face image data of multiple modalities is not added with differentiation is added to depth as input It is trained in degree neutral net, mode quantity can be two, or multiple, and the label of image is the label of people, no Mode is distinguished, the public characteristic space that training is obtained can choose the top layer of network, it is also possible to select the non-top layer of network; Technique effect can be the recognition effect of cross-module state between the various face mode of acquisition.3) it is similar to all multi-modal mode inside Similarity is weighted the mode of summation fusion between degree and cross-module morphotype state, obtains the multi-modal fusion phase of final two images Like spend, for test image mode can completely can also be imperfect, its modality type set can be training set matched moulds The complete or collected works of state type set can also be proper subclass;Technique effect be so that multi-modal recognition of face obtain very strong flexibility and Applicability, goes for many scenes (as only registered set is multi-modal, query image single mode etc.) of multi-modal recognition of face.
Therefore, the present invention is excavated and study face by designing multi-modal and cross-module state deep learning network, sufficiently Complementary characteristic between multi-modal, greatly improves the performance of recognition of face, and there is provided very strong multi-modal and cross-module state The flexibility of face recognition application and applicability.
Above-mentioned implementation method is not limitation of the present invention, and the present invention is also not limited to the example above, this technology neck Change, remodeling, addition or replacement that the technical staff in domain is made in the range of technical scheme, also belong to this hair Bright protection domain.

Claims (5)

1. a kind of multi-modal face identification method based on deep learning, it is characterised in that comprise the following steps:
(1) Face datection is carried out to RGB facial images, positioning feature point, alignment cuts, and makes the RGB mode faces after cutting Data set S0;According to the coordinate mapping relations between RGB mode and other mode, the characteristic point of other mode faces is found, and Cut human face data the collection S1, S2 ... for making other mode;
(2) a depth convolutional neural networks structure N1 for multi-modal fusion is designed, in this structure, first half is several Independent neutral net branch, input one mode of correspondence of each branch, then with specific network structure multiple mode Branch is fused into a branch, reconnects a series of neural network structure units;Then by S0, S1, S2 ... are sent to N1's In respective branches, N1 networks, the model for training is trained to be represented with M1;
Above-mentioned mode includes but is not limited to RGB mode, depth mode and near-infrared mode;In specific network structure, each point Branch has respective Classification Loss as the respective supervisory signals of each mode, and structure fusion method includes but is not limited to simple spy Levy splicing;
(3) a multi-modal shared depth convolutional neural networks structure N2 is designed, by S0, S1, S2 ... are not added with for distinguishing Rise in feeding N2, train N2 networks, the model for training to be represented with M2;
(4) extract feature stage, for registered set and the image of query set, its modality range in training set modality range, certain The different modalities of image can be expressed as I0, I1, I2 ..., and then I0, I1, I2 ... are extracted in model M 1 and M2 respectively Feature, feature can use F0, F0C, F1, F1C, F2, F2C ... to represent, it is the feature extracted from M2 that C is represented;
(5) similarity of the similarity S11, GF2 and PF2 of similarity S00, GF1 and PF1 respectively between calculating GF0 and PF0 S22;The similarity S01 between GF0C and PF1C is calculated, similar calculates cross-module state similarity S02, S03, S12, S13, S23……;
Above-mentioned GF0 represents that the F0 of registered set image, PF0 represent that the F0 of query set image, GF0C represent the F0C of registered set image, PF1C represents the F1C of query set image;
(6) summation fusion is weighted to all of registered set and query set similarity, final fusion similarity S is obtained, Recognition of face and confirming face are carried out on the similarity matrix that fusion similarity S is constituted.
2. the multi-modal face identification method based on deep learning according to claim 1, it is characterised in that:The nerve Network structure elements include but is not limited to convolutional layer, normalization layer, non-linear layer, pond layer and full articulamentum and distribution normalization Layer.
3. the multi-modal face identification method based on deep learning according to claim 1, it is characterised in that:The step (2) in, when training N1 networks, loss layer can use softmax withloss or use unknown losses layer.
4. the multi-modal face identification method based on deep learning according to claim 1, it is characterised in that:The step (3) in, S0 when S1, S2 ... are not added with the feeding N2 for distinguishing, if their port number is different, can take All it is changed into single pass mode and is normalized to same channels, or to repeat single channel identical to port number, is then re-fed into net It is trained in network structure.
5. the multi-modal face identification method based on deep learning according to claim 1, it is characterised in that:The step (4) in, for registered set and the image of query set, its modality range is within training set modality range.
CN201710122193.1A 2017-03-02 2017-03-02 Multi-mode face recognition method based on deep learning Active CN106909905B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710122193.1A CN106909905B (en) 2017-03-02 2017-03-02 Multi-mode face recognition method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710122193.1A CN106909905B (en) 2017-03-02 2017-03-02 Multi-mode face recognition method based on deep learning

Publications (2)

Publication Number Publication Date
CN106909905A true CN106909905A (en) 2017-06-30
CN106909905B CN106909905B (en) 2020-02-14

Family

ID=59186347

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710122193.1A Active CN106909905B (en) 2017-03-02 2017-03-02 Multi-mode face recognition method based on deep learning

Country Status (1)

Country Link
CN (1) CN106909905B (en)

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107273872A (en) * 2017-07-13 2017-10-20 北京大学深圳研究生院 The depth discrimination net model methodology recognized again for pedestrian in image or video
CN107423701A (en) * 2017-07-17 2017-12-01 北京智慧眼科技股份有限公司 The non-supervisory feature learning method and device of face based on production confrontation network
CN107480206A (en) * 2017-07-25 2017-12-15 杭州电子科技大学 A kind of picture material answering method based on multi-modal low-rank bilinearity pond
CN107506822A (en) * 2017-07-26 2017-12-22 天津大学 A kind of deep neural network method based on Space integration pond
CN107844744A (en) * 2017-10-09 2018-03-27 平安科技(深圳)有限公司 With reference to the face identification method, device and storage medium of depth information
CN108197587A (en) * 2018-01-18 2018-06-22 中科视拓(北京)科技有限公司 A kind of method that multi-modal recognition of face is carried out by face depth prediction
CN108229497A (en) * 2017-07-28 2018-06-29 北京市商汤科技开发有限公司 Image processing method, device, storage medium, computer program and electronic equipment
CN108229440A (en) * 2018-02-06 2018-06-29 北京奥开信息科技有限公司 One kind is based on Multi-sensor Fusion indoor human body gesture recognition method
CN108875777A (en) * 2018-05-03 2018-11-23 浙江大学 Kinds of fibers and blending rate recognition methods in textile fabric based on two-way neural network
CN108960337A (en) * 2018-07-18 2018-12-07 浙江大学 A kind of multi-modal complicated activity recognition method based on deep learning model
CN109284597A (en) * 2018-11-22 2019-01-29 北京旷视科技有限公司 A kind of face unlocking method, device, electronic equipment and computer-readable medium
CN109299639A (en) * 2017-07-25 2019-02-01 虹软(杭州)多媒体信息技术有限公司 A kind of method and apparatus for Expression Recognition
CN109583387A (en) * 2018-11-30 2019-04-05 龙马智芯(珠海横琴)科技有限公司 Identity identifying method and device
CN109815965A (en) * 2019-02-13 2019-05-28 腾讯科技(深圳)有限公司 A kind of image filtering method, device and storage medium
CN109934198A (en) * 2019-03-22 2019-06-25 北京市商汤科技开发有限公司 Face identification method and device
CN110009003A (en) * 2019-03-14 2019-07-12 北京旷视科技有限公司 Training method, the device and system of image procossing and image comparison model
CN110046551A (en) * 2019-03-18 2019-07-23 中国科学院深圳先进技术研究院 A kind of generation method and equipment of human face recognition model
CN110110120A (en) * 2018-06-11 2019-08-09 北方工业大学 A kind of image search method and device based on deep learning
WO2019169942A1 (en) * 2018-03-09 2019-09-12 华南理工大学 Anti-angle and occlusion interference fast face recognition method
CN110458828A (en) * 2019-08-12 2019-11-15 广东工业大学 A kind of laser welding defect identification method and device based on multi-modal fusion network
CN110674677A (en) * 2019-08-06 2020-01-10 厦门大学 Multi-mode multi-layer fusion deep neural network for anti-spoofing of human face
CN111316290A (en) * 2017-11-03 2020-06-19 通用电气公司 System and method for interactive representation learning migration through deep learning of feature ontologies
CN111369567A (en) * 2018-12-26 2020-07-03 腾讯科技(深圳)有限公司 Method and device for segmenting target object in three-dimensional image and electronic equipment
CN111401107A (en) * 2019-01-02 2020-07-10 上海大学 Multi-mode face recognition method based on feature fusion neural network
CN111611909A (en) * 2020-05-18 2020-09-01 桂林电子科技大学 Multi-subspace-domain self-adaptive face recognition method
CN111881706A (en) * 2019-11-27 2020-11-03 马上消费金融股份有限公司 Living body detection, image classification and model training method, device, equipment and medium
CN111937005A (en) * 2020-06-30 2020-11-13 北京小米移动软件有限公司 Biological feature recognition method, device, equipment and storage medium
CN112016524A (en) * 2020-09-25 2020-12-01 北京百度网讯科技有限公司 Model training method, face recognition device, face recognition equipment and medium
CN112149635A (en) * 2020-10-23 2020-12-29 北京百度网讯科技有限公司 Cross-modal face recognition model training method, device, equipment and storage medium
CN112464741A (en) * 2020-11-05 2021-03-09 马上消费金融股份有限公司 Face classification method, model training method, electronic device and storage medium
CN112926557A (en) * 2021-05-11 2021-06-08 北京的卢深视科技有限公司 Method for training multi-mode face recognition model and multi-mode face recognition method
CN113378984A (en) * 2021-07-05 2021-09-10 国药(武汉)医学实验室有限公司 Medical image classification method, system, terminal and storage medium
CN113903053A (en) * 2021-09-26 2022-01-07 厦门大学 Cross-modal pedestrian re-identification method based on unified intermediate modality
CN114627431A (en) * 2022-02-22 2022-06-14 安徽新识智能科技有限公司 Intelligent environment monitoring method and system based on Internet of things
CN114724041A (en) * 2022-06-02 2022-07-08 浙江天铂云科光电股份有限公司 Power equipment infrared chart identification method and system based on deep learning
CN118552794A (en) * 2024-07-25 2024-08-27 湖南军芃科技股份有限公司 Ore sorting identification method based on multichannel training and ore sorting machine

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103530648A (en) * 2013-10-14 2014-01-22 四川空港知觉科技有限公司 Face recognition method based on multi-frame images
CN104966093A (en) * 2015-05-25 2015-10-07 苏州珂锐铁电气科技有限公司 Dynamic texture identification method based on deep neural networks
CN105913025A (en) * 2016-04-12 2016-08-31 湖北工业大学 Deep learning face identification method based on multiple-characteristic fusion
CN106339702A (en) * 2016-11-03 2017-01-18 北京星宇联合投资管理有限公司 Multi-feature fusion based face identification method
CN106446754A (en) * 2015-08-11 2017-02-22 阿里巴巴集团控股有限公司 Image identification method, metric learning method, image source identification method and devices

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103530648A (en) * 2013-10-14 2014-01-22 四川空港知觉科技有限公司 Face recognition method based on multi-frame images
CN104966093A (en) * 2015-05-25 2015-10-07 苏州珂锐铁电气科技有限公司 Dynamic texture identification method based on deep neural networks
CN106446754A (en) * 2015-08-11 2017-02-22 阿里巴巴集团控股有限公司 Image identification method, metric learning method, image source identification method and devices
CN105913025A (en) * 2016-04-12 2016-08-31 湖北工业大学 Deep learning face identification method based on multiple-characteristic fusion
CN106339702A (en) * 2016-11-03 2017-01-18 北京星宇联合投资管理有限公司 Multi-feature fusion based face identification method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MENGYI LIU 等: "Exploiting Feature Hierarchies with Convolutional Neural Networks for Cultural Event Recognition", 《2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS》 *
SHIQING ZHANG 等: "Multimodal Deep Convolutional Neural Network for Multimodal Deep Convolutional Neural Network for Audio-Visual Emotion Recognition", 《ICMR"16》 *

Cited By (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107273872B (en) * 2017-07-13 2020-05-05 北京大学深圳研究生院 Depth discrimination network model method for re-identification of pedestrians in image or video
CN107273872A (en) * 2017-07-13 2017-10-20 北京大学深圳研究生院 The depth discrimination net model methodology recognized again for pedestrian in image or video
WO2019010950A1 (en) * 2017-07-13 2019-01-17 北京大学深圳研究生院 Depth discrimination network model method for pedestrian re-recognition in image or video
CN107423701A (en) * 2017-07-17 2017-12-01 北京智慧眼科技股份有限公司 The non-supervisory feature learning method and device of face based on production confrontation network
CN107480206A (en) * 2017-07-25 2017-12-15 杭州电子科技大学 A kind of picture material answering method based on multi-modal low-rank bilinearity pond
CN109299639A (en) * 2017-07-25 2019-02-01 虹软(杭州)多媒体信息技术有限公司 A kind of method and apparatus for Expression Recognition
CN112861760A (en) * 2017-07-25 2021-05-28 虹软科技股份有限公司 Method and device for facial expression recognition
US11023715B2 (en) * 2017-07-25 2021-06-01 Arcsoft Corporation Limited Method and apparatus for expression recognition
CN107480206B (en) * 2017-07-25 2020-06-12 杭州电子科技大学 Multi-mode low-rank bilinear pooling-based image content question-answering method
CN107506822A (en) * 2017-07-26 2017-12-22 天津大学 A kind of deep neural network method based on Space integration pond
CN107506822B (en) * 2017-07-26 2021-02-19 天津大学 Deep neural network method based on space fusion pooling
CN108229497A (en) * 2017-07-28 2018-06-29 北京市商汤科技开发有限公司 Image processing method, device, storage medium, computer program and electronic equipment
CN107844744A (en) * 2017-10-09 2018-03-27 平安科技(深圳)有限公司 With reference to the face identification method, device and storage medium of depth information
CN111316290B (en) * 2017-11-03 2024-01-12 通用电气公司 System and method for interactive representation learning migration through deep learning of feature ontologies
CN111316290A (en) * 2017-11-03 2020-06-19 通用电气公司 System and method for interactive representation learning migration through deep learning of feature ontologies
CN108197587B (en) * 2018-01-18 2021-08-03 中科视拓(北京)科技有限公司 Method for performing multi-mode face recognition through face depth prediction
CN108197587A (en) * 2018-01-18 2018-06-22 中科视拓(北京)科技有限公司 A kind of method that multi-modal recognition of face is carried out by face depth prediction
CN108229440A (en) * 2018-02-06 2018-06-29 北京奥开信息科技有限公司 One kind is based on Multi-sensor Fusion indoor human body gesture recognition method
US11417147B2 (en) 2018-03-09 2022-08-16 South China University Of Technology Angle interference resistant and occlusion interference resistant fast face recognition method
WO2019169942A1 (en) * 2018-03-09 2019-09-12 华南理工大学 Anti-angle and occlusion interference fast face recognition method
CN108875777B (en) * 2018-05-03 2022-03-15 浙江大学 Method for identifying fiber types and blending proportion in textile fabric based on double-path neural network
CN108875777A (en) * 2018-05-03 2018-11-23 浙江大学 Kinds of fibers and blending rate recognition methods in textile fabric based on two-way neural network
CN110110120B (en) * 2018-06-11 2021-05-25 北方工业大学 Image retrieval method and device based on deep learning
CN110110120A (en) * 2018-06-11 2019-08-09 北方工业大学 A kind of image search method and device based on deep learning
CN108960337A (en) * 2018-07-18 2018-12-07 浙江大学 A kind of multi-modal complicated activity recognition method based on deep learning model
CN108960337B (en) * 2018-07-18 2020-07-17 浙江大学 Multi-modal complex activity recognition method based on deep learning model
CN109284597A (en) * 2018-11-22 2019-01-29 北京旷视科技有限公司 A kind of face unlocking method, device, electronic equipment and computer-readable medium
CN109583387A (en) * 2018-11-30 2019-04-05 龙马智芯(珠海横琴)科技有限公司 Identity identifying method and device
CN111369567A (en) * 2018-12-26 2020-07-03 腾讯科技(深圳)有限公司 Method and device for segmenting target object in three-dimensional image and electronic equipment
CN111401107A (en) * 2019-01-02 2020-07-10 上海大学 Multi-mode face recognition method based on feature fusion neural network
CN111401107B (en) * 2019-01-02 2023-08-18 上海大学 Multi-mode face recognition method based on feature fusion neural network
CN109815965B (en) * 2019-02-13 2021-07-06 腾讯科技(深圳)有限公司 Image filtering method and device and storage medium
CN109815965A (en) * 2019-02-13 2019-05-28 腾讯科技(深圳)有限公司 A kind of image filtering method, device and storage medium
CN110009003A (en) * 2019-03-14 2019-07-12 北京旷视科技有限公司 Training method, the device and system of image procossing and image comparison model
CN110046551A (en) * 2019-03-18 2019-07-23 中国科学院深圳先进技术研究院 A kind of generation method and equipment of human face recognition model
WO2020192112A1 (en) * 2019-03-22 2020-10-01 北京市商汤科技开发有限公司 Facial recognition method and apparatus
CN109934198A (en) * 2019-03-22 2019-06-25 北京市商汤科技开发有限公司 Face identification method and device
CN109934198B (en) * 2019-03-22 2021-05-14 北京市商汤科技开发有限公司 Face recognition method and device
CN110674677A (en) * 2019-08-06 2020-01-10 厦门大学 Multi-mode multi-layer fusion deep neural network for anti-spoofing of human face
CN110458828B (en) * 2019-08-12 2023-02-10 广东工业大学 Laser welding defect identification method and device based on multi-mode fusion network
CN110458828A (en) * 2019-08-12 2019-11-15 广东工业大学 A kind of laser welding defect identification method and device based on multi-modal fusion network
CN111881706A (en) * 2019-11-27 2020-11-03 马上消费金融股份有限公司 Living body detection, image classification and model training method, device, equipment and medium
CN111611909A (en) * 2020-05-18 2020-09-01 桂林电子科技大学 Multi-subspace-domain self-adaptive face recognition method
WO2022000334A1 (en) * 2020-06-30 2022-01-06 北京小米移动软件有限公司 Biological feature recognition method and apparatus, and device and storage medium
CN111937005A (en) * 2020-06-30 2020-11-13 北京小米移动软件有限公司 Biological feature recognition method, device, equipment and storage medium
CN112016524A (en) * 2020-09-25 2020-12-01 北京百度网讯科技有限公司 Model training method, face recognition device, face recognition equipment and medium
CN112016524B (en) * 2020-09-25 2023-08-08 北京百度网讯科技有限公司 Model training method, face recognition device, equipment and medium
CN112149635A (en) * 2020-10-23 2020-12-29 北京百度网讯科技有限公司 Cross-modal face recognition model training method, device, equipment and storage medium
CN112464741B (en) * 2020-11-05 2021-11-26 马上消费金融股份有限公司 Face classification method, model training method, electronic device and storage medium
CN112464741A (en) * 2020-11-05 2021-03-09 马上消费金融股份有限公司 Face classification method, model training method, electronic device and storage medium
CN112926557A (en) * 2021-05-11 2021-06-08 北京的卢深视科技有限公司 Method for training multi-mode face recognition model and multi-mode face recognition method
CN113378984B (en) * 2021-07-05 2023-05-02 国药(武汉)医学实验室有限公司 Medical image classification method, system, terminal and storage medium
CN113378984A (en) * 2021-07-05 2021-09-10 国药(武汉)医学实验室有限公司 Medical image classification method, system, terminal and storage medium
CN113903053A (en) * 2021-09-26 2022-01-07 厦门大学 Cross-modal pedestrian re-identification method based on unified intermediate modality
CN113903053B (en) * 2021-09-26 2024-06-07 厦门大学 Cross-mode pedestrian re-identification method based on unified intermediate mode
CN114627431A (en) * 2022-02-22 2022-06-14 安徽新识智能科技有限公司 Intelligent environment monitoring method and system based on Internet of things
CN114724041A (en) * 2022-06-02 2022-07-08 浙江天铂云科光电股份有限公司 Power equipment infrared chart identification method and system based on deep learning
CN118552794A (en) * 2024-07-25 2024-08-27 湖南军芃科技股份有限公司 Ore sorting identification method based on multichannel training and ore sorting machine

Also Published As

Publication number Publication date
CN106909905B (en) 2020-02-14

Similar Documents

Publication Publication Date Title
CN106909905A (en) A kind of multi-modal face identification method based on deep learning
CN110111340A (en) The Weakly supervised example dividing method cut based on multichannel
JP6891351B2 (en) How to generate a human hairstyle based on multi-feature search and deformation
CN109145939B (en) Semantic segmentation method for small-target sensitive dual-channel convolutional neural network
CN113657349B (en) Human behavior recognition method based on multi-scale space-time diagram convolutional neural network
CN107766850B (en) Face recognition method based on combination of face attribute information
CN108182441B (en) Parallel multichannel convolutional neural network, construction method and image feature extraction method
CN105069400B (en) Facial image gender identifying system based on the sparse own coding of stack
CN109711281A (en) A kind of pedestrian based on deep learning identifies again identifies fusion method with feature
CN111310668B (en) Gait recognition method based on skeleton information
CN105095870B (en) Pedestrian based on transfer learning recognition methods again
CN105373777A (en) Face recognition method and device
CN113297955B (en) Sign language word recognition method based on multi-mode hierarchical information fusion
CN111291604A (en) Face attribute identification method, device, storage medium and processor
CN113255728A (en) Depression classification method based on map embedding and multi-modal brain network
CN109815826A (en) The generation method and device of face character model
CN109299701A (en) Expand the face age estimation method that more ethnic group features cooperate with selection based on GAN
KR20200075114A (en) System and Method for Matching Similarity between Image and Text
CN107424161B (en) Coarse-to-fine indoor scene image layout estimation method
CN110046550A (en) Pedestrian's Attribute Recognition system and method based on multilayer feature study
CN109409240A (en) A kind of SegNet remote sensing images semantic segmentation method of combination random walk
CN110321862B (en) Pedestrian re-identification method based on compact ternary loss
CN109558902A (en) A kind of fast target detection method
CN104240256A (en) Image salient detecting method based on layering sparse modeling
CN114613013A (en) End-to-end human behavior recognition method and model based on skeleton nodes

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant