WO2020155606A1 - 面部识别方法及装置、电子设备和存储介质 - Google Patents
面部识别方法及装置、电子设备和存储介质 Download PDFInfo
- Publication number
- WO2020155606A1 WO2020155606A1 PCT/CN2019/100859 CN2019100859W WO2020155606A1 WO 2020155606 A1 WO2020155606 A1 WO 2020155606A1 CN 2019100859 W CN2019100859 W CN 2019100859W WO 2020155606 A1 WO2020155606 A1 WO 2020155606A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- features
- attention
- cluster
- attribute
- target object
- Prior art date
Links
- 230000001815 facial effect Effects 0.000 title claims abstract description 86
- 238000000034 method Methods 0.000 title claims abstract description 78
- 238000003860 storage Methods 0.000 title claims abstract description 27
- 230000007246 mechanism Effects 0.000 claims abstract description 16
- 238000012545 processing Methods 0.000 claims description 51
- 238000000605 extraction Methods 0.000 claims description 34
- 238000013528 artificial neural network Methods 0.000 claims description 22
- 238000004590 computer program Methods 0.000 claims description 16
- 238000012937 correction Methods 0.000 claims description 14
- 238000012549 training Methods 0.000 claims description 12
- 230000008569 process Effects 0.000 claims description 8
- 230000003595 spectral effect Effects 0.000 claims description 6
- 238000010586 diagram Methods 0.000 description 20
- 230000006870 function Effects 0.000 description 11
- 238000004891 communication Methods 0.000 description 10
- 230000000694 effects Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 7
- 238000013527 convolutional neural network Methods 0.000 description 6
- 238000005457 optimization Methods 0.000 description 6
- 230000009471 action Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 5
- 239000011521 glass Substances 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 210000004709 eyebrow Anatomy 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/169—Holistic features and representations, i.e. based on the facial image taken as a whole
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2323—Non-hierarchical techniques based on graph theory, e.g. minimum spanning trees [MST] or graph cuts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
Definitions
- the present disclosure relates to the field of computer technology but is not limited to the field of computers, and in particular to a facial recognition method and device, electronic equipment and storage medium.
- Facial attribute prediction has a wide range of applications. For example, it is an extremely important part of the surveillance and security field. Effectively predicting the person's gender, age, whether to wear dangerous objects and other attributes, play an extremely important role in the application of facial attribute prediction. Correct attribute prediction can further improve the accuracy of facial recognition, so that facial recognition can be more widely used in various application scenarios.
- the present disclosure proposes a technical solution for facial recognition.
- a facial recognition method which includes: extracting attribute characteristics of a to-be-processed image including a target object to obtain N attribute characteristics of the target object, where N is an integer greater than 1;
- the force mechanism extracts attention features of the image to be processed to obtain N attention features of the target object; performs cluster processing on the N attention features to obtain M cluster attention features, where M is A positive integer and M ⁇ N; according to the N attribute features and the M cluster attention features, the facial recognition result of the target object is determined.
- performing cluster processing on the N attention features to obtain M cluster attention features includes: performing cluster processing on the N attention features to obtain M cluster attention features.
- the cluster set, each attention feature corresponds to one of the M cluster sets; the cluster attention features of each cluster set are determined respectively, and M cluster attention features are obtained.
- the method further includes: multiplying the N attribute features and the N attention features to obtain enhanced N attribute features,
- determining the facial recognition result of the target object according to the N attribute features and the M cluster attention features includes: comparing the enhanced N attribute features according to the M cluster attention features Respectively, the facial recognition result of the target object is obtained.
- determining the facial recognition result of the target object according to the N attribute features and the M cluster attention features includes: according to the M cluster attention feature pairs The N attribute features are respectively corrected to obtain the facial recognition result.
- the enhanced N attribute features are respectively modified according to the M cluster attention features to obtain the face recognition result of the target object, including: combining the enhanced N attributes
- the features are respectively multiplied by the cluster attention features corresponding to at least some of the N attribute features to obtain the facial recognition result.
- correcting the N attribute features according to the M cluster attention features to obtain the facial recognition result of the target object includes: combining the N attribute features with The cluster attention features corresponding to at least part of the N attribute features are respectively multiplied to obtain the facial recognition result.
- the method is implemented by a neural network.
- the neural network includes a multi-task convolutional network, multiple individual attention networks, and a clustering network.
- the multi-task convolutional network is used to The image to be processed performs attribute feature extraction, the multiple individual attention networks are used to perform attention feature extraction on the image to be processed, and the clustering network is used to perform clustering processing on the N attention features .
- the method further includes: in the process of training the neural network, adjusting the network parameters of multiple individual attention networks according to the network loss of the clustering network.
- the clustering processing includes spectral clustering
- the M cluster attention features are respectively cluster centers of the M cluster sets.
- a facial recognition device including: an attribute extraction module for performing attribute feature extraction on a to-be-processed image including a target object to obtain N attribute characteristics of the target object, where N is An integer greater than 1; an attention extraction module for extracting attention features of the image to be processed based on an attention mechanism to obtain N attention features of the target object; a clustering module for analyzing the N Attention features are clustered to obtain M cluster attention features, where M is a positive integer and M ⁇ N; the result determination module is used to determine the N attribute features and the M cluster attention features To determine the facial recognition result of the target object.
- the clustering module includes: a clustering sub-module, configured to perform clustering processing on the N attention features to obtain M cluster sets, and each attention feature is associated with M One of the cluster sets corresponds to one of the cluster sets; the feature determination sub-module is used to determine the cluster attention features of each cluster set to obtain M cluster attention features.
- the device further includes: an attribute enhancement module, configured to multiply the N attribute features and the N attention features to obtain enhanced N attribute features, where ,
- the result determining module includes:
- the first correction sub-module is configured to respectively correct the enhanced N attribute features according to the M cluster attention features to obtain the face recognition result of the target object.
- the result determination module includes: a second correction submodule, configured to respectively correct the N attribute features according to the M cluster attention features to obtain the facial recognition result.
- the first correction sub-module includes: a first multiplication sub-module, configured to compare the enhanced N attribute features with at least part of the N attribute features The cluster attention features are respectively multiplied to obtain the facial recognition result.
- the second correction submodule includes: a second multiplication submodule, configured to combine the N attribute features with at least part of the N attribute features corresponding to the aggregation The similar attention features are respectively multiplied to obtain the facial recognition result.
- the device is implemented by a neural network.
- the neural network includes a multi-task convolutional network, multiple individual attention networks, and a clustering network.
- the multi-task convolutional network is used to The image to be processed performs attribute feature extraction, the multiple individual attention networks are used to perform attention feature extraction on the image to be processed, and the clustering network is used to perform clustering processing on the N attention features .
- the device further includes: a parameter adjustment module, which is used to adjust the network of multiple individual attention networks according to the network loss of the clustering network during the training of the neural network parameter.
- the clustering processing includes spectral clustering
- the M cluster attention features are respectively cluster centers of the M cluster sets.
- an electronic device including: a processor; a memory for storing executable instructions of the processor; wherein the processor is configured to execute the above method.
- a computer-readable storage medium having computer program instructions stored thereon, and the computer program instructions implement the above method when executed by a processor.
- a computer program product is executed by a processor to implement the above method.
- attribute feature extraction and attention feature extraction can be performed on the image to be processed respectively to obtain multiple attribute features and attention features; the attention feature is clustered to obtain cluster attention features, and based on multiple attributes Features and clustering attention features determine the results of facial recognition, extracting attention features through multi-attention mechanisms and clustering similar attention features to optimize different local features and improve the recognition effect of facial attributes.
- Fig. 1 shows a flowchart of a face recognition method according to an embodiment of the present disclosure.
- Fig. 2 shows a schematic diagram of an application example of a face recognition method according to an embodiment of the present disclosure.
- Fig. 3 shows a comparison diagram of lip attention characteristics before and after optimization according to the present disclosure.
- Fig. 4 shows a block diagram of a facial recognition device according to an embodiment of the present disclosure.
- Fig. 5 shows a block diagram of an electronic device according to an embodiment of the present disclosure.
- Fig. 6 shows a block diagram of an electronic device according to an embodiment of the present disclosure.
- FIG. 1 shows a flowchart of a facial recognition method 100 according to an embodiment of the present disclosure.
- the facial recognition method 100 includes: in step 102, performing attribute feature extraction on a to-be-processed image including a target object to obtain N attribute characteristics of the target object, where N is an integer greater than one.
- step 104 attention feature extraction is performed on the image to be processed based on the attention mechanism to obtain N attention features of the target object.
- step 106 perform clustering processing on the N attention features to obtain M cluster attention features, where M is a positive integer and M ⁇ N.
- the facial recognition result of the target object is determined according to the N attribute features and the M cluster attention features.
- the method provided by the embodiments of this application can be applied to various image processing devices.
- the image processing device can recognize the face of the image target object through the execution of step 102, step 104, step 106, and step 108, for example, face recognition, etc. .
- the image processing equipment can be a security equipment, for example, equipment installed at the gate of a community, school, factory area, residence, etc., where monitoring is required.
- the image processing device may also be applied to a non-security field.
- the image processing device may be a ticketing device that performs ticket checking through facial recognition.
- the image processing device may be a payment device, which determines whether to make a payment based on the facial recognition result.
- the facial recognition method of the present application can be applied to various scenes that require facial recognition to obtain facial recognition results.
- attribute feature extraction and attention feature extraction on the image to be processed respectively to obtain multiple attribute features and attention features; cluster the attention features to obtain cluster attention features, and according to multiple attributes
- Features and clustering attention features determine the results of facial recognition, extracting attention features through multi-attention mechanisms and clustering similar attention features to optimize different local features and improve the recognition effect of facial attributes.
- the features required for different tasks may not be the features of the entire face, but only the partial features of the face. For example, if it is predicted whether a person wears glasses, only separate eye information may be needed, and the extra information may interfere with the results. Therefore, attention features can be added to improve the accuracy of prediction.
- the attention feature may be a feature specified to be extracted, and may be one or more of the attribute features.
- the attribute feature may be an overall feature and/or a local feature of the target object.
- the overall feature includes, but is not limited to, the overall feature of the face of the target object; the local feature may be a local feature in the face, for example, the feature of the eyes, the feature of the lips, and the like.
- facial attributes for example, multiple attributes related to the face such as gender, age, and wearing of a person
- multiple attributes can be jointly recognized and shared features.
- attention mechanism Attention Mechanism
- important features such as ears, mouth, nose, etc.
- the image to be processed including the target object may be subjected to attribute feature extraction to obtain N attribute characteristics of the target object.
- the target object may be, for example, a person in the image, and the image to be processed may be, for example, a face image including the target object.
- the convolutional neural network CNN can be used as the base network (Base Net) to perform feature extraction on the image to be processed to obtain the facial feature map; and then through, for example, the Multi-Task Convolution Neural Net (MTCNN)
- MTCNN Multi-Task Convolution Neural Net
- the obtained facial feature map is subjected to attribute feature extraction, thereby obtaining N attribute features of the target object.
- the multi-task convolutional neural network can use different types of networks such as VGG16, residual network, etc. The present disclosure does not limit the network types of the multi-task convolutional neural network.
- attention feature extraction may be performed on the image to be processed based on the attention mechanism to obtain N attention features of the target object.
- feature extraction of the image to be processed may be performed through a basic network to obtain a facial feature map, thereby achieving feature sharing; and then through multiple individual attention networks (Individual Attention Network, IAN), the facial feature map may be extracted with attention features.
- IAN Intelligent Attention Network
- the N attention features may be clustered in step 106 to obtain M cluster attention features.
- some facial features can be learned better, such as glasses, nose, etc.
- some fine-grained features such as earrings and eyebrows are not easy to learn. Therefore, it is possible to cluster all the similar parts of the attention features together, and optimize, especially correct, some of the original attention features that have poor learning effects.
- step 106 may include:
- the cluster attention features of each cluster set are determined respectively, and M cluster attention features are obtained.
- N attention features can be clustered into M categories (M ⁇ N), and after clustering processing, each attention feature corresponds to one of the M categories. After clustering, M cluster sets can be obtained. The cluster attention features of each cluster set are extracted separately, and M cluster attention features can be obtained. In this way, M cluster attention features can be used as representatives of N attention features, and then the facial recognition result of the target object can be determined according to the N attribute features and M cluster attention features.
- the clustering process may be spectral clustering, and the M cluster attention features are the cluster centers of each cluster set in the M cluster sets.
- all values in each cluster set can be averaged to obtain the cluster center.
- training can be performed through a preset training set to learn the corresponding weight of each attention feature, and each attention feature in the cluster set is weighted and averaged to obtain the combination of the clusters. Class center.
- the present disclosure does not limit the specific method of clustering.
- the attention features that are well-learned can help other attention features that are not well-learned, and the effect of the attention mechanism can be improved through the mutual help of related attention features.
- the facial recognition result of the target object may be determined according to the N attribute characteristics and the M cluster attention characteristics.
- the obtained N attribute features and M cluster attention features can be comprehensively considered through multiplication and other methods, and other methods can also be used for further processing to determine the face recognition result of the target object.
- step 108 may include:
- the N attribute features are respectively corrected according to the M cluster attention features to obtain the facial recognition result.
- M cluster attention features can be used to directly modify N attribute features.
- the step of correcting the N attribute features based on the M cluster attention features may include: combining the N attribute features with clusters corresponding to at least some of the N attribute features. The similar attention features are respectively multiplied to obtain the facial recognition result.
- At least some of the N attribute features in this embodiment include: each of the N attribute features, and fewer than N attribute features.
- the method may further include: multiplying the N attribute features and the N attention features to obtain enhanced N attribute features;
- Step 108 may include: respectively correcting the enhanced N attribute features according to the M cluster attention features to obtain the face recognition result of the target object.
- the N attention features and the N attribute features are respectively corrected to obtain the face recognition result of the target object.
- the step of correcting the enhanced N attribute features according to the M cluster attention features may include: combining the enhanced N attribute features with at least part of the N attribute features. The corresponding cluster attention features are multiplied respectively to obtain the facial recognition result.
- Fig. 2 shows a schematic diagram of an application example of a face recognition method according to an embodiment of the present disclosure.
- the method 200 is an application example of the facial recognition method 100.
- feature extraction can be performed on the image to be processed through the basic network to obtain a facial feature map; the attribute feature extraction on the facial feature map can obtain N attribute features 202; attention to the facial feature map Force feature extraction, N attention features 204 are obtained; N attention features 204 are clustered, and M cluster sets are obtained.
- Each of the N attention features corresponds to one of the M cluster sets; the cluster attention features of each cluster set are determined respectively, and M cluster attention features can be obtained .
- the N attribute features 202 are directly corrected (S208), and the facial recognition result of the target object can be determined .
- the predecessor of IAN clustering can be used to correct the facial recognition results of the target object, and the difficult attention features can be optimized through the attention features that are easy to learn.
- the N attention features 204 can be further multiplied by the N attribute features 202 to obtain enhanced N attribute features. Furthermore, according to the M cluster attention features, the enhanced N attribute features are respectively corrected to obtain the face recognition result of the target object. As a result, the attention feature can be further emphasized, and the attention feature that is not easy can be optimized through the attention feature that is easy to learn.
- the attention mechanism can be improved through mutual help of related attributes.
- the method can be implemented by a neural network.
- the neural network includes a multi-task convolutional network, multiple separate attention networks, and a clustering network.
- the multi-task convolutional network is used to Performing attribute feature extraction on the to-be-processed image
- the multiple individual attention networks are used to perform attention feature extraction on the to-be-processed image
- the clustering network is used to cluster the N attention features deal with.
- the multi-task convolutional network is used to extract the attribute features 202 of the image to be processed (facial feature map).
- MTCNN multi-task convolutional network
- VGG16 and Residual Network networks of different sizes, such as VGG16 and Residual Network, can be used to deal with different application scenarios.
- IAN Multiple individual attention networks
- attention features 204 from the image to be processed (facial feature map).
- facial feature map For example, the glasses and lips of a person are respectively targeted.
- IAN a simple convolutional neural network can be used for learning, or an attention mechanism such as residual attention can be used.
- the clustering network is used for clustering N attention features to obtain M cluster attention features.
- the present disclosure does not limit the network type of the clustering network.
- the method further includes: in the process of training the neural network, adjusting the network parameters of multiple individual attention networks according to the network loss of the clustering network.
- the aforementioned neural networks can be trained according to a preset training set.
- combined tasks are used for optimization.
- no artificial links may be involved, and independent learning is used to achieve optimization.
- the network parameters of multiple individual attention networks can be adjusted according to the network loss of the clustering network, so that the back propagation of the neural network can be used to optimize the individual attention network.
- a multi-task attribute learning network based on MTCNN can be used to train N attributes of the facial feature map at the same time to obtain N attribute features 202.
- the N attribute features 202 obtained through MTCNN are feature tensors, which can be expressed as (N, C, W, H).
- the value of (N, C, W, H) is determined by the specific neural network.
- N, C, W, and H are all natural numbers, and N represents the number of feature tensors; C represents the number of feature maps obtained, also known as the number of channels, which is usually much greater than 3; W represents the width of the feature map; H represents the feature The length of the graph.
- N attribute features 202 can be expressed as N attribute features (C, W, H), that is, N attribute feature tensors (number of channels, width, length). Subsequently, the N attribute features 202 can be used to predict the N facial attributes as a result of facial recognition.
- IAN may be used to train N attributes from a facial feature map, and an attention map (attention map) may be used to learn N attributes, so as to obtain N attention features 204.
- an attention map attention map
- each attention feature corresponds to a two-dimensional attention feature image (Mask)
- the feature tensor A of the obtained attention feature 204 can be expressed as (N, 1, W, H), that is, the number of channels here Is 1. Therefore, the N attention features 204 can be expressed as N feature tensors A(W, H), that is, N feature tensors A (width, length).
- N attention feature tensors A (width, length) can be multiplied by N attribute feature tensors, so that N attribute features The relevant features in the tensor are emphasized.
- the N attention features are clustered to obtain M cluster sets, which can be expressed as C 1 ,..., C M.
- the cluster centers extracted from each of the M cluster sets can be expressed as the first cluster attention feature (A x1 ,A y1 ),..., the Mth cluster attention feature (A xm ,A ym ).
- the above M cluster attention features can be expressed as (M, 1, W, H), that is, M two-dimensional tensors X _1 (width, length),..., X _M (width, length).
- N attribute feature tensors F (number of channels, width, length) obtained by the correction of N attention features A, determine the corresponding cluster class centers, and then use the corresponding M two-dimensional tensors X The center of the class makes the final modification to the N attribute features F. Since these M tensors X come from the output of the previous IAN, through such learning, IAN can also be optimized at the same time, and multiple attention features can be used to modify the N attribute features F at the same time.
- the correction method is to simultaneously multiply N attribute features F (number of channels, width, length) by M clustered tensors X _m , where m is in the range of [1,M] The natural number within. From this, the multiplied tensor FX (M, number of channels, width, length) can be obtained. Expand the multiplied tensor FX (M, number of channels, width, length) into (M ⁇ number of channels, width, length), and finally use the expanded results to predict features to obtain the final facial recognition result.
- FIG. 3 shows a comparison diagram of the lip attention characteristics before and after optimization according to the present disclosure.
- the lip attention feature image before optimization has more noise.
- the attention features of the lips can be better concentrated on the lips, and the noise is reduced.
- the facial recognition method of the embodiments of the present disclosure when training and predicting face attributes, it is possible to independently optimize the attention features and increase the strength of the training model to improve the accuracy of prediction, thereby better predicting the fine-grained attributes of the face , Such as whether it is equipped with a hat, necklace, whether to carry headphones, etc. According to the embodiments of the present disclosure, it can be applied to face attribute recognition in the fields of surveillance and security, etc., to improve the recognition rate of face attributes and improve the recognition effect.
- the present disclosure also provides facial recognition devices, electronic equipment, computer-readable storage media, and programs, all of which can be used to implement any facial recognition method provided in the present disclosure.
- facial recognition devices electronic equipment, computer-readable storage media, and programs, all of which can be used to implement any facial recognition method provided in the present disclosure.
- the writing order of each step does not mean a strict execution order, and the specific execution order of each step should be determined by its function and possible internal logic.
- Fig. 4 shows a block diagram of a facial recognition device according to an embodiment of the present disclosure. As shown in Fig. 4, the device includes:
- the attribute extraction module 41 is configured to perform attribute feature extraction on the image to be processed including the target object to obtain N attribute characteristics of the target object, where N is an integer greater than 1;
- the attention extraction module 42 is configured to perform attention feature extraction on the image to be processed based on the attention mechanism to obtain N attention features of the target object;
- the clustering module 43 is configured to perform clustering processing on the N attention features to obtain M cluster attention features, where M is a positive integer and M ⁇ N;
- the result determination module 44 is configured to determine the facial recognition result of the target object according to the N attribute features and the M cluster attention features.
- the clustering module includes: a clustering sub-module, configured to perform clustering processing on the N attention features to obtain M cluster sets, and each attention feature is associated with M One of the cluster sets corresponds to one of the cluster sets; the feature determination sub-module is used to determine the cluster attention features of each cluster set to obtain M cluster attention features.
- the device further includes: an attribute enhancement module, configured to multiply the N attribute features and the N attention features to obtain enhanced N attribute features, where ,
- the result determining module includes:
- the first correction sub-module is configured to respectively correct the enhanced N attribute features according to the M cluster attention features to obtain the face recognition result of the target object.
- the result determination module includes: a second correction submodule, configured to respectively correct the N attribute features according to the M cluster attention features to obtain the facial recognition result.
- the first correction submodule includes: a first multiplication submodule, configured to combine the enhanced N attribute features with the cluster attention corresponding to each of the attribute features The features are respectively multiplied to obtain the facial recognition result.
- the second correction sub-module includes: a second multiplication sub-module, configured to combine the N attribute features with the cluster attention features corresponding to each of the attribute features Respectively multiply to obtain the facial recognition result.
- the device is implemented by a neural network.
- the neural network includes a multi-task convolutional network, multiple individual attention networks, and a clustering network.
- the multi-task convolutional network is used to The image to be processed performs attribute feature extraction, the multiple individual attention networks are used to perform attention feature extraction on the image to be processed, and the clustering network is used to perform clustering processing on the N attention features .
- the device further includes: a parameter adjustment module, which is used to adjust the network of multiple individual attention networks according to the network loss of the clustering network during the training of the neural network parameter.
- the clustering processing includes spectral clustering
- the M cluster attention features are respectively cluster centers of the M cluster sets.
- the functions or modules contained in the apparatus provided in the embodiments of the present disclosure can be used to execute the methods described in the above method embodiments.
- the functions or modules contained in the apparatus provided in the embodiments of the present disclosure can be used to execute the methods described in the above method embodiments.
- the embodiments of the present disclosure also provide a computer-readable storage medium on which computer program instructions are stored, and the computer program instructions implement the foregoing method when executed by a processor.
- the computer-readable storage medium may be a non-volatile computer-readable storage medium.
- An embodiment of the present disclosure also provides an electronic device, including: a processor; a memory for storing executable instructions of the processor; wherein the processor is configured as the above method.
- the electronic device can be provided as a terminal, server or other form of device.
- the embodiments of the present disclosure also provide a computer program product, which implements the above method after being executed by a processor.
- FIG. 5 shows a block diagram of an electronic device 800 according to an embodiment of the present disclosure.
- the electronic device 800 may be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, and other terminals.
- the electronic device 800 may include one or more of the following components: a processing component 802, a memory 804, a power component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812, and a sensor component 814 , And communication component 816.
- the processing component 802 generally controls the overall operations of the electronic device 800, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
- the processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the foregoing method.
- the processing component 802 may include one or more modules to facilitate the interaction between the processing component 802 and other components.
- the processing component 802 may include a multimedia module to facilitate the interaction between the multimedia component 808 and the processing component 802.
- the memory 804 is configured to store various types of data to support operations in the electronic device 800. Examples of these data include instructions for any application or method operating on the electronic device 800, contact data, phone book data, messages, pictures, videos, etc.
- the memory 804 can be implemented by any type of volatile or non-volatile storage device or their combination, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable and Programmable Read Only Memory (EPROM), Programmable Read Only Memory (PROM), Read Only Memory (ROM), Magnetic Memory, Flash Memory, Magnetic Disk or Optical Disk.
- SRAM static random access memory
- EEPROM electrically erasable programmable read-only memory
- EPROM erasable and Programmable Read Only Memory
- PROM Programmable Read Only Memory
- ROM Read Only Memory
- Magnetic Memory Flash Memory
- Magnetic Disk Magnetic Disk or Optical Disk.
- the power supply component 806 provides power for various components of the electronic device 800.
- the power supply component 806 may include a power management system, one or more power supplies, and other components associated with the generation, management, and distribution of power for the electronic device 800.
- the multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user.
- the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
- the touch panel includes one or more touch sensors to sense touch, sliding, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure related to the touch or slide operation.
- the multimedia component 808 includes a front camera and/or a rear camera. When the electronic device 800 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
- the audio component 810 is configured to output and/or input audio signals.
- the audio component 810 includes a microphone (MIC).
- the microphone is configured to receive an external audio signal.
- the received audio signal may be further stored in the memory 804 or transmitted via the communication component 816.
- the audio component 810 further includes a speaker for outputting audio signals.
- the I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module.
- the peripheral interface module may be a keyboard, a click wheel, a button, and the like. These buttons may include but are not limited to: home button, volume button, start button, and lock button.
- the sensor component 814 includes one or more sensors for providing the electronic device 800 with various aspects of state evaluation.
- the sensor component 814 can detect the on/off status of the electronic device 800 and the relative positioning of the components.
- the component is the display and the keypad of the electronic device 800.
- the sensor component 814 can also detect the electronic device 800 or the electronic device 800.
- the position of the component changes, the presence or absence of contact between the user and the electronic device 800, the orientation or acceleration/deceleration of the electronic device 800, and the temperature change of the electronic device 800.
- the sensor component 814 may include a proximity sensor configured to detect the presence of nearby objects when there is no physical contact.
- the sensor component 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
- the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- the communication component 816 is configured to facilitate wired or wireless communication between the electronic device 800 and other devices.
- the electronic device 800 can access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof.
- the communication component 816 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel.
- the communication component 816 further includes a near field communication (NFC) module to facilitate short-range communication.
- the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
- RFID radio frequency identification
- IrDA infrared data association
- UWB ultra-wideband
- Bluetooth Bluetooth
- the electronic device 800 can be implemented by one or more application specific integrated circuits (ASIC), digital signal processors (DSP), digital signal processing devices (DSPD), programmable logic devices (PLD), field A programmable gate array (FPGA), controller, microcontroller, microprocessor, or other electronic components are implemented to implement the above methods.
- ASIC application specific integrated circuits
- DSP digital signal processors
- DSPD digital signal processing devices
- PLD programmable logic devices
- FPGA field A programmable gate array
- controller microcontroller, microprocessor, or other electronic components are implemented to implement the above methods.
- a non-volatile computer-readable storage medium such as the memory 804 including computer program instructions, which can be executed by the processor 820 of the electronic device 800 to complete the foregoing method.
- FIG. 6 shows a block diagram of an electronic device 1900 according to an embodiment of the present disclosure.
- the electronic device 1900 may be provided as a server. 6
- the electronic device 1900 includes a processing component 1922, which further includes one or more processors, and a memory resource represented by a memory 1932, for storing instructions executable by the processing component 1922, such as application programs.
- the application program stored in the memory 1932 may include one or more modules each corresponding to a set of instructions.
- the processing component 1922 is configured to execute instructions to perform the above-described methods.
- the electronic device 1900 may also include a power component 1926 configured to perform power management of the electronic device 1900, a wired or wireless network interface 1950 configured to connect the electronic device 1900 to a network, and an input output (I/O) interface 1958 .
- the electronic device 1900 can operate based on an operating system stored in the memory 1932, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or the like.
- a non-volatile computer-readable storage medium such as the memory 1932 including computer program instructions, which can be executed by the processing component 1922 of the electronic device 1900 to complete the foregoing method.
- the present disclosure may be a system, method, and/or computer program product.
- the computer program product may include a computer-readable storage medium loaded with computer-readable program instructions for enabling a processor to implement various aspects of the present disclosure.
- the computer-readable storage medium may be a tangible device that can hold and store instructions used by the instruction execution device.
- the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Non-exhaustive list of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) Or flash memory), static random access memory (SRAM), portable compact disk read-only memory (CD-ROM), digital versatile disk (DVD), memory stick, floppy disk, mechanical encoding device, such as a printer with instructions stored thereon
- RAM random access memory
- ROM read only memory
- EPROM erasable programmable read only memory
- flash memory flash memory
- SRAM static random access memory
- CD-ROM compact disk read-only memory
- DVD digital versatile disk
- memory stick floppy disk
- mechanical encoding device such as a printer with instructions stored thereon
- the computer-readable storage medium used here is not interpreted as a transient signal itself, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (for example, light pulses through fiber optic cables), or through wires Transmission of electrical signals.
- the computer-readable program instructions described herein can be downloaded from a computer-readable storage medium to various computing/processing devices, or downloaded to an external computer or external storage device via a network, such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, optical fiber transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
- the network adapter card or network interface in each computing/processing device receives computer-readable program instructions from the network, and forwards the computer-readable program instructions for storage in the computer-readable storage medium in each computing/processing device .
- the computer program instructions used to perform the operations of the present disclosure may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine-related instructions, microcode, firmware instructions, state setting data, or in one or more programming languages.
- Source code or object code written in any combination, the programming language includes object-oriented programming languages such as Smalltalk, C++, etc., and conventional procedural programming languages such as "C" language or similar programming languages.
- Computer readable program instructions can be executed entirely on the user's computer, partly on the user's computer, executed as a stand-alone software package, partly on the user's computer and partly executed on a remote computer, or entirely on the remote computer or server carried out.
- the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, using an Internet service provider to access the Internet connection).
- LAN local area network
- WAN wide area network
- an electronic circuit such as a programmable logic circuit, a field programmable gate array (FPGA), or a programmable logic array (PLA), can be customized by using the status information of the computer-readable program instructions.
- the computer-readable program instructions are executed to realize various aspects of the present disclosure.
- These computer-readable program instructions can be provided to the processors of general-purpose computers, special-purpose computers, or other programmable data processing devices, thereby producing a machine that makes these instructions when executed by the processors of the computer or other programmable data processing devices , A device that implements the functions/actions specified in one or more blocks in the flowcharts and/or block diagrams is produced. It is also possible to store these computer-readable program instructions in a computer-readable storage medium. These instructions make computers, programmable data processing apparatuses, and/or other devices work in a specific manner. Thus, the computer-readable medium storing the instructions includes An article of manufacture, which includes instructions for implementing various aspects of the functions/actions specified in one or more blocks in the flowchart and/or block diagram.
- each block in the flowchart or block diagram may represent a module, program segment, or part of an instruction, and the module, program segment, or part of an instruction contains one or more components for realizing the specified logical function.
- Executable instructions may also occur in a different order from the order marked in the drawings. For example, two consecutive blocks can actually be executed substantially in parallel, or they can sometimes be executed in the reverse order, depending on the functions involved.
- each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart can be implemented by a dedicated hardware-based system that performs the specified functions or actions Or it can be realized by a combination of dedicated hardware and computer instructions.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (21)
- 一种面部识别方法,包括:对包括目标对象的待处理图像进行属性特征提取,得到所述目标对象的N个属性特征,N为大于1的整数;基于注意力机制对所述待处理图像进行注意力特征提取,得到所述目标对象的N个注意力特征;对所述N个注意力特征进行聚类处理,得到M个聚类注意力特征,M为正整数且M<N;根据所述N个属性特征以及所述M个聚类注意力特征,确定所述目标对象的面部识别结果。
- 根据权利要求1所述的方法,其中,对所述N个注意力特征进行聚类处理,得到M个聚类注意力特征,包括:对所述N个注意力特征进行聚类处理,得到M个聚类集合,每个注意力特征与M个聚类集合中的一个聚类集合相对应;分别确定各个聚类集合的聚类注意力特征,得到M个聚类注意力特征。
- 根据权利要求1或2所述的方法,其中,所述方法还包括:将所述N个属性特征与所述N个注意力特征分别相乘,得到增强后的N个属性特征,其中,根据所述N个属性特征以及所述M个聚类注意力特征,确定所述目标对象的面部识别结果,包括:根据所述M个聚类注意力特征对增强后的N个属性特征分别进行修正,得到所述目标对象的面部识别结果。
- 根据权利要求1或2所述的方法,其中,根据所述N个属性特征以及所述M个聚类注意力特征,确定所述目标对象的面部识别结果,包括:根据所述M个聚类注意力特征对所述N个属性特征分别进行修正,得到所述面部识别结果。
- 根据权利要求3所述的方法,其中,根据所述M个聚类注意力特征对增强后的N个属性特征分别进行修正,得到所述目标对象的面部识别结果, 包括:将增强后的N个属性特征与所述N个属性特征中至少部分属性特征所对应的所述聚类注意力特征分别相乘,得到所述面部识别结果。
- 根据权利要求4所述的方法,其中,根据所述M个聚类注意力特征对所述N个属性特征分别进行修正,得到所述目标对象的面部识别结果,包括:将所述N个属性特征与所述N个属性特征中至少部分属性特征所对应的所述聚类注意力特征分别相乘,得到所述面部识别结果。
- 根据权利要求1至6中任一项所述的方法,其中,所述方法通过神经网络实现,所述神经网络包括多任务卷积网络、多个单独注意力网络以及聚类网络,所述多任务卷积网络用于对所述待处理图像进行属性特征提取,所述多个单独注意力网络用于对所述待处理图像进行注意力特征提取,所述聚类网络用于对所述N个注意力特征进行聚类处理。
- 根据权利要求7所述的方法,其中,所述方法还包括:在训练所述神经网络的过程中,根据所述聚类网络的网络损失,调整多个单独注意力网络的网络参数。
- 根据权利要求1-8中任意一项所述的方法,其中,所述聚类处理包括谱聚类,所述M个聚类注意力特征分别是所述M个聚类集合的类中心。
- 一种面部识别装置,其中,包括:属性提取模块,用于对包括目标对象的待处理图像进行属性特征提取,得到所述目标对象的N个属性特征,N为大于1的整数;注意力提取模块,用于基于注意力机制对所述待处理图像进行注意力特征提取,得到所述目标对象的N个注意力特征;聚类模块,用于对所述N个注意力特征进行聚类处理,得到M个聚类注意力特征,M为正整数且M<N;结果确定模块,用于根据所述N个属性特征以及所述M个聚类注意力特征,确定所述目标对象的面部识别结果。
- 根据权利要求10所述的装置,其中,所述聚类模块包括:聚类子模块,用于对所述N个注意力特征进行聚类处理,得到M个聚类集合,每个注意力特征与M个聚类集合中的一个聚类集合相对应;特征确定子模块,用于分别确定各个聚类集合的聚类注意力特征,得到M个聚类注意力特征。
- 根据权利要求10或11所述的装置,其中,所述装置还包括:属性增强模块,用于将所述N个属性特征与所述N个注意力特征分别相乘,得到增强后的N个属性特征,其中,所述结果确定模块包括:第一修正子模块,用于根据所述M个聚类注意力特征对增强后的N个属性特征分别进行修正,得到所述目标对象的面部识别结果。
- 根据权利要求10或11所述的装置,其中,所述结果确定模块包括:第二修正子模块,用于根据所述M个聚类注意力特征对所述N个属性特征分别进行修正,得到所述面部识别结果。
- 根据权利要求12所述的装置,其中,所述第一修正子模块包括:第一相乘子模块,用于将增强后的N个属性特征与各所述属性特征所对应的所述聚类注意力特征分别相乘,得到所述面部识别结果。
- 根据权利要求13所述的装置,其中,所述第二修正子模块包括:第二相乘子模块,用于将所述N个属性特征与各所述属性特征所对应的所述聚类注意力特征分别相乘,得到所述面部识别结果。
- 根据权利要求10至15中任一项所述的装置,其中,所述装置通过神经网络实现,所述神经网络包括多任务卷积网络、多个单独注意力网络以及聚类网络,所述多任务卷积网络用于对所述待处理图像进行属性特征提取,所述多个单独注意力网络用于对所述待处理图像进行注意力特征提取,所述聚类网络用于对所述N个注意力特征进行聚类处理。
- 根据权利要求16所述的装置,其中,所述装置还包括:参数调整模块,用于在训练所述神经网络的过程中,根据所述聚类网络的网络损失,调整多个单独注意力网络的网络参数。
- 根据权利要求10-17中任意一项所述的装置,其中,所述聚类处理包括谱聚类,所述M个聚类注意力特征分别是所述M个聚类集合的类中心。
- 一种电子设备,其中,包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为:执行权利要求1至9中任意一项所述的方法。
- 一种计算机可读存储介质,其上存储有计算机程序指令,其中,所述计算机程序指令被处理器执行时实现权利要求1至9中任意一项所述的方法。
- 一种计算机程序产品,所述计算机程序产品被处理器执行后,实现权利要求1至9中任意一项所述的方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020207018821A KR102390580B1 (ko) | 2019-02-02 | 2019-08-15 | 얼굴 인식 방법 및 장치, 전자 기기 및 저장 매체 |
SG11202006192YA SG11202006192YA (en) | 2019-02-02 | 2019-08-15 | Face recognition method and apparatus, electronic device, and storage medium |
JP2020533112A JP7038829B2 (ja) | 2019-02-02 | 2019-08-15 | 顔部認識方法及び装置、電子機器並びに記憶媒体 |
US16/907,406 US11455830B2 (en) | 2019-02-02 | 2020-06-22 | Face recognition method and apparatus, electronic device, and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910107458.X | 2019-02-02 | ||
CN201910107458.XA CN109800737B (zh) | 2019-02-02 | 2019-02-02 | 面部识别方法及装置、电子设备和存储介质 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/907,406 Continuation US11455830B2 (en) | 2019-02-02 | 2020-06-22 | Face recognition method and apparatus, electronic device, and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020155606A1 true WO2020155606A1 (zh) | 2020-08-06 |
Family
ID=66561887
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/100859 WO2020155606A1 (zh) | 2019-02-02 | 2019-08-15 | 面部识别方法及装置、电子设备和存储介质 |
Country Status (7)
Country | Link |
---|---|
US (1) | US11455830B2 (zh) |
JP (1) | JP7038829B2 (zh) |
KR (1) | KR102390580B1 (zh) |
CN (1) | CN109800737B (zh) |
SG (1) | SG11202006192YA (zh) |
TW (1) | TWI717923B (zh) |
WO (1) | WO2020155606A1 (zh) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112580694A (zh) * | 2020-12-01 | 2021-03-30 | 中国船舶重工集团公司第七0九研究所 | 基于联合注意力机制的小样本图像目标识别方法及系统 |
CN112733652A (zh) * | 2020-12-31 | 2021-04-30 | 深圳赛安特技术服务有限公司 | 图像目标识别方法、装置、计算机设备及可读存储介质 |
CN113049084A (zh) * | 2021-03-16 | 2021-06-29 | 电子科技大学 | 一种基于注意力机制的Resnet的分布式光纤传感信号识别方法 |
CN113052175A (zh) * | 2021-03-26 | 2021-06-29 | 北京百度网讯科技有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN113392263A (zh) * | 2021-06-24 | 2021-09-14 | 上海商汤科技开发有限公司 | 一种数据标注方法及装置、电子设备和存储介质 |
CN113470827A (zh) * | 2021-06-30 | 2021-10-01 | 上海商汤智能科技有限公司 | 分类方法及装置、电子设备和存储介质 |
CN113553905A (zh) * | 2021-06-16 | 2021-10-26 | 北京百度网讯科技有限公司 | 图像识别方法、装置及系统 |
CN114495294A (zh) * | 2021-12-03 | 2022-05-13 | 华中科技大学鄂州工业技术研究院 | 一种地铁闸机无感支付方法、装置及存储介质 |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109800737B (zh) | 2019-02-02 | 2021-06-25 | 深圳市商汤科技有限公司 | 面部识别方法及装置、电子设备和存储介质 |
CN110210572B (zh) * | 2019-06-10 | 2023-02-07 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、存储介质及设备 |
CN110458052B (zh) * | 2019-07-25 | 2023-04-07 | Oppo广东移动通信有限公司 | 基于增强现实的目标对象识别方法、装置、设备、介质 |
CN111339813B (zh) * | 2019-09-30 | 2022-09-27 | 深圳市商汤科技有限公司 | 人脸属性识别方法、装置、电子设备和存储介质 |
CN111274388B (zh) * | 2020-01-14 | 2024-05-10 | 平安科技(深圳)有限公司 | 一种文本聚类的方法及装置 |
CN111950643B (zh) * | 2020-08-18 | 2022-06-28 | 创新奇智(上海)科技有限公司 | 图像分类模型训练方法、图像分类方法及对应装置 |
CN112560756A (zh) * | 2020-12-24 | 2021-03-26 | 北京嘀嘀无限科技发展有限公司 | 识别人脸的方法、装置、电子设备和存储介质 |
CN113177533B (zh) * | 2021-05-28 | 2022-09-06 | 济南博观智能科技有限公司 | 一种人脸识别方法、装置及电子设备 |
CN113688729B (zh) * | 2021-08-24 | 2023-04-07 | 上海商汤科技开发有限公司 | 一种行为识别方法及装置、电子设备和存储介质 |
KR102664254B1 (ko) * | 2022-04-01 | 2024-05-08 | 한국생산기술연구원 | 비전 인공지능 기반 수신호 인식 장치 및 이를 이용한 수신호 인식 방법 |
KR102458179B1 (ko) | 2022-06-20 | 2022-10-24 | 주식회사 한국공간데이터 | 얼굴 속성을 이용한 얼굴 인식 방법 및 얼굴 인식 장치 |
CN117437684B (zh) * | 2023-12-14 | 2024-04-16 | 深圳须弥云图空间科技有限公司 | 一种基于修正注意力的图像识别方法和装置 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120313964A1 (en) * | 2011-06-13 | 2012-12-13 | Sony Corporation | Information processing apparatus, information processing method, and program |
CN105678304A (zh) * | 2015-12-30 | 2016-06-15 | 浙江宇视科技有限公司 | 一种车标识别方法及装置 |
CN106611421A (zh) * | 2016-12-30 | 2017-05-03 | 西安电子科技大学 | 基于特征学习和素描线段约束的sar图像分割方法 |
CN106845421A (zh) * | 2017-01-22 | 2017-06-13 | 北京飞搜科技有限公司 | 基于多区域特征与度量学习的人脸特征识别方法及系统 |
CN108229674A (zh) * | 2017-02-21 | 2018-06-29 | 北京市商汤科技开发有限公司 | 聚类用神经网络的训练方法和装置、聚类方法和装置 |
CN108960331A (zh) * | 2018-07-10 | 2018-12-07 | 重庆邮电大学 | 一种基于行人图像特征聚类的行人再识别方法 |
CN109800737A (zh) * | 2019-02-02 | 2019-05-24 | 深圳市商汤科技有限公司 | 面部识别方法及装置、电子设备和存储介质 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003346151A (ja) | 2002-05-29 | 2003-12-05 | Fuji Photo Film Co Ltd | 画像認識方法および装置並びにプログラム |
TWI528331B (zh) * | 2014-10-06 | 2016-04-01 | 由田新技股份有限公司 | 注意力偵測裝置、方法、電腦可讀取紀錄媒體、及其電腦程式產品 |
CN105096420A (zh) * | 2015-07-31 | 2015-11-25 | 北京旷视科技有限公司 | 门禁系统以及用于其的数据处理方法 |
JP6873600B2 (ja) | 2016-03-04 | 2021-05-19 | キヤノン株式会社 | 画像認識装置、画像認識方法及びプログラム |
US10831827B2 (en) * | 2016-04-01 | 2020-11-10 | International Business Machines Corporation | Automatic extraction of user mobility behaviors and interaction preferences using spatio-temporal data |
CN107909059A (zh) * | 2017-11-30 | 2018-04-13 | 中南大学 | 一种面向复杂城市场景中协同仿生视觉的交通标志牌检测与识别方法 |
CN108875521A (zh) * | 2017-12-20 | 2018-11-23 | 北京旷视科技有限公司 | 人脸检测方法、装置、系统和存储介质 |
CN109255369B (zh) * | 2018-08-09 | 2020-10-16 | 杭州易现先进科技有限公司 | 利用神经网络识别图片的方法及装置、介质和计算设备 |
CN109145867B (zh) * | 2018-09-07 | 2021-08-10 | 北京旷视科技有限公司 | 人体姿态估计方法、装置、系统、电子设备、存储介质 |
US11256960B2 (en) * | 2020-04-15 | 2022-02-22 | Adobe Inc. | Panoptic segmentation |
-
2019
- 2019-02-02 CN CN201910107458.XA patent/CN109800737B/zh active Active
- 2019-08-15 JP JP2020533112A patent/JP7038829B2/ja active Active
- 2019-08-15 KR KR1020207018821A patent/KR102390580B1/ko active IP Right Grant
- 2019-08-15 SG SG11202006192YA patent/SG11202006192YA/en unknown
- 2019-08-15 WO PCT/CN2019/100859 patent/WO2020155606A1/zh active Application Filing
- 2019-11-29 TW TW108143797A patent/TWI717923B/zh active
-
2020
- 2020-06-22 US US16/907,406 patent/US11455830B2/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120313964A1 (en) * | 2011-06-13 | 2012-12-13 | Sony Corporation | Information processing apparatus, information processing method, and program |
CN105678304A (zh) * | 2015-12-30 | 2016-06-15 | 浙江宇视科技有限公司 | 一种车标识别方法及装置 |
CN106611421A (zh) * | 2016-12-30 | 2017-05-03 | 西安电子科技大学 | 基于特征学习和素描线段约束的sar图像分割方法 |
CN106845421A (zh) * | 2017-01-22 | 2017-06-13 | 北京飞搜科技有限公司 | 基于多区域特征与度量学习的人脸特征识别方法及系统 |
CN108229674A (zh) * | 2017-02-21 | 2018-06-29 | 北京市商汤科技开发有限公司 | 聚类用神经网络的训练方法和装置、聚类方法和装置 |
CN108960331A (zh) * | 2018-07-10 | 2018-12-07 | 重庆邮电大学 | 一种基于行人图像特征聚类的行人再识别方法 |
CN109800737A (zh) * | 2019-02-02 | 2019-05-24 | 深圳市商汤科技有限公司 | 面部识别方法及装置、电子设备和存储介质 |
Non-Patent Citations (1)
Title |
---|
LUO, JIANHAO ET AL: "A Survey on Fine-Grained Image Categorization Using Deep Convolutional Features", ACTA AUTOMATICA SINICA, vol. 43, no. 8, 31 August 2017 (2017-08-31), pages 1306 - 1318 * |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112580694B (zh) * | 2020-12-01 | 2024-04-19 | 中国船舶重工集团公司第七0九研究所 | 基于联合注意力机制的小样本图像目标识别方法及系统 |
CN112580694A (zh) * | 2020-12-01 | 2021-03-30 | 中国船舶重工集团公司第七0九研究所 | 基于联合注意力机制的小样本图像目标识别方法及系统 |
CN112733652A (zh) * | 2020-12-31 | 2021-04-30 | 深圳赛安特技术服务有限公司 | 图像目标识别方法、装置、计算机设备及可读存储介质 |
CN112733652B (zh) * | 2020-12-31 | 2024-04-19 | 深圳赛安特技术服务有限公司 | 图像目标识别方法、装置、计算机设备及可读存储介质 |
CN113049084B (zh) * | 2021-03-16 | 2022-05-06 | 电子科技大学 | 一种基于注意力机制的Resnet的分布式光纤传感信号识别方法 |
CN113049084A (zh) * | 2021-03-16 | 2021-06-29 | 电子科技大学 | 一种基于注意力机制的Resnet的分布式光纤传感信号识别方法 |
CN113052175B (zh) * | 2021-03-26 | 2024-03-29 | 北京百度网讯科技有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN113052175A (zh) * | 2021-03-26 | 2021-06-29 | 北京百度网讯科技有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN113553905A (zh) * | 2021-06-16 | 2021-10-26 | 北京百度网讯科技有限公司 | 图像识别方法、装置及系统 |
CN113553905B (zh) * | 2021-06-16 | 2024-04-26 | 北京百度网讯科技有限公司 | 图像识别方法、装置及系统 |
CN113392263A (zh) * | 2021-06-24 | 2021-09-14 | 上海商汤科技开发有限公司 | 一种数据标注方法及装置、电子设备和存储介质 |
CN113470827A (zh) * | 2021-06-30 | 2021-10-01 | 上海商汤智能科技有限公司 | 分类方法及装置、电子设备和存储介质 |
CN114495294A (zh) * | 2021-12-03 | 2022-05-13 | 华中科技大学鄂州工业技术研究院 | 一种地铁闸机无感支付方法、装置及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
SG11202006192YA (en) | 2020-11-27 |
CN109800737A (zh) | 2019-05-24 |
KR20200096565A (ko) | 2020-08-12 |
JP7038829B2 (ja) | 2022-03-18 |
US20200320281A1 (en) | 2020-10-08 |
CN109800737B (zh) | 2021-06-25 |
US11455830B2 (en) | 2022-09-27 |
TWI717923B (zh) | 2021-02-01 |
KR102390580B1 (ko) | 2022-04-26 |
JP2021514497A (ja) | 2021-06-10 |
TW202030638A (zh) | 2020-08-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
TWI717923B (zh) | 面部識別方法及裝置、電子設備和儲存介質 | |
US20210326708A1 (en) | Neural network training method and apparatus, and image processing method and apparatus | |
WO2021196401A1 (zh) | 图像重建方法及装置、电子设备和存储介质 | |
WO2021155632A1 (zh) | 图像处理方法及装置、电子设备和存储介质 | |
WO2021056808A1 (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN110909815B (zh) | 神经网络训练、图像处理方法、装置及电子设备 | |
CN110598504B (zh) | 图像识别方法及装置、电子设备和存储介质 | |
TW202105244A (zh) | 圖像處理方法及裝置、電子設備和電腦可讀儲存介質 | |
JP2022523606A (ja) | 動画解析のためのゲーティングモデル | |
JP7165818B2 (ja) | ニューラルネットワークのトレーニング方法及び装置並びに画像生成方法及び装置 | |
TW202105202A (zh) | 影片處理方法及裝置、電子設備、儲存媒體和電腦程式 | |
CN110458218B (zh) | 图像分类方法及装置、分类网络训练方法及装置 | |
CN105335684B (zh) | 人脸检测方法及装置 | |
TWI735112B (zh) | 圖像生成方法、電子設備和儲存介質 | |
US20210012154A1 (en) | Network optimization method and apparatus, image processing method and apparatus, and storage medium | |
CN109165738B (zh) | 神经网络模型的优化方法及装置、电子设备和存储介质 | |
CN111259967B (zh) | 图像分类及神经网络训练方法、装置、设备及存储介质 | |
TW202032425A (zh) | 圖像處理方法及裝置、電子設備和儲存介質 | |
CN111242303B (zh) | 网络训练方法及装置、图像处理方法及装置 | |
WO2020114236A1 (zh) | 关键点检测方法、装置、电子设备及存储介质 | |
WO2020192113A1 (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN110135349A (zh) | 识别方法、装置、设备及存储介质 | |
WO2023024439A1 (zh) | 一种行为识别方法及装置、电子设备和存储介质 | |
CN109447258B (zh) | 神经网络模型的优化方法及装置、电子设备和存储介质 | |
WO2021056770A1 (zh) | 图像重建方法及装置、电子设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2020533112 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20207018821 Country of ref document: KR Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19912709 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 23/11/2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19912709 Country of ref document: EP Kind code of ref document: A1 |