CN113239844A - Intelligent cosmetic mirror system based on multi-head attention target detection - Google Patents

Intelligent cosmetic mirror system based on multi-head attention target detection Download PDF

Info

Publication number
CN113239844A
CN113239844A CN202110576729.3A CN202110576729A CN113239844A CN 113239844 A CN113239844 A CN 113239844A CN 202110576729 A CN202110576729 A CN 202110576729A CN 113239844 A CN113239844 A CN 113239844A
Authority
CN
China
Prior art keywords
target detection
control unit
makeup
head attention
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110576729.3A
Other languages
Chinese (zh)
Other versions
CN113239844B (en
Inventor
刘斌毓
张丽平
夏劲松
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin University of Science and Technology
Original Assignee
Harbin University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin University of Science and Technology filed Critical Harbin University of Science and Technology
Priority to CN202110576729.3A priority Critical patent/CN113239844B/en
Publication of CN113239844A publication Critical patent/CN113239844A/en
Application granted granted Critical
Publication of CN113239844B publication Critical patent/CN113239844B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • AHUMAN NECESSITIES
    • A45HAND OR TRAVELLING ARTICLES
    • A45DHAIRDRESSING OR SHAVING EQUIPMENT; EQUIPMENT FOR COSMETICS OR COSMETIC TREATMENTS, e.g. FOR MANICURING OR PEDICURING
    • A45D42/00Hand, pocket, or shaving mirrors
    • A45D42/08Shaving mirrors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0631Item recommendations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0641Shopping interfaces
    • G06Q30/0643Graphical representation of items or shoppers

Abstract

The invention discloses an intelligent cosmetic mirror system based on multi-head attention target detection, which comprises an image acquisition unit, a target detection unit and a control unit, wherein the image acquisition unit is used for acquiring an image of a user; the image acquisition unit is used for acquiring a facial image of a user; the target detection unit extracts the features of the user image, detects different face areas, evaluates the makeup of the different face areas and makes up in a high-definition mode; the control unit combines the information fed back by the target detection unit to generate a current makeup preview image on the cosmetic mirror; on the basis of the traditional cosmetic mirror, target detection and a multi-head attention mechanism are added, an artificial intelligent middle-depth learning technology such as an antagonistic network is generated, different parts of a human face are identified by using the target detection technology, the makeup information of different parts is compared with a cloud database, the current makeup is scored, high-definition makeup of specific parts is performed, and the final makeup effect is predicted, so that a user can visually see the makeup trying effect of the user, and the time of the user is saved.

Description

Intelligent cosmetic mirror system based on multi-head attention target detection
Technical Field
The invention relates to the technical field of cosmetic mirrors, in particular to an intelligent cosmetic mirror system based on multi-head attention target detection.
Background
The cosmetic mirror is a necessity of daily life of a female group, and the traditional cosmetic mirror has a single function and only has a basic function of imaging or light supplement. When a user only tries some makeup, the user must completely finish the makeup on the face and then wipe the makeup, which easily causes great damage to the skin of the user, and wastes not only a large amount of cosmetics but also time of the user. In addition, when the user is making up, the user does not know whether the current makeup effect is expected or not, and the user does not know which makeup is more suitable for the current makeup. Therefore, how to improve the use efficiency of the cosmetic mirror is a problem to be solved urgently.
The intelligent cosmetic mirror system based on the multi-head attention target detection is designed by utilizing the technology of generating an confrontation network based on the multi-head attention target detection, the problems can be solved, and the service efficiency of the cosmetic mirror is effectively improved.
Disclosure of Invention
The invention provides an intelligent cosmetic mirror system based on multi-head attention target detection, which aims to overcome the defects of the prior art. In order to achieve the purpose, the invention adopts the following specific technical scheme:
the method for extracting the face features of the face image by using the residual error network comprises the following steps:
(1) extracting the feature information of the face image through the convolution layer, wherein the feature extraction formula is as follows:
Figure BDA0003084594810000011
Yp=f(zp)
wherein Wp∈Rc×w×hIn the form of a three-dimensional convolution kernel,
Figure BDA0003084594810000012
representing a convolution operation, X being the input face image data, bpFor convolution biasing, f (-) is the Relu activation function, YpAre human face features. Wherein the Relu activation function calculation formula is as follows:
Figure BDA0003084594810000013
(2) residual error connection is carried out on the face features output by the sub-networks in different residual error modules, and the specific calculation formula is as follows:
Figure BDA0003084594810000014
(3) inputting the face feature map finally output by the residual error network into a multi-head attention coding neural network, wherein a multi-head attention calculation formula is as follows:
Figure BDA0003084594810000021
wherein KiIs the ith key, ViIs the ith value, qiFor the ith query, Wk,Wv,Wq,vTA projection matrix of keys, values, queries, activation vectors, tanh (-), softmax (-), is an activation function.
(4) Defining a face region target detection loss function as:
Figure BDA0003084594810000022
and training the target detection network based on the multi-head attention mechanism.
(5) The makeup information of different parts of the face image is transmitted to a cloud server through a control unit wireless communication module by utilizing a face region target prediction frame generated by a target detection unit, the cloud server compares the makeup information with relevant makeup information in a cloud database to give a score of the makeup of the current user, and the control unit feeds the score back to the user through a makeup mirror after receiving the score. And for the too low makeup score, the server recommends a corresponding cosmetic to the user according to the current face area.
(6) And inputting the face features obtained by the residual error network into a generated confrontation network to obtain a final predicted effect image of the current makeup of the user.
The invention has the beneficial effects that: compared with other target detection methods, the method disclosed by the invention has the advantages that the target frame is predicted by directly utilizing the characteristic information of the image, the design process of artificially carrying out an anchor frame in the traditional target detection method is omitted, and the end-to-end target detection is really realized. The target detection technology can help the user to know the current makeup information more quickly, save a large amount of time of the user and improve the use experience of the user.
Drawings
FIG. 1 is a schematic diagram of a target detection unit human face feature extraction residual error network used in the present invention.
FIG. 2 is a schematic diagram of an object detection network based on a multi-head attention mechanism of an object detection unit used in the present invention.
Fig. 3 is a schematic diagram of a generator in a target detection unit generation countermeasure network used in the present invention.
Fig. 4 is a schematic diagram of an arbiter in a target detection unit generation countermeasure network used in the present invention.
Description of reference numerals: 1-face image data; 2-residual module 1; 3-residual module 2; 4-residual module 3; 5-residual module 4; 6-sequence embedding layer; 7-multi-head attention target detection; 8-target prediction box; 9-residual module 5; 10-residual module 6; 11-residual module 7; 12-residual module 8; 13-average pooling layer 1; 14-residual module 9; 15-residual module 10; 16-residual module 11; 17-residual block 12; average pooling layer 2.
Detailed Description
The invention will be further elucidated and described with reference to the drawings and to the detailed description of an embodiment of the invention. It is to be understood that the described embodiments are merely exemplary of some, and not all, embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
FIG. 1 is a schematic diagram of a target detection unit human face feature extraction residual error network used in the present invention.
As shown in fig. 1, the method for extracting features of a face image by using a residual error network comprises the following steps:
(1) 1, inputting a plurality of images obtained by an image acquisition unit into a residual error network as a batch;
(2) firstly, extracting features of the image through a 2 first residual error module, wherein residual errors are connected among residual error sub-networks, and finally, the first residual error module outputs a 56 x 256 face feature map through a plurality of convolution kernels of 1 x 1 and 3 x 3;
(3) inputting the 56 × 256 face feature maps into a 3 second residual module, wherein residual sub-networks are connected with each other through residual errors, and finally the second residual module outputs 28 × 512 face feature maps through a plurality of convolution kernels of 1 × 1 and 3 × 3;
(4) inputting the 28 × 512 face feature map into 4 third residual modules, wherein residual sub-networks are connected with each other through residual errors, and finally the third residual module outputs 14 × 1024 face feature map through a plurality of convolution kernels of 1 × 1 and 3 × 3;
(5) and (3) inputting the 14 × 1024 face feature map into a 5 th residual module, wherein residual sub-networks are connected with one another through residual errors, and finally the fourth residual module outputs a 7 × 2048 face feature map through a plurality of convolution kernels of 1 × 1 and 3 × 3.
FIG. 2 is a schematic diagram of an object detection network based on a multi-head attention mechanism of an object detection unit used in the present invention.
As shown in fig. 2, the method for identifying different regions of a human face by using a multi-head attention target detection network comprises the following steps:
(1) carrying out Hadamard product on coding information of different areas of the face in the face characteristic image and the face image, inputting the Hadamard product result into a 6-sequence embedding layer, inputting the result into a 7-multi-head attention coding neural network, and carrying out different subspace projection on the Hadamard product by using keys, values and a projection matrix of query respectively;
(2) calculating attention scores of different heads by using a zooming dot product;
(3) normalizing the attention score by a softmax (·) activation function;
(4) respectively normalizing and regularizing the attention scores of different heads, and inputting a final sequence result 8 into a feedforward neural network;
(5) predicting frames of different parts of the face through different full-connection layers, inputting final information of the predicting frames into a control unit, marking the different parts of the face on a cosmetic mirror according to the different predicting frames by the control unit, capturing makeup information of the different parts, and transmitting the makeup information to a cloud server for makeup effect evaluation.
Fig. 3 is a schematic diagram of a generator in a target detection unit generation countermeasure network used in the present invention.
Fig. 4 is a schematic diagram of an arbiter in a target detection unit generation countermeasure network used in the present invention.
As shown in fig. 3 and 4, the method for predicting the final makeup effect of the face by using the generator and the discriminator for generating the confrontation network comprises the following steps:
(1) inputting the face feature map finally output by the residual error network into a generation countermeasure network generator, wherein the generator generates a predicted image of the final effect of the current makeup of the user through a plurality of different convolution layers of 9, 10, 11, 12-3 x 3 and a plurality of different pooling layers of 13-2 x 2;
(2) inputting the predicted image of the final effect of the current makeup of the user generated by the generator into an arbiter, the arbiter utilizes a plurality of 14, 15, 16, 17-3 x 3 convolution layers and 18 pooling layers to extract the characteristics of the image, the characteristics are input into a control unit, the control unit is combined with a cloud database to inquire a makeup effect image which best accords with the characteristics of the current makeup, the image is fed back to the arbiter, the arbiter utilizes the inner product by pixel to calculate the similarity between the predicted image of the final effect of the current makeup of the user generated by the generator and the makeup effect image fed back to the control unit, the confidence coefficient of the predicted image of the final effect of the current makeup of the user generated by the generator is obtained through cross entropy loss, if the confidence coefficient is greater than a preset threshold value of 0.8, the predicted image of the final effect of the current makeup of the user generated by the generator is input into the control unit, and the control unit is drawn on a cosmetic mirror, and if the current makeup final effect prediction graph generated by the generator is smaller than the threshold, inputting the current makeup final effect prediction graph generated by the generator into the generator, and enabling the generator to regenerate the current makeup final effect prediction graph until the current makeup final effect prediction graph is larger than the preset threshold.
Examples
To verify the effectiveness of multi-head attention target detection, a comparison was made on the well-known target detection data set COCO with the current best target detection method, fast R-CNN network. Wherein the Faster RCNN-FPN is a fast R-CNN network with regional suggestions, and the Faster RCNN-R101-FPN is a master R-CNN network with ResNet101 as a backbone network. The results are shown in Table 1, where AP represents the mean accuracy, AP-50 represents the AP measurement at a threshold of IoU of 0.5, AP-75 represents the measurement at a threshold of IoU of 0.75, AP-S represents the AP measurement of the target box at pixel areas less than 32 x 32, AP-M represents the measurement of the target box at pixel areas between 32 x 32 and 96 x 96, and AP-L represents the AP measurement of the target box at pixel areas greater than 96 x 96.
Figure BDA0003084594810000051
Table 1 test results of the validity experiment of the present invention for multi-head attention target detection.

Claims (8)

1. An intelligent cosmetic mirror system based on multi-head attention target detection is characterized by comprising: the system comprises an image acquisition unit, a target detection unit and a control unit; the image acquisition unit acquires a face image of a user under the control of the control unit through an external camera; the target detection unit utilizes a residual error network to extract human face features for the collected human face images under the control of the control unit, and the extracted human face features are respectively input into a coding neural network based on a multi-head attention mechanism and a generator for generating a confrontation network; inputting the output of the coding neural network into a decoding neural network to obtain a human face area detection frame, and inputting the human face makeup image output by the generator into a discriminator for generating a countermeasure network to obtain a final makeup predicted image; the result of the target detection unit is displayed on the cosmetic mirror and presented to the user under the control of the control unit; the control unit comprises a wireless communication module, a voice input module, a voice output module, a clock module, a storage module, an arithmetic logic operation module and a microprogram conversion module.
2. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the residual error network of the target detection unit comprises 3 residual error modules 1, 4 residual error modules 2, 6 residual error modules 3 and 3 residual error modules 4, and residual error connection is carried out among the residual error modules; wherein the residual module 1 comprises a convolution layer of 1 × 64, a convolution layer of 3 × 64, and a convolution layer of 1 × 256; residual module 2 contains one convolution layer of 1 × 128, one convolution layer of 3 × 128, and one convolution layer of 1 × 512; residual module 3 contains one convolution layer of 1 × 256, one convolution layer of 3 × 256, and one convolution layer of 1 × 1024; residual module 4 contains one convolution layer of 1 × 512, one convolution layer of 3 × 512, and one convolution layer of 1 × 2048; the final residual network outputs a 7 × 2048 feature map.
3. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the coding neural network based on the multi-head attention mechanism of the target detection unit comprises a sequence embedding layer, a multi-head attention layer, a network layer for regularizing the output of the multi-head attention layer, a feedforward neural network layer and a network layer for regularizing the output of the feedforward neural network; the input of the coding neural network based on the multi-head attention mechanism is a 7 × 2048 feature map finally output by a residual error network, the feature map is converted into 49 sequence type data of 1 × 2048 through a sequence embedding layer and is input into a multi-head attention layer, the multi-head attention layer captures the human face region features in the sequence data, the human face region features are input into the feed-forward neural network through a regularization layer, and the final sequence data related to the human face region features are obtained after regularization.
4. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the generation countermeasure network of the target detection unit includes a generator for outputting data of a specified type from random input data, a discriminator for judging output data of the generator based on real data; the input of the generator is a 7 × 2048 feature map finally output by the residual error network, and an effect map after face makeup is predicted is output; the input of the discriminator is the makeup effect picture output by the generator, when the confidence level of the makeup effect picture generated by the generator calculated by the discriminator is larger than the threshold value, the makeup effect picture is output to the control unit, the control unit can display the makeup effect picture on the cosmetic mirror, otherwise, the discriminator can input the makeup effect picture into the generator again and requires the generator to regenerate the makeup effect picture until the confidence level of the generated makeup effect picture is larger than the threshold value.
5. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the wireless communication module that the control unit contained is used for establishing the connection between control unit and high in the clouds server, mobile terminal, mobile network, realizes the data interaction and the control interaction of control unit and corresponding part.
6. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the voice input module contained in the control unit receives the voice control command of the user and completes the execution and the corresponding of the corresponding command.
7. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the voice output module contained in the control unit converts the internal control signal into voice information and outputs the voice information to the loudspeaker to prompt the user of the related information.
8. The intelligent cosmetic mirror system based on multi-head attention target detection according to claim 1, characterized in that: the clock module contained in the control unit generates periodic pulse signals, so that the components in the control unit can execute commands in order; the control unit comprises a storage module for storing data generated in the running of the microprogram, an execution result of a corresponding control command, a result generated by the target detection unit and input data of each network for target detection; the arithmetic logic operation module of the control unit performs corresponding arithmetic operation and logic operation; the microprogram conversion module contained in the control unit is used for converting other programs into microprograms executable by the control unit.
CN202110576729.3A 2021-05-26 2021-05-26 Intelligent cosmetic mirror system based on multi-head attention target detection Active CN113239844B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110576729.3A CN113239844B (en) 2021-05-26 2021-05-26 Intelligent cosmetic mirror system based on multi-head attention target detection

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110576729.3A CN113239844B (en) 2021-05-26 2021-05-26 Intelligent cosmetic mirror system based on multi-head attention target detection

Publications (2)

Publication Number Publication Date
CN113239844A true CN113239844A (en) 2021-08-10
CN113239844B CN113239844B (en) 2022-11-01

Family

ID=77138938

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110576729.3A Active CN113239844B (en) 2021-05-26 2021-05-26 Intelligent cosmetic mirror system based on multi-head attention target detection

Country Status (1)

Country Link
CN (1) CN113239844B (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10381105B1 (en) * 2017-01-24 2019-08-13 Bao Personalized beauty system
CN110807332A (en) * 2019-10-30 2020-02-18 腾讯科技(深圳)有限公司 Training method of semantic understanding model, semantic processing method, semantic processing device and storage medium
US20200134428A1 (en) * 2018-10-29 2020-04-30 Nec Laboratories America, Inc. Self-attentive attributed network embedding
WO2020171550A1 (en) * 2019-02-18 2020-08-27 Samsung Electronics Co., Ltd. Image processing method and apparatus, electronic device and computer readable storage medium
CN111639596A (en) * 2020-05-29 2020-09-08 上海锘科智能科技有限公司 Anti-glasses-shielding face recognition method based on attention mechanism and residual error network
CN111861945A (en) * 2020-09-21 2020-10-30 浙江大学 Text-guided image restoration method and system
CN112017301A (en) * 2020-07-24 2020-12-01 武汉纺织大学 Style migration model and method for specific relevant area of clothing image
CN112084841A (en) * 2020-07-27 2020-12-15 齐鲁工业大学 Cross-modal image multi-style subtitle generation method and system
CN112232156A (en) * 2020-09-30 2021-01-15 河海大学 Remote sensing scene classification method based on multi-head attention generation countermeasure network
WO2021022521A1 (en) * 2019-08-07 2021-02-11 华为技术有限公司 Method for processing data, and method and device for training neural network model

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10381105B1 (en) * 2017-01-24 2019-08-13 Bao Personalized beauty system
US20200321074A1 (en) * 2017-01-24 2020-10-08 Ha Tran Personalized cosmetic system
US20200134428A1 (en) * 2018-10-29 2020-04-30 Nec Laboratories America, Inc. Self-attentive attributed network embedding
WO2020171550A1 (en) * 2019-02-18 2020-08-27 Samsung Electronics Co., Ltd. Image processing method and apparatus, electronic device and computer readable storage medium
WO2021022521A1 (en) * 2019-08-07 2021-02-11 华为技术有限公司 Method for processing data, and method and device for training neural network model
CN110807332A (en) * 2019-10-30 2020-02-18 腾讯科技(深圳)有限公司 Training method of semantic understanding model, semantic processing method, semantic processing device and storage medium
CN111639596A (en) * 2020-05-29 2020-09-08 上海锘科智能科技有限公司 Anti-glasses-shielding face recognition method based on attention mechanism and residual error network
CN112017301A (en) * 2020-07-24 2020-12-01 武汉纺织大学 Style migration model and method for specific relevant area of clothing image
CN112084841A (en) * 2020-07-27 2020-12-15 齐鲁工业大学 Cross-modal image multi-style subtitle generation method and system
CN111861945A (en) * 2020-09-21 2020-10-30 浙江大学 Text-guided image restoration method and system
CN112232156A (en) * 2020-09-30 2021-01-15 河海大学 Remote sensing scene classification method based on multi-head attention generation countermeasure network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
常新旭等: "融合多头自注意力机制的语音增强方法", 《西安电子科技大学学报》 *
石开等: "多头注意力与语义视频标注", 《计算机工程与应用》 *

Also Published As

Publication number Publication date
CN113239844B (en) 2022-11-01

Similar Documents

Publication Publication Date Title
Lee et al. From big to small: Multi-scale local planar guidance for monocular depth estimation
CN110689008A (en) Monocular image-oriented three-dimensional object detection method based on three-dimensional reconstruction
CN110674688B (en) Face recognition model acquisition method, system and medium for video monitoring scene
CN109508663A (en) A kind of pedestrian's recognition methods again based on multi-level supervision network
JP7166388B2 (en) License plate recognition method, license plate recognition model training method and apparatus
US9299011B2 (en) Signal processing apparatus, signal processing method, output apparatus, output method, and program for learning and restoring signals with sparse coefficients
CN114202740A (en) Pedestrian re-identification method based on multi-scale feature fusion
CN111523377A (en) Multi-task human body posture estimation and behavior recognition method
CN112581409A (en) Image defogging method based on end-to-end multiple information distillation network
CN112163508A (en) Character recognition method and system based on real scene and OCR terminal
KR20110063989A (en) Video object detection apparatus and method thereof
CN115328319B (en) Intelligent control method and device based on light-weight gesture recognition
CN104346630A (en) Cloud flower identifying method based on heterogeneous feature fusion
CN112906493A (en) Cross-modal pedestrian re-identification method based on cross-correlation attention mechanism
CN115188066A (en) Moving target detection system and method based on cooperative attention and multi-scale fusion
CN109002776B (en) Face recognition method, system, computer device and computer-readable storage medium
CN113065506B (en) Human body posture recognition method and system
CN112016592B (en) Domain adaptive semantic segmentation method and device based on cross domain category perception
US11810366B1 (en) Joint modeling method and apparatus for enhancing local features of pedestrians
CN113239844A (en) Intelligent cosmetic mirror system based on multi-head attention target detection
WO2022127576A1 (en) Site model updating method and system
He et al. MTRFN: Multiscale temporal receptive field network for compressed video action recognition at edge servers
Zhu et al. HDRD-Net: High-resolution detail-recovering image deraining network
CN114663910A (en) Multi-mode learning state analysis system
CN113837230A (en) Image description generation method based on adaptive attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant