CN112200024B - Two-dimensional facial expression recognition method through three-dimensional deformable model learning - Google Patents

Two-dimensional facial expression recognition method through three-dimensional deformable model learning Download PDF

Info

Publication number
CN112200024B
CN112200024B CN202011018179.5A CN202011018179A CN112200024B CN 112200024 B CN112200024 B CN 112200024B CN 202011018179 A CN202011018179 A CN 202011018179A CN 112200024 B CN112200024 B CN 112200024B
Authority
CN
China
Prior art keywords
dimensional
expression
point cloud
loss
decoder
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011018179.5A
Other languages
Chinese (zh)
Other versions
CN112200024A (en
Inventor
时迎琰
邹乔莎
史传进
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fudan University
Original Assignee
Fudan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fudan University filed Critical Fudan University
Priority to CN202011018179.5A priority Critical patent/CN112200024B/en
Publication of CN112200024A publication Critical patent/CN112200024A/en
Application granted granted Critical
Publication of CN112200024B publication Critical patent/CN112200024B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the technical field of image processing, and particularly relates to a two-dimensional facial expression recognition method based on three-dimensional deformable model learning. According to the method, the three-dimensional expression point cloud corresponding to the two-dimensional facial expression is reconstructed, the network structure of an encoder-decoder is adopted, the expression classifier, expression classification loss and three-dimensional point cloud regression loss are combined, and under end-to-end combined training, the finally obtained expression parameters not only enhance the discriminability and improve the accuracy of two-dimensional facial expression classification, but also keep the generation performance, and the corresponding three-dimensional facial expression point cloud can be recovered. The invention can realize two functions of two-dimensional facial expression classification and three-dimensional facial expression point cloud generation at the same time, and has wide application prospect.

Description

Two-dimensional facial expression recognition method through three-dimensional deformable model learning
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a two-dimensional facial expression recognition method through three-dimensional deformable model learning.
Background
The facial expression recognition technology plays an important role in emotional artificial intelligence, and is widely applied to digital entertainment, driver monitoring, children autism diagnosis and other human-computer interaction systems. Therefore, reliable intelligent facial expression recognition technology is receiving more and more attention from academia and industry.
The three-dimensional deformable model is a parameterized face three-dimensional point cloud model based on statistics, wherein expression parameters can describe different actions of the face. In order to classify two-dimensional facial expressions by using expression parameters of three-dimensional deformable models, expNet [1] Using ResNet 101 [2] And directly regressing the expression parameters of the two-dimensional face picture as a regressor, and then classifying the obtained expression parameters by using a k-nearest neighbor classification algorithm to obtain a final expression category result.
The extraction stage and the classification stage of the facial expression characteristics are separated by the ExpNet, and the classification effect of end-to-end training cannot be achieved. Therefore, expNet still has room to optimize.
In order to solve the problems, the invention provides a two-dimensional facial expression recognition method through three-dimensional deformable model learning. The invention can realize two functions of two-dimensional facial expression classification and three-dimensional facial expression point cloud generation at the same time, and has wide application prospect.
Reference:
[1]Feng-Ju Chang, Anh Tuan Tran, Tal Hassner, Iacopo Masi, Ram Nevatia, and Gérard G. Medioni. 2018. ExpNet: Landmark-Free, Deep, 3D Facial Expressions. In 13th IEEE International Conference on Automatic Face & Gesture Recognition, FG 2018, Xi’an, China, May 15-19, 2018. IEEE Computer Society, 122–129. https://doi.org/10.1109/FG.2018.00027
[2] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of Conference on Computer Vision and Pattern Recognition, June 2016.
[3] Pascal Paysan, Reinhard Knothe, Brian Amberg, Sami Romdhani, and Thomas Vetter. 2009. A 3D Face Model for Pose and Illumination Invariant Face Recognition. In Sixth IEEE International Conference on Advanced Video and Signal Based Surveillance, AVSS 2009, 2-4 September 2009, Genova, Italy, Stefano Tubaro and Jean-Luc Dugelay (Eds.). IEEE Computer Society, 296–301. https://doi.org/10.1109/AVSS.2009.58
[4] Chen Cao, Yanlin Weng, Shun Zhou, Yiying Tong, and Kun Zhou. 2014. Face-Warehouse:A 3D Facial Expression Database for Visual Computing. IEEE Trans. Vis. Comput.Graph. 20, 3 (2014), 413–425. https://doi.org/10.1109/TVCG.2013.249
[5]Weiyang Liu, YandongWen, Zhiding Yu, Ming Li, Bhiksha Raj, and Le Song. 2017. SphereFace: Deep Hypersphere Embedding for Face Recognition. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017.IEEE Computer Society, 6738–6746. https://doi.org/10.1109/CVPR.2017.713
[6] Patrick Lucey, Jeffrey F. Cohn, Takeo Kanade, Jason M. Saragih, Zara Ambadar, and Iain A. Matthews. 2010. The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR Workshops 2010, San Francisco, CA, USA, 13-18 June, 2010. IEEE Computer Society, 94–101. https://doi.org/10.1109/CVPRW.2010.5543262
[7] Guoying Zhao, Xiaohua Huang, Matti Taini, Stan Z. Li, and Matti Pietikäinen. 2011. Facial expression recognition from near-infrared videos. Image Vis. Comput. 29, 9 (2011), 607–619. https://doi.org/10.1016/j.imavis.2011.07.002。
disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a two-dimensional facial expression recognition method based on three-dimensional deformable model learning.
The invention provides a two-dimensional facial expression recognition method based on a three-dimensional deformable model learning technology, which comprises the following steps: reconstructing a three-dimensional facial expression point cloud corresponding to a two-dimensional facial expression, adopting a network structure of an encoder-decoder, combining an expression classifier, expression classification loss and three-dimensional point cloud regression loss, and finally obtaining expression parameters under end-to-end joint training; the method not only enhances the discriminability and improves the accuracy of two-dimensional facial expression classification, but also keeps the generativeness and can recover the corresponding three-dimensional facial expression point cloud;
the two-dimensional facial expression is a facial image to be subjected to expression classification;
the three-dimensional facial expression point cloud is a part of the three-dimensional space after the calm face point cloud is removed from the face point cloud, namely the point cloud corresponding to the facial action;
the expression parameters are codes of facial expressions, and corresponding three-dimensional facial expression point clouds can be recovered through a decoder;
the encoder is used for encoding the input two-dimensional facial expression and outputting expression parameters;
the decoder is used for decoding the input expression parameters and outputting three-dimensional facial expression point clouds;
the expression classifier adopts a SoftMax classification layer and outputs the prediction probability of each expression;
the expression classification loss is applied to the expression parameters through an expression classifier by adopting cross entropy loss;
the three-dimensional point cloud regression loss is applied to the output of a decoder by adopting MSE mean square error loss.
The method provided by the invention adopts end-to-end joint training; the concrete mode is as follows:
the method comprises the following steps of firstly, a pre-training stage: respectively training an encoder by using the two-dimensional face picture and the expression parameters of the corresponding three-dimensional deformable model, and training a decoder by using the expression parameters of the three-dimensional deformable model and the corresponding three-dimensional face expression point cloud; thus, the encoder and the decoder have preliminary functions;
then a joint training phase: performing joint training on the pre-trained encoder and the pre-trained decoder, namely inputting expression parameters output by the encoder into the decoder, and generating three-dimensional facial expression point clouds corresponding to the facial pictures end to end; in the combined training process, expression classification loss is applied to expression parameters output by the encoder through the expression classifier, so that the distribution of the expression parameters in the space is more discriminative, the inter-class distance is increased, and the accuracy of expression classification is improved; and meanwhile, the mean square error loss is applied to the three-dimensional facial expression point cloud output by the decoder, so that the finally obtained expression parameters can still generate the corresponding three-dimensional facial expression point cloud.
The invention has the technical effects that the expression type prediction of the input face image can be realized through the encoder and the expression classifier, and the three-dimensional face expression point cloud can be generated through the encoder and the decoder. The two functions are realized simultaneously, and the method has wide application prospect.
Drawings
Fig. 1 is a simplified schematic diagram of the framework of the present invention.
Detailed Description
The present invention will be described more fully hereinafter in the reference to the accompanying drawings, which provide preferred embodiments of the invention, and which are not to be considered as limited to the embodiments set forth herein.
Fig. 1 shows a simplified schematic diagram of a two-dimensional facial expression recognition method through three-dimensional deformable model learning according to the present invention. The three-dimensional deformable Model selects the base Face Model [3] And faceWarehouse [4] Then the expression parameters are 29-dimensional vectors, the three-dimensional point cloud has 53215 vertexes, and the encoder selects SphereFace [5] Adding a 29-dimensional full-connection layer at the top end of the decoder for outputting 29-dimensional expression parameters, selecting a multi-layer perceptron with two layers, the dimension of the middle layer is 1000, and outputting the layerIs 159645; the expression classifier adopts a SoftMax classification layer (SoftMax is a common function); the expression classification loss is applied to the expression parameters through an expression classifier by adopting cross entropy loss; the three-dimensional point cloud regression loss is applied to the output of the decoder by adopting MSE mean square error loss.
In particular, assume thatIIs the input two-dimensional face picture,Ein order to be an encoder, the encoder is,Din order to be a decoder, the decoder,p=[p 1 , p 2 , …, p 29 ]is an expression parameter of 29 dimensions and is,S = [x 1 , y 1 , z 1 , x 2 , y 2 , z 2 , …, x 53215 , y 53215 , z 53215 ]is a point cloud coordinate of 159645 dimensions, thenp=E(I),S=D(p)=D(E(I) SoftMax classification layer of expression classifierP j =exp(w j p+ b j )/Σ l exp(w l p+b l ) In whichP j Is as followsjThe probability of a similar expression is,exp() As a natural exponential function, with a cross-entropy penalty ofH(P,Q)=-Σ j Q j log(P j ) In whichPIs the probability predicted by the expression classifier,Qis a true probability distribution of the class,MSEthe loss of mean square error isL(S,T)= Σ i (S i -T i ) 2 /159645WhereinS i Is the output of the decoder for each dimension,T i is each dimension of the real point cloud.
In the pre-training stage, the encoder is trained by using the two-dimensional face picture and the expression parameters of the corresponding three-dimensional deformable model, and the decoder is trained by using the expression parameters of the three-dimensional deformable model and the corresponding three-dimensional face expression point cloud.
And in the joint training stage, expression classification loss and three-dimensional point cloud regression loss are applied simultaneously, and the pre-trained encoder and decoder are subjected to end-to-end joint training.
Specifically, in a joint training stage, two loss constraints, namely expression classification loss and three-dimensional point cloud regression loss, are simultaneously applied to a neural network of an encoder-decoder structure; in order to make the spatial distribution of the expression parameters more separable, the weight of the three-dimensional point cloud regression loss can be reduced at the initial stage of the joint training, and the weights of the expression classification loss and the three-dimensional point cloud regression loss are 1; when the expression classification loss is saturated, in order to ensure that the generated three-dimensional face expression is more natural and real, the weight of the three-dimensional point cloud regression loss is increased, and the weights of the expression classification loss and the three-dimensional point cloud regression loss are 1.
Finally, expression category prediction of the input face image can be realized through an encoder and an expression classifier, and a three-dimensional face expression point cloud can be generated through the encoder and a decoder.
The method of The invention is applied to two common facial expression data sets of The Extended Cohn-Kanade database [6] And The Oulu-CASIA database [7] In the above, the expression classification accuracy rates are respectively 77.2% and 72.8% by adopting the test condition of subject-independent 10 fold cross validation and adopting the average value of the classification accuracy rates of the expression categories as the evaluation index. While using ResNet 34 [2] And support vector machine to implement ExpNet [1] The final expression classification accuracy rates of the feature extraction and classification processes are respectively 63.3% and 56.4%, and the method is obviously lagged behind the method. Therefore, the method of the invention utilizes the end-to-end encoder-decoder structure to learn the three-dimensional deformable model, and can effectively improve the classification effect of the two-dimensional facial expression pictures.
The foregoing description of the embodiments of the present invention has been provided by way of specific examples, and other advantages and benefits of the present invention will become apparent to those skilled in the art from the disclosure herein. The invention is capable of other and different embodiments and of being practiced or of being carried out in various ways, and its several details are capable of modification in various respects, all without departing from the spirit and scope of the present invention.

Claims (4)

1. A two-dimensional facial expression recognition method based on three-dimensional deformable model learning is characterized in that a three-dimensional deformable model learning technology is used for reconstructing three-dimensional facial expression point clouds corresponding to two-dimensional facial expressions, and expression parameters are finally obtained under end-to-end joint training by adopting a network structure of an encoder-decoder and combining an expression classifier, expression classification loss and three-dimensional point cloud regression loss; wherein:
the two-dimensional facial expression is a facial image to be subjected to expression classification;
the three-dimensional facial expression point cloud is a part of the three-dimensional space after the calm face point cloud is removed, namely the point cloud corresponding to the facial action;
the expression parameters are codes of facial expressions, and corresponding three-dimensional facial expression point clouds can be recovered through a decoder;
the encoder is used for encoding the input two-dimensional facial expression and outputting expression parameters;
the decoder is used for decoding the input expression parameters and outputting three-dimensional facial expression point clouds;
the expression classifier adopts a SoftMax classification layer and outputs the prediction probability of each expression;
the expression classification loss is applied to the expression parameters through an expression classifier by adopting cross entropy loss;
the three-dimensional point cloud regression loss is applied to the output of the decoder by adopting MSE mean square error loss.
2. The two-dimensional facial expression recognition method based on three-dimensional deformable model learning as claimed in claim 1, characterized in that the method employs end-to-end joint training; the concrete method is as follows:
the first is a pre-training phase: respectively training an encoder by using the two-dimensional face picture and the expression parameters of the corresponding three-dimensional deformable model, and training a decoder by using the expression parameters of the three-dimensional deformable model and the corresponding three-dimensional face expression point cloud; thus, the encoder and the decoder have preliminary functions;
then a joint training phase: jointly training the pre-trained encoder and decoder, namely inputting expression parameters output by the encoder into the decoder, and generating three-dimensional facial expression point clouds corresponding to the facial pictures end to end; in the combined training process, expression classification loss is applied to expression parameters output by the encoder through the expression classifier, so that the distribution of the expression parameters in the space is more discriminative, the inter-class distance is increased, and the accuracy of expression classification is improved; and meanwhile, the mean square error loss is applied to the three-dimensional facial expression point cloud output by the decoder, so that the finally obtained expression parameters can still generate the corresponding three-dimensional facial expression point cloud.
3. The two-dimensional facial expression recognition method based on three-dimensional deformable Model learning of claim 2, characterized in that the three-dimensional deformable Model adopts Basel Face Model and faceWarehouse, so that the expression parameters are 29-dimensional vectors, and the three-dimensional point cloud has 53215 vertexes; the encoder adopts SphereFace, and a layer of 29-dimensional full-connection layer is added at the top end of the encoder to output 29-dimensional expression parameters; the decoder employs a two-layer multi-layer perceptron, the dimension of the middle layer being 1000, and the dimension of the output layer being 159645.
4. The two-dimensional facial expression recognition method based on three-dimensional deformable model learning of claim 2, characterized in that in the joint training phase, two loss constraints of expression classification loss and three-dimensional point cloud regression loss are simultaneously applied to the neural network of the encoder-decoder structure; in order to make the spatial distribution of expression parameters more separable, in the initial stage of the joint training, the weight of the three-dimensional point cloud regression loss is reduced, and the weights of the expression classification loss and the three-dimensional point cloud regression loss are 1; when the expression classification loss is saturated, in order to ensure that the generated three-dimensional face expression is more natural and real, the weight of the three-dimensional point cloud regression loss is increased, and the weights of the expression classification loss and the three-dimensional point cloud regression loss are 1.
CN202011018179.5A 2020-09-24 2020-09-24 Two-dimensional facial expression recognition method through three-dimensional deformable model learning Active CN112200024B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011018179.5A CN112200024B (en) 2020-09-24 2020-09-24 Two-dimensional facial expression recognition method through three-dimensional deformable model learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011018179.5A CN112200024B (en) 2020-09-24 2020-09-24 Two-dimensional facial expression recognition method through three-dimensional deformable model learning

Publications (2)

Publication Number Publication Date
CN112200024A CN112200024A (en) 2021-01-08
CN112200024B true CN112200024B (en) 2022-10-11

Family

ID=74006606

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011018179.5A Active CN112200024B (en) 2020-09-24 2020-09-24 Two-dimensional facial expression recognition method through three-dimensional deformable model learning

Country Status (1)

Country Link
CN (1) CN112200024B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117011963B (en) * 2023-10-07 2023-12-08 四川金投科技股份有限公司 Intelligent lock and intelligent door control system based on electronic key

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108171770A (en) * 2018-01-18 2018-06-15 中科视拓(北京)科技有限公司 A kind of human face expression edit methods based on production confrontation network
CN108805977A (en) * 2018-06-06 2018-11-13 浙江大学 A kind of face three-dimensional rebuilding method based on end-to-end convolutional neural networks
CN111445582A (en) * 2019-01-16 2020-07-24 南京大学 Single-image human face three-dimensional reconstruction method based on illumination prior

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20170000748A (en) * 2015-06-24 2017-01-03 삼성전자주식회사 Method and apparatus for face recognition
FR3051951B1 (en) * 2016-05-27 2018-06-15 Mimi Hearing Technologies GmbH METHOD FOR PRODUCING A DEFORMABLE MODEL IN THREE DIMENSIONS OF AN ELEMENT, AND SYSTEM THEREOF
US10380413B2 (en) * 2017-07-13 2019-08-13 Robert Bosch Gmbh System and method for pose-invariant face alignment
KR102260214B1 (en) * 2019-08-21 2021-06-03 엘지전자 주식회사 Method of analysis sleeping and intelligent device performing sleep analysis based on artificial intelligence

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108171770A (en) * 2018-01-18 2018-06-15 中科视拓(北京)科技有限公司 A kind of human face expression edit methods based on production confrontation network
CN108805977A (en) * 2018-06-06 2018-11-13 浙江大学 A kind of face three-dimensional rebuilding method based on end-to-end convolutional neural networks
CN111445582A (en) * 2019-01-16 2020-07-24 南京大学 Single-image human face three-dimensional reconstruction method based on illumination prior

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Fully Automated Facial Expression Recognition Using 3D Morphable Model andMesh-Local Binary Pattern";Hela Bejaoui et al.;《International Conference on Advanced Concepts for Intelligent Vision System》;20171130;全文 *
"MeshGAN: Non-linear 3D Morphable Models of Faces";Shiyang Cheng et al.;《arXiv》;20190325;全文 *
"三维人脸表情获取及重建技术综述";王珊等;《系统仿真学报》;20180730;全文 *
"三维可形变人脸重建与实时表情驱动系统";吴迪;《中国优秀硕士学位论文全文数据库 信息科技辑》;20200315;全文 *

Also Published As

Publication number Publication date
CN112200024A (en) 2021-01-08

Similar Documents

Publication Publication Date Title
CN111652236A (en) Lightweight fine-grained image identification method for cross-layer feature interaction in weak supervision scene
Chen et al. Pointgpt: Auto-regressively generative pre-training from point clouds
Deng et al. A voxel graph cnn for object classification with event cameras
WO2023070695A1 (en) Infrared image conversion training method and apparatus, device and storage medium
CN112200024B (en) Two-dimensional facial expression recognition method through three-dimensional deformable model learning
CN114332573A (en) Multi-mode information fusion recognition method and system based on attention mechanism
CN113822368A (en) Anchor-free incremental target detection method
Zhao et al. Similarity-aware fusion network for 3d semantic segmentation
CN110111365B (en) Training method and device based on deep learning and target tracking method and device
CN114972794A (en) Three-dimensional object recognition method based on multi-view Pooll transducer
CN105956604B (en) Action identification method based on two-layer space-time neighborhood characteristics
Song et al. A potential vision-based measurements technology: Information flow fusion detection method using RGB-thermal infrared images
Savadi Hosseini et al. A hybrid deep learning architecture using 3d cnns and grus for human action recognition
Bian et al. Unsupervised domain adaptation for point cloud semantic segmentation via graph matching
Qin et al. Application of video scene semantic recognition technology in smart video
Zhou Video expression recognition method based on spatiotemporal recurrent neural network and feature fusion
Zhu et al. Vehicle type recognition algorithm based on improved network in network
Chen et al. S-votenet: Deep hough voting with spherical proposal for 3d object detection
Chen et al. Application of data-driven iterative learning algorithm in transmission line defect detection
Song et al. Srrm: Semantic region relation model for indoor scene recognition
CN115497085A (en) Point cloud completion method and system based on multi-resolution dual-feature folding
CN103793720A (en) Method and system for positioning eyes
Ameur et al. Unconstrained face verification based on monogenic binary pattern and convolutional neural network
Gu et al. Analysis of art classroom teaching behavior based on intelligent image recognition
Shi et al. Pose-robust facial expression recognition by 3D morphable model learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant