WO2003030084A2 - Face recognition from a temporal sequence of face images - Google Patents
Face recognition from a temporal sequence of face images Download PDFInfo
- Publication number
- WO2003030084A2 WO2003030084A2 PCT/IB2002/003690 IB0203690W WO03030084A2 WO 2003030084 A2 WO2003030084 A2 WO 2003030084A2 IB 0203690 W IB0203690 W IB 0203690W WO 03030084 A2 WO03030084 A2 WO 03030084A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- images
- image
- face
- probe
- higher resolution
- Prior art date
Links
- 230000002123 temporal effect Effects 0.000 title claims abstract description 19
- 238000000034 method Methods 0.000 claims abstract description 48
- 239000000523 sample Substances 0.000 claims abstract description 31
- 238000012549 training Methods 0.000 claims abstract description 19
- 230000001815 facial effect Effects 0.000 claims abstract description 17
- 239000011159 matrix material Substances 0.000 claims description 11
- 230000004913 activation Effects 0.000 claims description 7
- 238000001994 activation Methods 0.000 claims description 7
- 238000012360 testing method Methods 0.000 claims description 7
- 238000001514 detection method Methods 0.000 claims description 4
- 238000003064 k means clustering Methods 0.000 claims description 3
- 102000016550 Complement Factor H Human genes 0.000 claims description 2
- 108010053085 Complement Factor H Proteins 0.000 claims description 2
- 239000013598 vector Substances 0.000 description 15
- 238000010586 diagram Methods 0.000 description 4
- 238000013528 artificial neural network Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 3
- 238000003909 pattern recognition Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000010420 art technique Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
Definitions
- the present invention relates to face recognition systems and particularly, to a system and method for performing face recognition using a temporal sequence of face images in order to improve the robustness of recognition.
- Face recognition is an important research area in human computer interaction and many algorithms and classifier devices for recognizing faces have been proposed.
- face recognition systems store a full facial template obtained from multiple instances of a subject's face during training of the classifier device, and compare a single probe (test) image against the stored templates to recognize the individual.
- Fig. 1 illustrates a traditional classifier device 10 comprising, for example, a Radial Basis Function (RBF) network having a layer 12 of input nodes, a hidden layer 14 comprising radial basis functions and an output layer 18 for providing a classification.
- RBF Radial Basis Function
- a description of an RBF classifier device is available from commonly-owned, co-pending Unites States Patent Application Serial No. 09/794,443 entitled Classification of objects through model ensembles filed February 27, 2001, the whole contents and disclosure of which is incorporated by reference as if fully set forth herein.
- a single probe (test) image 25 including input vectors 26 comprising data representing pixel values of the image is compared against the stored templates for face recognition.
- face recognition from a single face image is a difficult problem, especially when that face image is not completely frontal.
- a video clip of an individual is available for such a face recognition task.
- a lot of temporal information is wasted.
- a system and method for classifying facial images from a temporal sequence of images comprising the steps of: a) training a classifier device for recognizing facial images, said classifier device being trained with input data associated with a full facial image; b) obtaining a plurality of probe images of said temporal sequence of images; c) aligning each of said probe images with respect to each other; d) combining said images to form a higher resolution image; and, e) classifying said higher resolution image according to a classification method performed by said trained classifier device.
- the system and method of the invention enables the combination of several partial views of a face image to create a better single view of the face for recognition.
- the success rate of the face recognition is related to the resolution of the image, the higher the resolution, the higher the success rate. Therefore, the classifier is trained with the high-resolution images. If a single low-resolution image is received, the recognizer will still work, but if a temporal sequence is received, a high-resolution image is created and the classifier will work even better.
- Fig. 1 is a diagram depicting an RBF classifier device 10 applied for face recognition and classification according to prior art techniques
- Fig. 2 is a diagram depicting an RBF classifier device 10' implemented for face recognition in accordance with the principles of the invention.
- Fig. 3 is a diagram depicting how a high resolution image is created after warping.
- Fig. 2 illustrates a proposed classifier 10' of the invention that enables multiple probe images 40 of the same individual from a sequence of images are used simultaneously. It is understood that for pmposes of description an RBF network 10' may be used, however, any classification method/device may be implemented.
- a shape that will correspond to any given head may be produced, with a pre-set precision, i.e., the higher the number of points the better precision; 4) View morphing techniques, whereby given an image and a 3-D structure of the scene, an exact image may be created that will correspond to an image obtained from the same camera in the arbitrary- position of the scene. Some view morphing techniques do not require an exact, but only an approximate 3-D structure of the scene and still provide very good results such as described in the reference to S.J. Gortler, R. Grzeszczuk, R. Szelisky and M.F.
- one or more faces comprising the plurality of images 40 is oriented differently in each probe image and is not fully visible on each probe image. If just one of the probe images (for instance, one without a frontal view) is used instead, current face recognition systems may not be able to recognize the individual from this single non-frontal face image since they require a face image that may be, at most, ⁇ 15° from the fully frontal position.
- the multiple probe images are combined together into a single higher resolution image.
- these images are aligned with each other based on correspondences from the warping methods applied in accordance with the teachings of commonly-owned, co-pending U.S. Patent Application Serial No. 09/966406 [Attorney Docket 702053, Atty D# 14901] and, once this is performed, at most pixel points (i, j), there are as many pixels available as the number of probe images. It is understood that after alignment, there may be some locations where not all the probe images contribute to after warping them. The resolution is simply increased as there are many pixel values available at each location.
- Fig. 3 is a diagram depicting conceptually how a high-resolution image is created after warping. As shown in Fig. 3, points 50a-50d points denote pixels of an image 45 at locations corresponding to a frontal view of a face. Points 60 correspond to the position of points from other images from the given temporal sequence 40 after warping them into image 45.
- Points 75 correspond to the inserted pixels of a resulting high-resolution image.
- the image value at these locations is computed as an interpolation of the points 60.
- One method for doing this is to fit a surface to points 50a-50d and points 60 (any polynomial would do) and then estimate value of the polynomial at the location of interpolated points 75.
- the successive face images i.e., probe images, are extracted from test sequence automatically from the output of some face detection/tracking algorithm well known in the art, such as the system described in the reference to A. J. Colmenarez and T. S. Huang entitled “Face detection with information-based maximum discrimination," Proc. IEEE Computer Vision and Pattern Recognition, Puerto Rico, USA, pp. 782-787, 1997, the whole contents and disclosure of which is incorporated by reference as if fully set forth herein.
- a Radial Basis Function (“RBF") classifier such as shown in Fig. 2, is implemented, but it is understood that any classification method/device may be implemented.
- RBF Radial Basis Function
- a description of an RBF classifier device is available from commonly- owned, co-pending Unites States Patent Application Serial No. 09/794,443 entitled Classification of objects through model ensembles filed February 27, 2001, the whole contents and disclosure of which is incorporated by reference as if fully set forth herein.
- the construction of an RBF network as disclosed in commonly-owned, co- pending Unites States Patent Application Serial No. 09/794,443, is now described with reference to Fig. 2. As shown in Fig.
- the RBF network classifier 10' is structured in accordance with a traditional three-layer back-propagation network including a first input layer 12 made up of source nodes (e.g., k sensory units); a second or hidden layer 14 comprising i nodes whose function is to cluster the data and reduce its dimensionality; and, a third or output layer 18 comprising / nodes whose function is to supply the responses 20 of the network 10' to the activation patterns applied to the input layer 12.
- the transformation from the input space to the hidden-unit space is non-linear, whereas the transformation from the hidden-unit space to the output space is linear.
- an RBF classifier network 10' may be viewed in two ways: 1) to interpret the RBF classifier as a set of kernel functions that expand input vectors into a high-dimensional space in order to take advantage of the mathematical fact that a classification problem cast into a high-dimensional space is more likely to be linearly separable than one in a low-dimensional space; and, 2) to interpret the RBF classifier as a function-mapping interpolation method that tries to construct hypersurfaces, one for each class, by taking a linear combination of the Basis Functions (BF).
- BF Basis Functions
- An unknown input vector is classified as belonging to the class associated with the hypersurface with the largest output at that point.
- the BFs do not serve as a basis for a high-dimensional space, but as components in a finite expansion of the desired hypersurface where the component coefficients, (the weights) have to be trained.
- BF nodes in the hidden layer 14 i.e., called Basis Function (BF) nodes
- ⁇ , 2 represents the diagonal entries of the covariance matrix of Gaussian pulse (i).
- each BF node (i) outputs a scalar value y t reflecting the activation of the BF caused by that input as represented by equation 1) as follows:
- ⁇ is a proportionality constant for the variance
- ⁇ , ⁇ and ⁇ ,* are the k components of the mean and variance vectors, respectively, of basis node (i). Inputs that are close to the center of the Gaussian BF result in higher activations, while those that are far away result in lower activations.
- each output node 18 of the RBF network forms a linear combination of the BF node activations
- Z j is the output of the h output node
- v is the activation of the i th BF node
- wy is the weight 24 connecting the i th BF node to the h output node
- w ⁇ j is the bias or threshold of the h output node. This bias comes from the weights associated with a BF node that has a constant unit output regardless of the input.
- An unknown vector X is classified as belonging to the class associated with the output nodey with the largest output z 7 .
- the weights wy in the linear network are not solved using iterative minimization methods such as gradient descent. They are determined quickly and exactly using a matrix pseudo-inverse technique such as described in above- mentioned reference to C. M. Bishop, "Neural Networks for Pattern Recognition,” Clarendon Press, Oxford, 1997.
- a detailed algorithmic description of the preferable RBF classifier that may be implemented in the present invention is provided herein in Tables 1 and 2. As shown in Table 1, initially, the size of the RBF network 10' is determined by selecting F, the number of BFs nodes.
- F is problem-specific and usually depends on the dimensionality of the problem and the complexity of the decision regions to be formed. In general, F can be determined empirically by trying a variety of Fs, or it can set to some constant number, usually larger than the input dimension of the problem. After F is set, the mean ⁇ 7 and variance ⁇ vectors of the BFs may be determined using a variety of methods. They can be trained along with the output weights using a back-propagation gradient descent technique, but this usually requires a long training time and may lead to suboptimal local minima. Alternatively, the means and variances may be determined before training the output weights. Training of the networks would then involve only determining the weights.
- the BF means (centers) and variances (widths) are normally chosen so as to cover the space of interest.
- Different techniques may be used as known in the art: for example, one technique implements a grid of equally spaced BFs that sample the input space; another technique implements a clustering algorithm such as k-means to determine the set of BF centers; other techniques implement chosen random vectors from the training set as BF centers, making sure that each class is represented.
- the BF variances or widths ⁇ f 2 may be set. They can be fixed to some global value or set to reflect the density of the data vectors in the vicinity of the BF center.
- a global proportionality factor H for the variances is included to allow for rescaling of the BF widths. By searching the space of H for values that result in good performance, its proper value is determined.
- the next step is to train the output weights wy in the linear network.
- Individual training patterns X(p) and their class labels C(p) are presented to the classifier, and the resulting BF node outputs y ⁇ (p), are computed.
- These and desired outputs d p) are then used to determine the Ex F correlation matrix "R" and the Fx M output matrix "B".
- each training pattern produces one R and B matrices.
- the final R and B matrices are the result of the sum of N individual R and B matrices, where N is the total number of training patterns. Once all N patterns have been presented to the classifier, the output weights wy are determined.
- the final correlation matrix R is inverted and is used to determine each w, . 1.
- classification is performed by presenting an unknown input vector X test to the trained classifier and computing the resulting BF node outputs y . . These values are then used, along with the weights w tj , to compute the output values Z j . The input vector X test is then classified as belonging to the class associated with the output nodey ' with the largest z, output.
- the RBF input comprises a temporal sequence of n size normalized facial gray-scale images fed to the network RBF network 10' as one-dimensional, i.e., 1-D vectors 30.
- the hidden (unsupervised) layer 14 implements an "enhanced" k-means clustering procedure, such as described in S. Gutta, J. Huang, P.
- the number of clusters may vary, in steps of 5, for instance, from 1/5 of the number of traimng images to n, the total number of training images.
- the width ⁇ of the Gaussian for each cluster is set to the maximum (the distance between the center of the cluster and the farthest away member - within class diameter, the distance between the center of the cluster and closest pattern from all other clusters) multiplied by an overlap factor o, here equal to 2.
- the width is further dynamically refined using different proportionality constants h.
- the hidden layer 14 yields the equivalent of a functional shape base, where each cluster node encodes some common characteristics across the shape space.
- the output (supervised) layer maps face encodings ('expansions') along such a space to their corresponding ID classes and finds the corresponding expansion ('weight') coefficients using pseudo-inverse techniques. Note that the number of clusters is frozen for that configuration (number of clusters and specific proportionality constant h) which yields 100 % accuracy on ID classification when tested on the same training images.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2003533210A JP2005512172A (en) | 2001-09-28 | 2002-09-10 | Facial recognition from time series of facial images |
EP02762710A EP1586071A2 (en) | 2001-09-28 | 2002-09-10 | Face recognition from a temporal sequence of face images |
KR10-2004-7004558A KR20040037179A (en) | 2001-09-28 | 2002-09-10 | Face recognition from a temporal sequence of face images |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/966,409 US20030063781A1 (en) | 2001-09-28 | 2001-09-28 | Face recognition from a temporal sequence of face images |
US09/966,409 | 2001-09-28 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2003030084A2 true WO2003030084A2 (en) | 2003-04-10 |
WO2003030084A3 WO2003030084A3 (en) | 2005-08-25 |
Family
ID=25511355
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2002/003690 WO2003030084A2 (en) | 2001-09-28 | 2002-09-10 | Face recognition from a temporal sequence of face images |
Country Status (6)
Country | Link |
---|---|
US (1) | US20030063781A1 (en) |
EP (1) | EP1586071A2 (en) |
JP (1) | JP2005512172A (en) |
KR (1) | KR20040037179A (en) |
CN (1) | CN1636226A (en) |
WO (1) | WO2003030084A2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100643303B1 (en) | 2004-12-07 | 2006-11-10 | 삼성전자주식회사 | Method and apparatus for detecting multi-view face |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2003096269A1 (en) * | 2002-05-10 | 2003-11-20 | Sony Corporation | Information processing apparatus and method |
CN1797420A (en) * | 2004-12-30 | 2006-07-05 | 中国科学院自动化研究所 | Method for recognizing human face based on statistical texture analysis |
US20060217925A1 (en) * | 2005-03-23 | 2006-09-28 | Taron Maxime G | Methods for entity identification |
JP4686505B2 (en) * | 2007-06-19 | 2011-05-25 | 株式会社東芝 | Time-series data classification apparatus, time-series data classification method, and time-series data processing apparatus |
KR101363017B1 (en) * | 2007-08-23 | 2014-02-12 | 삼성전자주식회사 | System and methed for taking pictures and classifying the pictures taken |
SG152952A1 (en) * | 2007-12-05 | 2009-06-29 | Gemini Info Pte Ltd | Method for automatically producing video cartoon with superimposed faces from cartoon template |
US9405995B2 (en) * | 2008-07-14 | 2016-08-02 | Lockheed Martin Corporation | Method and apparatus for facial identification |
US8948476B2 (en) | 2010-12-20 | 2015-02-03 | St. Jude Medical, Atrial Fibrillation Division, Inc. | Determination of cardiac geometry responsive to doppler based imaging of blood flow characteristics |
US20100168557A1 (en) * | 2008-12-30 | 2010-07-01 | Deno D Curtis | Multi-electrode ablation sensing catheter and system |
US8900150B2 (en) | 2008-12-30 | 2014-12-02 | St. Jude Medical, Atrial Fibrillation Division, Inc. | Intracardiac imaging system utilizing a multipurpose catheter |
US9610118B2 (en) * | 2008-12-31 | 2017-04-04 | St. Jude Medical, Atrial Fibrillation Division, Inc. | Method and apparatus for the cancellation of motion artifacts in medical interventional navigation |
US9928406B2 (en) | 2012-10-01 | 2018-03-27 | The Regents Of The University Of California | Unified face representation for individual recognition in surveillance videos and vehicle logo super-resolution system |
CN104318215B (en) * | 2014-10-27 | 2017-09-19 | 中国科学院自动化研究所 | A kind of cross-view face identification method based on domain robust convolution feature learning |
US10860887B2 (en) | 2015-11-16 | 2020-12-08 | Samsung Electronics Co., Ltd. | Method and apparatus for recognizing object, and method and apparatus for training recognition model |
US10417533B2 (en) * | 2016-08-09 | 2019-09-17 | Cognex Corporation | Selection of balanced-probe sites for 3-D alignment algorithms |
US11714881B2 (en) | 2021-05-27 | 2023-08-01 | Microsoft Technology Licensing, Llc | Image processing for stream of input images with enforced identity penalty |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5686960A (en) * | 1992-01-14 | 1997-11-11 | Michael Sussman | Image input device having optical deflection elements for capturing multiple sub-images |
US5251037A (en) * | 1992-02-18 | 1993-10-05 | Hughes Training, Inc. | Method and apparatus for generating high resolution CCD camera images |
JP2989364B2 (en) * | 1992-03-12 | 1999-12-13 | シャープ株式会社 | Image processing apparatus and image processing method |
US5341174A (en) * | 1992-08-17 | 1994-08-23 | Wright State University | Motion compensated resolution conversion system |
US5696848A (en) * | 1995-03-09 | 1997-12-09 | Eastman Kodak Company | System for creating a high resolution image from a sequence of lower resolution motion images |
US6496594B1 (en) * | 1998-10-22 | 2002-12-17 | Francine J. Prokoski | Method and apparatus for aligning and comparing images of the face and body from different imagers |
US6650704B1 (en) * | 1999-10-25 | 2003-11-18 | Irvine Sensors Corporation | Method of producing a high quality, high resolution image from a sequence of low quality, low resolution images that are undersampled and subject to jitter |
US6778705B2 (en) * | 2001-02-27 | 2004-08-17 | Koninklijke Philips Electronics N.V. | Classification of objects through model ensembles |
-
2001
- 2001-09-28 US US09/966,409 patent/US20030063781A1/en not_active Abandoned
-
2002
- 2002-09-10 JP JP2003533210A patent/JP2005512172A/en not_active Withdrawn
- 2002-09-10 CN CNA028189973A patent/CN1636226A/en active Pending
- 2002-09-10 WO PCT/IB2002/003690 patent/WO2003030084A2/en not_active Application Discontinuation
- 2002-09-10 KR KR10-2004-7004558A patent/KR20040037179A/en not_active Application Discontinuation
- 2002-09-10 EP EP02762710A patent/EP1586071A2/en not_active Withdrawn
Non-Patent Citations (6)
Title |
---|
BAKER S ET AL: "HALLUCINATING FACES" PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, XX, XX, March 2000 (2000-03), pages 83-88, XP001028178 * |
BAKER S ET AL: "Super-resolution optical flow" TECHNICAL REPORT, ROBOTICS INSTITUTE, CARNEGIE MELLON UNIVERSITY, CMU-RI-TR-99-36, October 1999 (1999-10) - October 1999 (1999-10), pages 1-13, XP002223179 * |
KOH L H ET AL: "An integrated face detection and recognition system" IMAGE ANALYSIS AND PROCESSING, 1999. PROCEEDINGS. INTERNATIONAL CONFERENCE ON VENICE, ITALY 27-29 SEPT. 1999, LOS ALAMITOS, CA, USA,IEEE COMPUT. SOC, US, 27 September 1999 (1999-09-27), pages 532-537, XP010354302 ISBN: 0-7695-0040-4 * |
LIN S-H ET AL: "FACE RECOGNITION/DETECTION BY PROBABILISTIC DECISION-BASED NEURAL NETWORK" IEEE TRANSACTIONS ON NEURAL NETWORKS, IEEE INC, NEW YORK, US, vol. 8, no. 1, 1997, pages 114-132, XP000640676 ISSN: 1045-9227 * |
QIAN CHEN ET AL: "Head pose estimation using both color and feature information" PATTERN RECOGNITION, 2000. PROCEEDINGS. 15TH INTERNATIONAL CONFERENCE ON SEPTEMBER 3-7, 2000, LOS ALAMITOS, CA, USA,IEEE COMPUT. SOC, US, 3 September 2000 (2000-09-03), pages 842-845, XP010533945 ISBN: 0-7695-0750-6 * |
TOLBA A S ET AL: "Combined classifiers for invariant face recognition" INFORMATION INTELLIGENCE AND SYSTEMS, 1999. PROCEEDINGS. 1999 INTERNATIONAL CONFERENCE ON BETHESDA, MD, USA 31 OCT.-3 NOV. 1999, LOS ALAMITOS, CA, USA,IEEE COMPUT. SOC, US, 31 October 1999 (1999-10-31), pages 350-359, XP010362277 ISBN: 0-7695-0446-9 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100643303B1 (en) | 2004-12-07 | 2006-11-10 | 삼성전자주식회사 | Method and apparatus for detecting multi-view face |
Also Published As
Publication number | Publication date |
---|---|
WO2003030084A3 (en) | 2005-08-25 |
KR20040037179A (en) | 2004-05-04 |
JP2005512172A (en) | 2005-04-28 |
CN1636226A (en) | 2005-07-06 |
US20030063781A1 (en) | 2003-04-03 |
EP1586071A2 (en) | 2005-10-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Kumar et al. | Object detection system based on convolution neural networks using single shot multi-box detector | |
US10949649B2 (en) | Real-time tracking of facial features in unconstrained video | |
Yang et al. | Extraction of 2d motion trajectories and its application to hand gesture recognition | |
Moghaddam et al. | Probabilistic visual learning for object representation | |
Moghaddam et al. | Bayesian face recognition using deformable intensity surfaces | |
EP1433118B1 (en) | System and method of face recognition using portions of learned model | |
US6628821B1 (en) | Canonical correlation analysis of image/control-point location coupling for the automatic location of control points | |
EP1586071A2 (en) | Face recognition from a temporal sequence of face images | |
Moeini et al. | Real-world and rapid face recognition toward pose and expression variations via feature library matrix | |
JP2005512201A5 (en) | ||
JP2005512201A (en) | Face recognition using kernel fisher face | |
Liang et al. | Accurate face alignment using shape constrained Markov network | |
Li et al. | A data-driven approach for facial expression retargeting in video | |
Akakın et al. | Robust classification of face and head gestures in video | |
Xu et al. | A high resolution grammatical model for face representation and sketching | |
WO2003030089A1 (en) | System and method of face recognition through 1/2 faces | |
Riaz et al. | Age-invariant face recognition using gender specific 3D aging modeling | |
US20030063795A1 (en) | Face recognition through warping | |
Li | Hand shape estimation for south african sign language | |
Elgarrai et al. | Offline face recognition system based on gaborfisher descriptors and hidden markov models | |
Mestri et al. | Analysis of feature extraction and classification models for lip-reading | |
Huang et al. | Subface hidden Markov models coupled with a universal occlusion model for partially occluded face recognition | |
Kumar et al. | Facial emotion recognition system using entire feature vectors and supervised classifier | |
Spurlock et al. | Multi-view action recognition one camera at a time | |
Hüsken et al. | Evaluation of implicit 3D modeling for pose-invariant face recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): CN JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FR GB GR IE IT LU MC NL PT SE SK TR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2003533210 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2002762710 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 20028189973 Country of ref document: CN Ref document number: 1020047004558 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2002762710 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2002762710 Country of ref document: EP |