CN107704810A - A kind of expression recognition method suitable for medical treatment and nursing - Google Patents

A kind of expression recognition method suitable for medical treatment and nursing Download PDF

Info

Publication number
CN107704810A
CN107704810A CN201710828671.0A CN201710828671A CN107704810A CN 107704810 A CN107704810 A CN 107704810A CN 201710828671 A CN201710828671 A CN 201710828671A CN 107704810 A CN107704810 A CN 107704810A
Authority
CN
China
Prior art keywords
shape
mrow
point
expression
msub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710828671.0A
Other languages
Chinese (zh)
Inventor
刘健
孙瑜
葛天洁
贾新旺
秦岭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Science and Technology
Original Assignee
Nanjing University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Science and Technology filed Critical Nanjing University of Science and Technology
Priority to CN201710828671.0A priority Critical patent/CN107704810A/en
Publication of CN107704810A publication Critical patent/CN107704810A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a kind of expression recognition method suitable for medical treatment and nursing, comprise the following steps:Active shape model is established using training image;Test image is input in active shape model and searches for human face characteristic point;Extraction includes the mood vector characteristics of angle information and range information, and human face characteristic point is substituted into mood vector characteristics, and the information of mood vector characteristics is inputted and voted into SVM classifier, and who gets the most votes's class is affiliated class.

Description

A kind of expression recognition method suitable for medical treatment and nursing
Technical field
The present invention relates to a kind of image processing techniques, particularly a kind of expression recognition method suitable for medical treatment and nursing.
Background technology
Expression recognition technology is in recent years with the rapid development such as machine learning of some association areas, at image Reason, become the technology of a focus development.The influence of expression recognition system and potentiality are generalized to vast application simultaneously In occasion, such as man-machine interaction, intelligent robot, driver status supervision etc..Expression recognition system is computer understanding The premise of people's emotion, and people explore intelligence, understand the effective way of intelligence.Personalizing for computer how is realized, is made It can adaptively provide most friendly operation ring according to contents such as the states of the environment of surrounding and object for communicatee Border, oneself is intelligent robot asking of coming into that people's daily life must solve through turning into the target of man-machine interface of future generation development Topic, to establishing multi information intelligent man-machine interaction system important in inhibiting.Meanwhile as China human mortality old-age group trend is aggravated, Medical treatment and nursing industry welcomes rapid development, on the road of medical care industry intelligent development, to the Expression Recognition of old physical disabilities, The identification of particularly negative feeling has great importance in actual applications.
Expression recognition technology generally comprises three parts content:(1) Face datection;(2) extraction of human face expression feature; (3) classification of expressive features.Wherein human face expression feature extraction be in whole system the most core the step of, feature extraction is direct Have influence on the precision, robustness and real-time of identification.The method of common face characteristic extraction includes:Based on geometric properties, table See feature, the method for behavioral characteristics.Generally, although expression recognition accurately extracts expression by development for many years Feature is still exactly the technical barrier of a urgent need to resolve so as to carry out the identification of strong robustness human face expression, while carries out expression certainly The accuracy and real-time of dynamic identification still have to be hoisted.
The content of the invention
It is an object of the invention to provide a kind of expression recognition method suitable for medical treatment and nursing, comprise the following steps:
Step 1, active shape model is established using training image;
Step 2, test image is input in active shape model and searches for human face characteristic point;
Step 3, extraction includes the mood vector characteristics of angle information and range information, by human face characteristic point substitution mood to In measure feature, and the information of mood vector characteristics is inputted and voted into SVM classifier, who gets the most votes's class is affiliated Class.
Using the above method, step 1 specifically includes:
Step 1.1, N training images are obtained and n mark point is marked on training image and represents flat shape with X
X=[x1,x2,...xj...xn,y1,y2,...yj...yn]T (1)
Wherein, (xj,yj) it is j-th of coordinate for marking point;
Step 1.2, the average shape of training image is obtained
Step 1.3, the characteristic vector of the covariance matrix of average shape is obtained
Step 1.4, by characteristic vectorIn answer characteristic value to arrange from big to small and q characteristic vector forms before therefrom choosing Matrix Φs
Φs=[Φ1,...Φj,...Φq] (4)
Step 1.5, active shape model Y is obtaineds
Wherein, bsIt is the parameter set of shape.
Using the above method, middle test image is input in active shape model searches for face characteristic according to the following steps Point:
Step 2.1, i=0 is made, uses average shapeAs initialization shape Xi
Step 2.2, to current shape, mahalanobis distance is calculated for each mark point, selects the point of minimum range to be used as and is somebody's turn to do Mark the new position (x of pointj',yj');
Step 2.3, adjusting parameter tx、ty、θ、s、bs, make (xj',yj') withMatching, (tx,ty) it is figure As translation distance, θ is image rotation angle, and s is image scaling size, bsFor the parameter set of shape, M represents the mistake of adjustment Journey, M-1For M inverse transformation;
Step 2.4, makeIf Xi=Xi+1, then exit;Otherwise, i=i+1,2.2 are gone to step.
Using the above method, the angle information angle triangle for three human face characteristic points in step 3, if this three Angular angle is in the angular range corresponding to an expression, then expression is then fallen into corresponding to the angular range corresponding to the angle Expression in;Range information is the distance between two human face characteristic points in step 3, if the distance corresponding to an expression away from From in the range of, then expression is then fallen into the expression corresponding to the distance range corresponding to the distance.
The present invention proposes a kind of new Extraction of Geometrical Features method based on ASM, calculates vector angle and distance feature, Make Expression Recognition more accurate.
The present invention is described further with reference to Figure of description.
Brief description of the drawings
Fig. 1 is expression recognition schematic flow sheet.
Fig. 2 is the manual calibration maps of human face characteristic point in embodiment.
Fig. 3 is face mood vector characteristics angle and distance feature schematic diagram in embodiment.
Embodiment
With reference to Fig. 1, a kind of expression recognition method suitable for medical treatment and nursing, comprise the following steps:
Step 1, active shape model is established using training image;
Step 2, test image is input in active shape model and searches for human face characteristic point;
Step 3, extraction includes the mood vector characteristics of angle information and range information, by human face characteristic point substitution mood to In measure feature, and the information of mood vector characteristics is inputted and voted into SVM classifier, who gets the most votes's class is affiliated Class.
" class " can simply be interpreted as the recognition result of svm classifier, and identification process is to utilize software LIBSVM adjusting parameters Classification results are obtained, are prior arts, therefore do not illustrate design parameter in the patent.Last recognition result is glad respectively, in Property and sad three major types.Obtain this result and think that identification is completed.
Specifically
Step 1:Active shape model is established using training image
An ASM model is established first with training image, marks a number of point on training picture manually, these Point all concentrates on as countenance changes and changes most places, such as eyebrow, lip, eyes etc..If these point marks Inaccuracy, recognition result can decline.The facial image of characteristic point mark is as shown in Figure 2.
Active shape model (ASM) is a kind of Object shape description technology, is a kind of feature matching method based on model, It both can neatly change the shape of model to adapt to the uncertain characteristic of target shape, and the change of shape is controlled in mould Type allow in the range of, so as to ensure model change when will not be affected by various factors and there is irrational shape.Its base This thought is to choose one group of training sample, and the shape of sample is described with one group of characteristic point, and then the shape of each sample is carried out Registering (so that shape is similar as much as possible), carries out statistics using principal component method to the shape vector after these registrations and builds Mould obtains the statistical description of body form, finally contour of object is searched in new image using the model of foundation, depending on Position goes out target object.
Assuming that n point forms an object, N is the number of all training images, and vectorial X represents flat shape:
X=[x1,x2,...xj...xn,y1,y2,...yj...yn]T
Wherein, (xj,yj) coordinate of j-th of mark point is represented, n is the number for forming all points of body form.
The average shape of one object, it can be obtained from following formula:
The characteristic vector of the covariance matrix of average shapeExpression formula is:
Therefrom select q maximum characteristic vector of character pair value, composition matrix Φs, expression formula is:
Then obtaining ASM models from the training atlas of manual feature point for calibration is
Wherein, YsIt is the shape of target object,It is average shape, bsIt is the parameter set of shape, ΦsIt is association The matrix of the characteristic vector composition of variance matrix, it is orthogonal.
Step 2:Test image is input in ASM models, scans for matching.
Step 201:I=0 is made, uses average shapeAs initialization shape Xi
Step 202:To current shape in each mark point, mahalanobis distance is calculated, selects that point conduct of ultimate range New position (the x of the calibration pointj',yj');
Step 203:The transformation parameter t such as adjustment translation, rotation and yardstickx,ty, θ, s, and form parameter bs, make (xj', yj') withMatching, orderM represents such a operation, is by image translation (tx、 ty), image rotation θ, graphical rule scaling s, bsThe parameter set of shape, can using visual interpretation as:Arbitrary shape can be with Approximate representation is " deformation " to average shape, and this " deformation " is to variously-shaped change weighting by form parameter With model.These parameters are adjusted, are commonly called as " aliging ", in order to make the optimal mark glyph calculated under mahalanobis distance Close ASM models.M-1That is M inverse transformation, I did not attempted this algorithm in person, was seen in bibliography;
Step 204:If Xi=Xi+1Then exit, otherwise go to step 202.
Mahalanobis distance calculating process in step 202 includes:
First to every width training image feature point for calibration edge by its angular bisector direction, using the characteristic point of demarcation in The heart takes npIndividual pixel, the following vector of its gray value composition
Gray scale can be obtained prolong the derivative in angular bisector direction and be:
After the derivative is standardized, have
To carrying out identical calculating on each image, obtain j-th of mark point on all images half-tone information it is flat Average
Wherein, i represents i width images, represents j-th of mark point of jth width image
Calculate covariance
For standardized grayscale vector h' corresponding to a certain mark point in test imagej, with Average normalized gray scale derivative The mahalanobis distance of vector
In summary, the matching of ASM models is exactly to utilize the profile of the target object in algorithm search test image.Searching Suo Zhong, the optimized parameter of target shape is obtained by comparison reference model and test image, object wheel is searched in new image Exterior feature, so as to orient target object.
Further, the present invention is improved for the method for existing extraction geometric properties, in order that it is more applicable In medical treatment and nursing industry, i.e., there is more unitized judgement to patient's passiveness expression (including sadness, pain etc.), to geometric properties Improved, it is proposed that mood vector characteristics (Motion Vector Feature)
Easily found from the picture in JAFFE expressions storehouse, people's horizontal length of mouth when glad and sad be it is different, But the direct length of mouth is extracted to carry out judgement be unscientific, and because the mouth of people may be big, Ke Neng little, but the nose of people It is rigid, it is not easy to change, just form a triangle with three points on nose and corners of the mouth both sides, the triangle can be utilized Some features carry out expression identification, it is specific as follows:
Triangle OAB, it is extracted first mood vector characteristics angle information.Assuming that it is respectively with 3 points of coordinate:
Wherein i=1,2 ... N
Wherein, N is the number of facial image,J=43,44,49 is the seat on three summits in the i-th width image Mark represents that the label of the point of nose is 43, and the label at corners of the mouth both ends is 44,49 respectively.
It is possible to further calculateWithAngle, can be used for that difference is sad, neutral glad expression, Formula is as follows:
By observing JAFFE expression datas storehouse, it can be seen that when happiness, the mouth of people is usually what is opened, and sad When, the mouth of people closes, and extracts distance feature, that is, represents four points of upper lip and lower mouth
The distance between four corresponding points of lip, can be obtained by following formula
d5961=| y59-y61|
Wherein, y59It is the 59th point of ordinate, y61It is the 61st point of ordinate, d5961It is the vertical seat of the 59th point and the 61st point The absolute value of the difference of mark, remaining three similar.
Because every face there are 66 points, the coordinate of these points, also drop it off in mood vector characteristics, it is possible to use Following vector expression represents the shape of face:
X=[x1,x2,...xj...xn,y1,y2,...yj...yn]T
The feature thus needed, it will send these characteristic informations in SVM classifier voted below, Who gets the most votes's class is affiliated class.

Claims (4)

1. a kind of expression recognition method suitable for medical treatment and nursing, it is characterised in that comprise the following steps:
Step 1, active shape model is established using training image;
Step 2, test image is input in active shape model and searches for human face characteristic point;
Step 3, extraction includes the mood vector characteristics of angle information and range information, and it is special that human face characteristic point is substituted into mood vector In sign, and the information of mood vector characteristics is inputted and voted into SVM classifier, who gets the most votes's class is affiliated class.
2. according to the method for claim 1, it is characterised in that step 1 specifically includes:
Step 1.1, N training images are obtained and n mark point is marked on training image and represents flat shape with X
X=[x1,x2,...xj...xn,y1,y2,...yj...yn]T (1)
Wherein, (xj,yj) it is j-th of coordinate for marking point;
Step 1.2, the average shape of training image is obtained
<mrow> <mover> <mi>X</mi> <mo>&amp;OverBar;</mo> </mover> <mo>=</mo> <mfrac> <mn>1</mn> <mi>N</mi> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>X</mi> <mi>j</mi> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow>
Step 1.3, the characteristic vector of the covariance matrix of average shape is obtained
Step 1.4, by characteristic vectorIn answer characteristic value to arrange from big to small and q characteristic vector forms matrix Φ before therefrom choosings
Φs=[Φ1,...Φj,...Φq] (4)
Step 1.5, active shape model Y is obtaineds
<mrow> <msub> <mi>Y</mi> <mi>s</mi> </msub> <mo>=</mo> <mover> <mi>X</mi> <mo>&amp;OverBar;</mo> </mover> <mo>+</mo> <msub> <mi>&amp;Phi;</mi> <mi>s</mi> </msub> <msub> <mi>b</mi> <mi>s</mi> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>5</mn> <mo>)</mo> </mrow> </mrow>
Wherein, bsIt is the parameter set of shape.
3. according to claim, the method described in 2, it is characterised in that in by test image be input in active shape model by Following steps search for human face characteristic point:
Step 2.1, i=0 is made, uses average shapeAs initialization shape Xi
Step 2.2, to current shape, mahalanobis distance is calculated for each mark point, selects the point of minimum range as the mark New position (the x of pointj',yj');
Step 2.3, adjusting parameter tx、ty、θ、s、bs, make (xj',yj') withMatching, (tx,ty) it is image translation Distance, θ are image rotation angle, and s is image scaling size, bsFor the parameter set of shape, M represents the process of adjustment, M-1 For M inverse transformation;
Step 2.4, makeIf Xi=Xi+1, then exit;Otherwise, i=i+1,2.2 are gone to step.
4. according to the method for claim 1, it is characterised in that angle information is three human face characteristic point institute structures in step 3 Into the angle of triangle, if the angle of the triangle in the angular range corresponding to an expression, expression corresponding to the angle Then fall into the expression corresponding to the angular range;
Range information is the distance between two human face characteristic points in step 3, if the distance corresponding to an expression apart from model In enclosing, then expression is then fallen into the expression corresponding to the distance range corresponding to the distance.
CN201710828671.0A 2017-09-14 2017-09-14 A kind of expression recognition method suitable for medical treatment and nursing Pending CN107704810A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710828671.0A CN107704810A (en) 2017-09-14 2017-09-14 A kind of expression recognition method suitable for medical treatment and nursing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710828671.0A CN107704810A (en) 2017-09-14 2017-09-14 A kind of expression recognition method suitable for medical treatment and nursing

Publications (1)

Publication Number Publication Date
CN107704810A true CN107704810A (en) 2018-02-16

Family

ID=61171744

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710828671.0A Pending CN107704810A (en) 2017-09-14 2017-09-14 A kind of expression recognition method suitable for medical treatment and nursing

Country Status (1)

Country Link
CN (1) CN107704810A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109409296A (en) * 2018-10-30 2019-03-01 河北工业大学 The video feeling recognition methods that facial expression recognition and speech emotion recognition are merged
CN109753950A (en) * 2019-02-11 2019-05-14 河北工业大学 Dynamic human face expression recognition method
CN111127521A (en) * 2019-10-25 2020-05-08 上海联影智能医疗科技有限公司 System and method for generating and tracking the shape of an object

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101777131A (en) * 2010-02-05 2010-07-14 西安电子科技大学 Method and device for identifying human face through double models
CN102831388A (en) * 2012-05-23 2012-12-19 上海交通大学 Method and system for detecting real-time characteristic point based on expanded active shape model
CN104036255A (en) * 2014-06-21 2014-09-10 电子科技大学 Facial expression recognition method
CN104376333A (en) * 2014-09-25 2015-02-25 电子科技大学 Facial expression recognition method based on random forests
CN104951743A (en) * 2015-03-04 2015-09-30 苏州大学 Active-shape-model-algorithm-based method for analyzing face expression

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101777131A (en) * 2010-02-05 2010-07-14 西安电子科技大学 Method and device for identifying human face through double models
CN102831388A (en) * 2012-05-23 2012-12-19 上海交通大学 Method and system for detecting real-time characteristic point based on expanded active shape model
CN104036255A (en) * 2014-06-21 2014-09-10 电子科技大学 Facial expression recognition method
CN104376333A (en) * 2014-09-25 2015-02-25 电子科技大学 Facial expression recognition method based on random forests
CN104951743A (en) * 2015-03-04 2015-09-30 苏州大学 Active-shape-model-algorithm-based method for analyzing face expression

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHEN CHIUNG HSIEH ET AL: "A Facial Expression Classification System Based on Active Shape Model and Support Vector Machine", 《2011 INTERNATIONAL SYMPOSIUM ON COMPUTER SCIENCE AND SOCIETY》 *
KE WANG ET AL: "Real-time facial expressions recognition system for service robot based-on ASM and SVMs", 《2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION》 *
YU ZHENGHONG ET AL: "Research of Facial Expression Recognition Based on ASM Model and RS-SVM", 《2014 FIFTH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS DESIGN AND ENGINEERING APPLICATIONS》 *
王林路等: "基于情感几何特征和支持向量机的人脸表情识别研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109409296A (en) * 2018-10-30 2019-03-01 河北工业大学 The video feeling recognition methods that facial expression recognition and speech emotion recognition are merged
CN109409296B (en) * 2018-10-30 2020-12-01 河北工业大学 Video emotion recognition method integrating facial expression recognition and voice emotion recognition
CN109753950A (en) * 2019-02-11 2019-05-14 河北工业大学 Dynamic human face expression recognition method
CN109753950B (en) * 2019-02-11 2020-08-04 河北工业大学 Dynamic facial expression recognition method
CN111127521A (en) * 2019-10-25 2020-05-08 上海联影智能医疗科技有限公司 System and method for generating and tracking the shape of an object
CN111127521B (en) * 2019-10-25 2024-03-01 上海联影智能医疗科技有限公司 System and method for generating and tracking shape of target

Similar Documents

Publication Publication Date Title
CN109325398B (en) Human face attribute analysis method based on transfer learning
CN104850825B (en) A kind of facial image face value calculating method based on convolutional neural networks
CN105447473B (en) A kind of any attitude facial expression recognizing method based on PCANet-CNN
Lemaire et al. Fully automatic 3D facial expression recognition using differential mean curvature maps and histograms of oriented gradients
CN100382751C (en) Canthus and pupil location method based on VPP and improved SUSAN
WO2020103417A1 (en) Bmi evaluation method and device, and computer readable storage medium
CN108614999B (en) Eye opening and closing state detection method based on deep learning
CN107704810A (en) A kind of expression recognition method suitable for medical treatment and nursing
CN112232184B (en) Multi-angle face recognition method based on deep learning and space conversion network
CN101763503A (en) Face recognition method of attitude robust
JP2018055470A (en) Facial expression recognition method, facial expression recognition apparatus, computer program, and advertisement management system
CN103902992B (en) Human face recognition method
CN111985532B (en) Scene-level context-aware emotion recognition deep network method
CN1687957A (en) Man face characteristic point positioning method of combining local searching and movable appearance model
CN106980819A (en) Similarity judgement system based on human face five-sense-organ
CN1776711A (en) Method for searching new position of feature point using support vector processor multiclass classifier
CN107392151A (en) Face image various dimensions emotion judgement system and method based on neutral net
CN104794441B (en) Human face characteristic positioning method based on active shape model and POEM texture models under complex background
CN106897706B (en) A kind of Emotion identification device
CN109409298A (en) A kind of Eye-controlling focus method based on video processing
CN111611912B (en) Detection method for pedestrian head-falling abnormal behavior based on human body joint point
CN1936925A (en) Method for judging characteristic point place using Bayes network classification device image
CN108898623A (en) Method for tracking target and equipment
CN114758382A (en) Face AU detection model establishing method and application based on adaptive patch learning
TW201040846A (en) Face detection apparatus and face detection method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180216