CN101046959A - Identity identification method based on lid speech characteristic - Google Patents

Identity identification method based on lid speech characteristic Download PDF

Info

Publication number
CN101046959A
CN101046959A CNA2007100400038A CN200710040003A CN101046959A CN 101046959 A CN101046959 A CN 101046959A CN A2007100400038 A CNA2007100400038 A CN A2007100400038A CN 200710040003 A CN200710040003 A CN 200710040003A CN 101046959 A CN101046959 A CN 101046959A
Authority
CN
China
Prior art keywords
lip
model
speech characteristic
sigma
segment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CNA2007100400038A
Other languages
Chinese (zh)
Inventor
王士林
刘功申
林祥
李翔
李生红
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Original Assignee
Shanghai Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University filed Critical Shanghai Jiaotong University
Priority to CNA2007100400038A priority Critical patent/CN101046959A/en
Publication of CN101046959A publication Critical patent/CN101046959A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Collating Specific Patterns (AREA)

Abstract

The present invention relates to a speechmaker identity authentication method based on lip movement in the field of biological characteristics and mode identification technology. Said method includes the following steps: inputting lip image sequence, dividing lip region based on improved fuzzy aggregation algorithm, modeling lip region based on point drive and extracting various lip characteristics based on lip region model.

Description

Identity identifying method based on lid speech characteristic
Technical field
What the present invention relates to is a kind of method of field of information security technology, particularly a kind of identity identifying method based on lid speech characteristic.
Background technology
Along with the arriving of information society, the importance of security of system is day by day remarkable.Identity identifying technology is the prerequisite of most of security of system accurately.Traditional identity identifying method is often based on special article (as key, I.D., smart card or the like) or based on specific knowledge (as password, ID (identity number) card No. or the like).There is the problem of easy loss or damage in the former, and the latter then exists password to pass into silence easily or problem such as be cracked.The more important thing is that because the authentication information of these identity can duplicate and share often, system can't judge user's real identity by these information.Therefore, when using these traditional identity recognition methodss, above-mentioned limitation brings hidden danger can for inevitably the safety of system.Yet,, but be expected to overcome the above problems well if in identity authorization system, introduce biological information.
As everyone knows, everyone has the speech style of self uniqueness, even when telling about identical language, each one mode of telling about also is not quite similar.These each one exclusive locutions are mainly reflected in: one, acoustic information aspect---different separately sound spectrum; Two, visual information aspect---different separately lip motion process.Research shows that also lid speech characteristic is the same with features such as fingerprint, looks, voice, has certain identity distinguishing ability, can be regarded as a kind of new bio feature in identification field.
Carry out aspect the authentication using lid speech characteristic, more influential in the world achievement in research has: people such as the Luettin of electronic engineering of Britain Sheffield university propose with lip shape and oral cavity interior intensity distributed intelligence as feature, and use hidden Markov model (Hidden Markov Model, HMM) identifying algorithm of differentiating, this method can reach 92% discrimination nearly in the experiment of carrying out for 12 users; The identity identifying method that the people such as Wark of Queensland ,Australia University of Science and Technology propose based on the lip feature extraction, wherein use principal component analysis (PCA) (PCA) and linear discriminant analysis (LDA) to extract the lip feature, and used gauss hybrid models (GMM) to finish identification and certification work.This method can reach 90% discrimination nearly in the experiment of carrying out for 37 users.Above-mentioned technology has reflected the feasibility of using lip motion to carry out identification to a certain extent, for the research of this respect is from now on had laid a good foundation.
Find through literature search prior art, Kanak etc. are at 2003 international voice sound signal processing conferences (Proceedings of IEEE International Conference on Acoustics, Speech and SignalProcessing, vol.3, " uniting the speaker identity authentication techniques of looking Audio Processing " of delivering pp.561-564) (Joint audio-video processing for biometric speaker identification), proposing a kind of employing intrinsic lip (eigenlip) in this article is feature, and the individual layer hidden Markov model is the speaker identity authentication method based on lid speech characteristic of classification authentication mechanism.Its deficiency is: 1) the extraction accuracy deficiency of lid speech characteristic; 2) lacking the lid speech characteristic that the identity resolving ability is arranged expresses; 3) lack the ID authentication mechanism that is fit to lid speech characteristic.
Summary of the invention
The objective of the invention is to overcome deficiency of the prior art, propose a kind of identity identifying method, achieve identification, judge the consistance of itself and legal identity for speaker identity based on lid speech characteristic, when obtaining high discrimination, guaranteed higher processing speed.
The present invention is achieved by the following technical solutions: at first carry out the automatic extraction of lid speech characteristic by lip region dividing method and lip modeling method, carry out the analysis of various lid speech characteristic identity resolving abilitys by the method for mutual information assessment then, from various lid speech characteristics, choose optimum lid speech characteristic combination, at last, according to the characteristics of lid speech characteristic, adopt multi-level hidden Markov model to carry out the identification and the authentication of speaker identity.
Described lip region dividing method specifically is a kind of improved fuzzy Classified Algorithms Applied, and its objective function is:
J m = J CLR + J SPA = J CLR , OBJ + J CLR , BKG + J SPA , OBJ + J SPA , BKG
= Σ r = 1 N Σ s = 1 M u 0 , r , s m d 0 , r , s 2 + Σ r = 1 N Σ s = 1 M u 1 , r , s m d 1 , r , s 2 + Σ r = 1 N Σ s = 1 M u 0 , r , s m g OBJ ( r , s ) + Σ r = 1 N Σ s = 1 M u 1 , r , s m g BKG ( r , s )
And membership function is obeyed Σ i = 0 1 u i , r , s = 1 . Wherein, the subscript CLR of objective function J, SPA, OBJ and BKG represent color part, space segment, lip target part and the background parts of objective function respectively; u I, r, sRepresent coordinate for (r, pixel s) is for the degree of membership of i class; d I, r, sRepresenting coordinate is (r, the color distance between pixel s) and the i class color center; g OBJAnd g BKGDenotation coordination is (r, the space length between pixel s) and the center, i space-like respectively.Can obtain the degree of membership of all pixels through iteration optimization, promptly be considered to the lip pixel greater than 0.5 picture element for the target class degree of membership about the lip class.
Described lip region modeling method is meant: a lip geometric model of forming by three quafric curves in 16 o'clock, and in the process of model optimization, the iteration optimization process of " point drives " of employing.In each iterative loop, optimizing process is only adjusted 16 model silhouette points and is made and wherein comprise lip pixel as much as possible and the least possible non-lip pixel.
Described method by the mutual information assessment is carried out the analysis of various lid speech characteristic identity resolving abilitys, adopts following formula to realize: I ( X ; Y ) = Σ XY P ( xy ) log 2 P ( x | y ) P ( x ) , I (X wherein; Y) represent feature set X and teller to gather mutual information between the Y, P (xy), P (x|y) and P (x) represent the prior probability of joint probability, conditional probability and the x of x and y respectively.Because often take the mutual information of discrete form to express (promptly calculating discrete probability by histogrammic form) in signature analysis, determining of histogrammic quantization method is an important ring of impact analysis accuracy.The present invention adopts the histogram of uniform quantization to calculate discrete probability, and wherein histogrammic dimension calculates by the following method: histogrammic dimension calculates by the following method: log 2N+1 (if feature Gaussian distribution), otherwise equal
Figure A20071004000300072
N is the training sample sum.
Described characteristics according to lid speech characteristic adopt multi-level hidden Markov model to carry out the identification and the authentication of speaker identity, and concrete steps are as follows:
1. the basic exercise sequence is cut apart: the basic exercise sequence of so-called lip refers to that two ends change the violent lip motion segment of smooth intermediate change.Note i characteristics combination constantly is f Lip, i, then i+1 characteristics combination constantly is f Lip, i+1If, Δ=‖ f Lip, i+1-f Lip, i‖>T (T is default thresholding), then time i is considered to a lip mutation time point, just the starting point of basic exercise sequence (or terminating point).After having determined all lip mutation time points, whole signal language sequence can be divided into several basic exercise sequences.
2. customer group hidden Markov model (Hidden Markov Model, HMM) foundation in storehouse: the basic exercise sequence of telling about certain segment in the signal language with all users in the training set is a training sample, the optimal characteristics collection is a training characteristics, HMM is a disaggregated model continuously, sets up the HMM model bank of customer group about all basic exercise sequences in the signal language; Wherein, the HMM model parameter is set to: 6 states are model from left to right, and Bao Mu-Wei Erqi (Baum-Welch) algorithm is a training algorithm, and Viterbi (Viterbi) algorithm is a recognizer;
3. specific teller's " lip print model " foundation:, analyze the matching degree of this segment in specific teller and the customer group HMM storehouse for all basic exercise sequences in the signal language.If matching degree higher (greater than predefined threshold value) illustrates that the user tells about this segment and do not have uniqueness.Otherwise, illustrating that the user tells about segment and has suitable characteristics, the basic exercise sequence of telling about this segment with this user is taked and the identical model parameter in customer group HMM storehouse as training set, sets up " lip print model " with specific teller's characteristics.
4. teller's identification and authentication: for teller's identification, promptly from customer group, discern teller's identity, the present invention adopts the matching degree of all users " lip print model " in analytical test speech segment and the sample storehouse, therefrom choose matching degree the highest be judged as the teller.For teller's authentication, the present invention is mated by analytical test speech segment and specific teller's " lip print model ", passes through if matching degree surpasses threshold value then authenticates, otherwise then regards as the disabled user.
The technical solution used in the present invention is better than traditional identity identifying method based on lid speech characteristic on performance, its reason is: first, the automatic extractive technique of lid speech characteristic in the scheme can obtain lip region information more accurately by taking into account color and space distribution information; The second, can obtain more to have the characteristics combination of identity resolving ability by the method for mutual information assessment; The 3rd, the framework of multi-level hidden Markov model meets the characteristics of lid speech characteristic, and higher classification accuracy can be provided.
The present invention is directed to the concrete characteristics of lip image sequence, proposed to be fit to extraction, analysis and the sorting technique of such biological characteristic, and realized the authentication of speaker identity.The present invention has passed through performance test, and table with test results understands that the present invention has higher accuracy for speaker identity identification and authentication, i.e. identification rate is 98.07%, and error rates such as authentication are 2.31%.Simultaneously, the present invention has guaranteed higher processing speed when obtaining high discrimination, have broad application prospects.
Description of drawings
Fig. 1 the inventive method process flow diagram
Wherein: (a) training flow process; (b) identifying procedure
The extraction method process flow diagram of Fig. 2 lid speech characteristic of the present invention
Fig. 3 lip segmentation result synoptic diagram
Wherein: (a) the former figure of lip; (b) lip segmentation result
Fig. 4 lip geometric model synoptic diagram
Fig. 5 the present invention is fit to the identity identifying and authenticating method flow diagram of lid speech characteristic characteristics
Wherein: (a) training flow process; (b) identifying procedure
Embodiment
Below in conjunction with accompanying drawing embodiments of the invention are elaborated: present embodiment is being to implement under the prerequisite with the technical solution of the present invention, provided detailed embodiment and concrete operating process, but protection scope of the present invention is not limited to following embodiment.
As shown in Figure 1, present embodiment at first carries out the automatic extraction of lid speech characteristic, mainly comprises: the 1) geometric properties of lip, the length and the width information of lip region are promptly described; 2) contour feature of lip is promptly described the information of lip contour shape; 3) lip region internal feature is promptly described information such as tooth that lip inside may comprise, tongue; 4) spectrum signature of lip is promptly described the feature of lip motion at frequency domain.Then by (Mutual Information, method MI) are made accurate, deep assessment for the identity resolving ability that above-mentioned each category feature has based on mutual information.At last, finish identification and authentication by multi-level hidden Markov model for speaker identity.
As shown in Figure 2, the automatic extraction flow process of lid speech characteristic.It mainly comprises following three parts: lip region is cut apart, lip region modeling and various lid speech characteristic extract.Each part concrete steps is as follows:
● lip region is cut apart:
Owing to the color distance of rgb space and the lack of uniformity of human eye vision, therefore at first the RGB color space is converted into the CIE-LABUV color space of relative equilibrium, so the colouring information of each pixel can be expressed as { L, a, b, u, v} in the image I.Then, whole lip region cutting procedure can be described as two class classification problems for all pixels in the image, promptly is divided into lip pixel class and non-lip pixel class.At last, take into account color and spatial information, the inventor adopts and improves the classification that fuzzy Classified Algorithms Applied is finished pixel.Concrete objective function is:
J m = J CLR + J SPA = J CLR , OBJ + J CLR , BKG + J SPA , OBJ + J SPA , BKG
= Σ r = 1 N Σ s = 1 M u 0 , r , s m d 0 , r , s 2 + Σ r = 1 N Σ s = 1 M u 1 , r , s m d 1 , r , s 2 + Σ r = 1 N Σ s = 1 M u 0 , r , s m g OBJ ( r , s ) + Σ r = 1 N Σ s = 1 M u 1 , r , s m g BKG ( r , s )
And membership function is obeyed Σ i = 0 1 u i , r , s = 1 . Wherein, the subscript CLR of objective function J, SPA, OBJ and BKG represent color part, space segment, lip target part and the background parts of objective function respectively; u I, r, sRepresent coordinate for (r, pixel s) is for the degree of membership of i class; d I, r, sRepresenting coordinate is (r, the color distance between pixel s) and the i class color center; g OBJAnd g BKGDenotation coordination is (r, the space length between pixel s) and the center, i space-like respectively.
Can obtain the degree of membership of all pixels through iteration optimization (iterative optimization), promptly be considered to the lip pixel greater than 0.5 picture element for the target class degree of membership about the lip class.The lip segmentation result as shown in Figure 3.
● the lip region modeling:
If adopt the pixel mode to describe lip region, have the too high deficiency of intrinsic dimensionality.So need to adopt lip model description lip pixel, thereby reduce dimension, classification effectiveness and accuracy rate after improving.The present invention adopts 16 lip geometric models (as shown in Figure 4), and this model is made up of three quafric curves, has both reduced the low complex degree of model, guarantees the rationality of lip model again.
In the process of model optimization, the iteration optimization process of " point drives " that the present invention adopts.In each iterative loop, optimizing process is only adjusted 16 model silhouette points and is made and wherein comprise lip pixel as much as possible and the least possible non-lip pixel.
● various lid speech characteristics extract:
From the lip model, can obtain following feature easily: lip geometric properties (f Geo); Lip shape feature (f Shape, comprise the dynamic shape aspect of model etc.); Lip textural characteristics (f Texture, comprise the dynamic appearance aspect of model etc.); Lip internal feature (f Inner, comprise tooth, tongue feature etc.); Lip spectrum signature (f Freq, comprise various frequency domain characters).
In view of the identity distinguishing ability of various lip features can obtain with different tellers' mutual information by calculating this feature, the present invention adopts the mutual information analytical approach to analyze the identity resolving ability of various features.The mutual information computing formula is as follows: I ( X ; Y ) = Σ XY P ( xy ) log 2 P ( x | y ) P ( x ) . Because often take the mutual information of discrete form to express (promptly calculating discrete probability by histogrammic form) in signature analysis, determining of histogrammic quantization method is an important ring of impact analysis accuracy.The present invention adopts the histogram of uniform quantization to calculate discrete probability, and wherein histogrammic dimension calculates by the following method: log 2N+1 (if feature Gaussian distribution), otherwise equal
Figure A20071004000300102
N is the training sample sum.
By the mutual information analysis, the inventor chooses following feature construction optimal characteristics combination, owing to wherein comprised the abundantest identity resolving ability: { f Geo, f Inner, f Texture, i Ca(f wherein Texture, icaThe lip textural characteristics that expression obtains by independent component analysis (ICA)).
As shown in Figure 5, be fit to the speaker identity identification authentication method based on multi-level hidden Markov model of lid speech characteristic characteristics, specific as follows:
1. the basic exercise sequence is cut apart: the basic exercise sequence of so-called lip refers to that two ends change the violent lip motion segment of smooth intermediate change.Note i characteristics combination constantly is f Lip, i, then i+1 characteristics combination constantly is f Lip, i+1If, Δ=‖ f Lip, i+1-f Lip, i‖>T, then time i is considered to a lip mutation time point, the just starting point of basic exercise sequence (or terminating point).After having determined all lip mutation time points, whole signal language sequence can be divided into several basic exercise sequences.
2. customer group hidden Markov model (Hidden Markov Model, HMM) foundation in storehouse: the basic exercise sequence of telling about certain segment in the signal language with all users in the training set is a training sample, the optimal characteristics collection is a training characteristics, HMM is a disaggregated model continuously, sets up the HMM model bank of customer group about all basic exercise sequences in the signal language; Wherein, the HMM model parameter is set to: 6 states are model from left to right, and Bao Mu-Wei Erqi (Baum-Welch) algorithm is a training algorithm, and Viterbi (Viterbi) algorithm is a recognizer.
3. specific teller's " lip print model " foundation:, analyze the matching degree of this segment in specific teller and the customer group HMM storehouse for all basic exercise sequences in the signal language.If matching degree higher (greater than predefined threshold value) illustrates that the user tells about this segment and do not have uniqueness.Otherwise, illustrating that the user tells about segment and has suitable characteristics, the basic exercise sequence of telling about this segment with this user is taked and the identical model parameter in customer group HMM storehouse as training set, sets up " lip print model " with specific teller's characteristics.
4. teller's identification and authentication: for teller's identification, promptly from customer group, discern teller's identity, the present invention adopts the matching degree of all users " lip print model " in analytical test speech segment and the sample storehouse, therefrom choose matching degree the highest be judged as the teller.For teller's authentication, the present invention is mated by analytical test speech segment and specific teller's " lip print model ", passes through if matching degree surpasses threshold value then authenticates, otherwise then regards as the disabled user.
In order to implement concrete grammar of the present invention, the inventor has designed and Implemented the identity authorization system based on lid speech characteristic.By for 40 tellers (29 male 11 woman), tell about the experiment and the assessment in the lip image sequence storehouse (comprise about 36,000 open one's mouth lip image) of particular hint language.Native system has higher accuracy for the authentication of speaker identity, i.e. identification rate is 98.07%, and error rates such as authentication are 2.31%.

Claims (6)

1, a kind of identity identifying method based on lid speech characteristic, it is characterized in that, at first carry out the automatic extraction of lid speech characteristic by lip region dividing method and lip modeling method, carry out the analysis of various lid speech characteristic identity resolving abilitys by the method for mutual information assessment then, from various lid speech characteristics, choose optimum lid speech characteristic combination, at last, according to the characteristics of lid speech characteristic, adopt multi-level hidden Markov model to carry out the identification and the authentication of speaker identity.
2, the identity identifying method based on lid speech characteristic according to claim 1 is characterized in that, described lip region dividing method, and its objective function is:
J m = J CLR + J SPA = J CLR , OBJ + J CLR , BKG + J SPA , OBJ + J SPA , BKG
= Σ r = 1 N Σ s = 1 M u 0 , r , s m d 0 , r , s 2 + Σ r = 1 N Σ s = 1 M u 1 , r , s m d 1 , r , s 2 + Σ r = 1 N Σ s = 1 M u 0 , r , s m g OBJ ( r , s ) + Σ r = 1 N Σ s = 1 M u 1 , r , s m g BKG ( r , s )
And membership function is obeyed Σ i = 1 1 u i , r , s = 1 , Wherein, the subscript CLR of objective function J, SPA, OBJ and BKG represent color part, space segment, lip target part and the background parts of objective function respectively; u I, r, sRepresent coordinate for (r, pixel s) is for the degree of membership of i class; d I, r, sRepresenting coordinate is (r, the color distance between pixel s) and the i class color center; g OBJAnd g BKGDenotation coordination is (r, the space length between pixel s) and the center, i space-like respectively; Obtain the degree of membership of all pixels through iteration optimization, promptly be considered to the lip pixel greater than 0.5 picture element for the target class degree of membership about the lip class.
3, the identity identifying method based on lid speech characteristic according to claim 1, it is characterized in that, described lip region modeling method, be meant: a lip geometric model of forming by three quafric curves in 16 o'clock, in the process of model optimization, the iteration optimization process of " point drives " of adopting, in each iterative loop, optimizing process is only adjusted 16 model silhouette points and is made and wherein comprise lip pixel as much as possible and the least possible non-lip pixel.
4, the identity identifying method based on lid speech characteristic according to claim 1 is characterized in that, described lid speech characteristic extracts automatically, extracts following feature: the lip geometric properties; The lip shape feature that comprises the dynamic shape aspect of model; The lip textural characteristics that comprises the dynamic appearance aspect of model; The lip internal feature that comprises tooth, tongue feature; Various lip spectrum signatures.
5, the identity identifying method based on lid speech characteristic according to claim 1 is characterized in that, described method by the mutual information assessment is carried out the analysis of various lid speech characteristic identity resolving abilitys, adopts following formula to realize:
I ( X ; Y ) = Σ XY P ( xy ) log 2 P ( x | y ) P ( x ) , I (X wherein; Y) represent feature set X and teller to gather mutual information between the Y, P (xy), P (x|y) and P (x) represent the prior probability of joint probability, conditional probability and the x of x and y respectively; Adopt the histogram of uniform quantization to calculate discrete probability, wherein histogrammic dimension calculates by the following method: if be characterized as Gaussian distribution, then dimension is log 2N+1, otherwise equal log 2 N + 1 + log 2 ( 1 + κ · N 6 ) , N is the training sample sum.
6, the identity identifying method based on lid speech characteristic according to claim 1 is characterized in that, described characteristics according to lid speech characteristic adopt multi-level hidden Markov model to carry out the identification and the authentication of speaker identity, and concrete steps are as follows:
1. the basic exercise sequence is cut apart: the basic exercise sequence of so-called lip, refer to that two ends change the violent lip motion segment of smooth intermediate change, and note i characteristics combination constantly is f Lip, i, then i+1 characteristics combination constantly is f Lip, i+1If, Δ=‖ f Lip, i+1-f Lip, i‖>T, then time i is considered to a lip mutation time point, i.e. the starting point of basic exercise sequence or terminating point, after having determined all lip mutation time points, whole signal language sequence can be divided into several basic exercise sequences;
2. the foundation in customer group hidden Markov model storehouse: the basic exercise sequence of telling about certain segment in the signal language with all users in the training set is a training sample, the optimal characteristics collection is a training characteristics, hidden Markov model is a disaggregated model continuously, sets up the hidden Markov model storehouse of customer group about all basic exercise sequences in the signal language; Wherein, the hidden Markov model parameter is set to: 6 states are model from left to right, and Bao Mu-Wei Erqi algorithm is a training algorithm, and viterbi algorithm is a recognizer;
3. specific teller's " lip print model " foundation: for all basic exercise sequences in the signal language, analyze the matching degree of this segment in specific teller and the customer group hidden Markov model storehouse, if matching degree is greater than predefined threshold value, illustrate that the user tells about this segment and do not have uniqueness, otherwise, illustrate that the user tells about segment and has characteristics, tell about the basic exercise sequence of this segment as training set with this user, take and the identical model parameter in customer group hidden Markov model storehouse, set up " lip print model " with specific teller's characteristics;
4. teller's identification and authentication: for teller's identification, adopt the matching degree of all users " lip print model " in analytical test speech segment and the sample storehouse, therefrom choose matching degree the highest be judged as the teller; For teller's authentication, " lip print model " by analytical test speech segment and specific teller mates, and passes through if matching degree surpasses threshold value then authenticates, otherwise then regards as the disabled user.
CNA2007100400038A 2007-04-26 2007-04-26 Identity identification method based on lid speech characteristic Pending CN101046959A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CNA2007100400038A CN101046959A (en) 2007-04-26 2007-04-26 Identity identification method based on lid speech characteristic

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CNA2007100400038A CN101046959A (en) 2007-04-26 2007-04-26 Identity identification method based on lid speech characteristic

Publications (1)

Publication Number Publication Date
CN101046959A true CN101046959A (en) 2007-10-03

Family

ID=38771506

Family Applications (1)

Application Number Title Priority Date Filing Date
CNA2007100400038A Pending CN101046959A (en) 2007-04-26 2007-04-26 Identity identification method based on lid speech characteristic

Country Status (1)

Country Link
CN (1) CN101046959A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102013103A (en) * 2010-12-03 2011-04-13 上海交通大学 Method for dynamically tracking lip in real time
CN102542280A (en) * 2010-12-26 2012-07-04 上海量明科技发展有限公司 Recognition method and system aiming at different lip-language mouth shapes with same content
CN103324918A (en) * 2013-06-25 2013-09-25 浙江中烟工业有限责任公司 Identity authentication method with face identification and lip identification matched
CN104361276A (en) * 2014-11-18 2015-02-18 新开普电子股份有限公司 Multi-mode biometric authentication method and multi-mode biometric authentication system
CN104680144A (en) * 2015-03-02 2015-06-03 华为技术有限公司 Lip language recognition method and device based on projection extreme learning machine
CN104834900A (en) * 2015-04-15 2015-08-12 常州飞寻视讯信息科技有限公司 Method and system for vivo detection in combination with acoustic image signal
CN105787428A (en) * 2016-01-08 2016-07-20 上海交通大学 Method for lip feature-based identity authentication based on sparse coding
CN106295501A (en) * 2016-07-22 2017-01-04 中国科学院自动化研究所 The degree of depth based on lip movement study personal identification method
CN106529379A (en) * 2015-09-15 2017-03-22 阿里巴巴集团控股有限公司 Method and device for recognizing living body
CN106778496A (en) * 2016-11-22 2017-05-31 重庆中科云丛科技有限公司 Biopsy method and device
CN108351972A (en) * 2015-10-28 2018-07-31 皇家飞利浦有限公司 Device and method for the anatomic shape for indicating biology
CN111986674A (en) * 2020-08-13 2020-11-24 广州仿真机器人有限公司 Intelligent voice recognition method based on three-level feature acquisition
CN112053160A (en) * 2020-09-03 2020-12-08 中国银行股份有限公司 Intelligent bracelet for lip language recognition, lip language recognition system and method

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102013103A (en) * 2010-12-03 2011-04-13 上海交通大学 Method for dynamically tracking lip in real time
CN102013103B (en) * 2010-12-03 2013-04-03 上海交通大学 Method for dynamically tracking lip in real time
CN102542280A (en) * 2010-12-26 2012-07-04 上海量明科技发展有限公司 Recognition method and system aiming at different lip-language mouth shapes with same content
CN102542280B (en) * 2010-12-26 2016-09-28 上海量明科技发展有限公司 The recognition methods of the different lip reading shape of the mouth as one speaks and system for same content
CN103324918A (en) * 2013-06-25 2013-09-25 浙江中烟工业有限责任公司 Identity authentication method with face identification and lip identification matched
CN103324918B (en) * 2013-06-25 2016-04-27 浙江中烟工业有限责任公司 The identity identifying method that a kind of recognition of face matches with lipreading recognition
CN104361276A (en) * 2014-11-18 2015-02-18 新开普电子股份有限公司 Multi-mode biometric authentication method and multi-mode biometric authentication system
CN104361276B (en) * 2014-11-18 2017-07-18 新开普电子股份有限公司 A kind of multi-modal biological characteristic identity identifying method and system
CN104680144A (en) * 2015-03-02 2015-06-03 华为技术有限公司 Lip language recognition method and device based on projection extreme learning machine
CN104834900A (en) * 2015-04-15 2015-08-12 常州飞寻视讯信息科技有限公司 Method and system for vivo detection in combination with acoustic image signal
CN104834900B (en) * 2015-04-15 2017-12-19 常州飞寻视讯信息科技有限公司 A kind of method and system combined audio-visual signal and carry out In vivo detection
CN106529379A (en) * 2015-09-15 2017-03-22 阿里巴巴集团控股有限公司 Method and device for recognizing living body
CN108351972A (en) * 2015-10-28 2018-07-31 皇家飞利浦有限公司 Device and method for the anatomic shape for indicating biology
CN105787428A (en) * 2016-01-08 2016-07-20 上海交通大学 Method for lip feature-based identity authentication based on sparse coding
CN106295501A (en) * 2016-07-22 2017-01-04 中国科学院自动化研究所 The degree of depth based on lip movement study personal identification method
CN106778496A (en) * 2016-11-22 2017-05-31 重庆中科云丛科技有限公司 Biopsy method and device
CN111986674A (en) * 2020-08-13 2020-11-24 广州仿真机器人有限公司 Intelligent voice recognition method based on three-level feature acquisition
CN112053160A (en) * 2020-09-03 2020-12-08 中国银行股份有限公司 Intelligent bracelet for lip language recognition, lip language recognition system and method
CN112053160B (en) * 2020-09-03 2024-04-23 中国银行股份有限公司 Intelligent bracelet for lip language identification, lip language identification system and method

Similar Documents

Publication Publication Date Title
CN101046959A (en) Identity identification method based on lid speech characteristic
CN107729835B (en) Expression recognition method based on fusion of traditional features of face key point region and face global depth features
CN106127156A (en) Robot interactive method based on vocal print and recognition of face
Soltane et al. Face and speech based multi-modal biometric authentication
Fallah et al. A new online signature verification system based on combining Mellin transform, MFCC and neural network
Pirlo et al. Multidomain verification of dynamic signatures using local stability analysis
Daramola et al. Offline signature recognition using hidden markov model (HMM)
CN103092329A (en) Lip reading technology based lip language input method
CN108256307B (en) Hybrid enhanced intelligent cognitive method of intelligent business travel motor home
Kryszczuk et al. Reliability-based decision fusion in multimodal biometric verification systems
CN103927532B (en) Person's handwriting method for registering based on stroke feature
CN108681737B (en) Method for extracting image features under complex illumination
CN103218624A (en) Recognition method and recognition device based on biological characteristics
CN101178767A (en) Recognizing layer amalgamation for human face and iris mixed recognition
CN110263726A (en) A kind of finger vein identification method and device based on depth correlation feature learning
Hu et al. Fingerprint classification based on genetic programming
CN106980845B (en) Face key point positioning method based on structured modeling
Guo A hidden markov model fingerprint matching approach
Soltane et al. Soft decision level fusion approach to a combined behavioral speech-signature biometrics verification
CN116884415A (en) Voiceprint recognition method based on DV-Softmax loss function
CN116612542A (en) Multi-mode biological feature consistency-based audio and video character recognition method and system
CN115795394A (en) Biological feature fusion identity recognition method for hierarchical multi-modal and advanced incremental learning
Lavanya et al. Performance evaluation of fingerprint identification based on DCT and DWT using multiple matching techniques
CN111310546B (en) Method for extracting and authenticating writing rhythm characteristics in online handwriting authentication
Jung et al. Live-scanned fingerprint classification with markov models modified by GA

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C12 Rejection of a patent application after its publication
RJ01 Rejection of invention patent application after publication

Open date: 20071003