CN103280219A - Android platform-based voiceprint recognition method - Google Patents

Android platform-based voiceprint recognition method Download PDF

Info

Publication number
CN103280219A
CN103280219A CN2013101828255A CN201310182825A CN103280219A CN 103280219 A CN103280219 A CN 103280219A CN 2013101828255 A CN2013101828255 A CN 2013101828255A CN 201310182825 A CN201310182825 A CN 201310182825A CN 103280219 A CN103280219 A CN 103280219A
Authority
CN
China
Prior art keywords
mel
voiceprint
voice
feature extraction
frequency
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2013101828255A
Other languages
Chinese (zh)
Inventor
刘海亮
蒋德东
曹彩凤
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Yat Sen University
National Sun Yat Sen University
Original Assignee
National Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National Sun Yat Sen University filed Critical National Sun Yat Sen University
Priority to CN2013101828255A priority Critical patent/CN103280219A/en
Publication of CN103280219A publication Critical patent/CN103280219A/en
Pending legal-status Critical Current

Links

Images

Abstract

The invention discloses an android platform-based voiceprint recognition method, which comprises the following steps: inputting a voice password by a user through a mobile phone; acquiring voiceprint characteristic parameters by using a Mel-cepstrum technology; inputting the same password when the user logs in the system for the second time; meanwhile, acquiring voiceprint information to be recognized through characteristic extraction; finally comparing the obtained confidence coefficient with a critical value, getting the output result of being me when the confidence coefficient is less than the critical value, and otherwise getting the result of being not me so as to finish a certification process. By implementing the method, a voiceprint recognition system is provided. According to the system, the operation is more convenient and quicker and the actual requirements of uses can be met better.

Description

A kind of voiceprint authentication method based on the Android platform
Technical field
The present invention relates to technique computes machine field, be specifically related to a kind of voiceprint authentication method based on the Android platform.
Background technology
Along with the development of intelligent mobile phone system, the application and development of smart mobile phone has become the focus that people pay close attention to.Along with Android mobile phone fashionable in the whole world, will certainly drive increasing sharply of Android application software quantity.The application software of Android mobile phone has related to fields of society at present.The exploitation of Android mobile phone aspect voice now begun, and will provide better to the user, serve more efficiently.Therefore, the vocal print Study on Technology is bound to become the Android mobile phone in the future in important research direction of voice development field.The voiceprint groundwork is the characteristic information that obtains in speaker's the voice, and these information can comprise speaker's psychologic status, behavior disposition etc.By analyzing the real identity that these information can confirm the speaker.Therefore, need carry out feature extraction to speaker's voice, obtain characteristic parameter.The characteristic parameter that speaker's front and back voice are obtained carries out then, and then judges speaker's true identity.It substantially all is to adopt computer as operation platform that traditional vocal print technology is used.Because computer has shortcomings such as the inconvenience of carrying, and has brought limitation to application.And will adopt smart mobile phone as carrier based on the voiceprint system of Android platform, and use conveniently like this, the user only need just can realize some authentication operations by voice, has saved the time and operation is more convenient.
Linear prediction cepstrum coefficient (LPCC) is the characteristic parameter extraction method that generally adopts at present.Usually be applied in conjunction and the continuous speech recognition, usually research has adopted three kinds of main linear predictions to deduce parameter, be linear prediction reflection coefficient, line spectrum pair coefficient and linear prediction cepstrum coefficient, and the application in the conjunction speech recognition system, and carry out Computer Simulation.
Artificial neural network algorithm is exactly the second way of anthropomorphic dummy's thinking.This is a nonlinear kinetics system, and its characteristic is distributed storage and the concurrent collaborative processing of information.Though single neuronic structure is extremely simple, function is limited, and the behavior that the network system that a large amount of neurons constitute can realize is extremely colourful.Therefore, this ability to model to the human cognitive system of neural network has determined to be applied to speech processes the incomparable advantage of additive method.
Present voiceprint mainly with computing machine as the operation carrier, most of linear prediction cepstrum coefficient (LPCC) method that adopts of vocal print feature extraction obtains characteristic parameter, this Feature Extraction Technology obtains parameter and is not suitable for voiceprint usually.The vocal print parameter matching adopts artificial neural network algorithm and Hidden Markov algorithm etc. mostly.These algorithm more complicated may influence the total system performance if operate on the intelligent mobile phone platform.
The major defect of existing voiceprint has: one, authentication rate is generally on the low side, and needing repeatedly usually, input could realize authentication.Two, running environment is had relatively high expectations, major part all requires just can normally move on computers.Software and hardware to computing machine all has requirement.Three, authentication result very easily is subjected to the interference of external environment, a little less than the anti-interference.Extraneous slight noise may make the result change.This also is the difficult problem that present vocal print Verification System generally runs into.
Summary of the invention
Purpose of the present invention provides the voiceprint system based on the Android platform, by the improvement to the voiceprint matching algorithm, has reduced the matching algorithm time complexity, has improved system running speed, makes system's operation more quick.
The invention provides a kind of voiceprint authentication method based on the Android platform, comprise the steps:
The user is by mobile phone input voice password; Adopt Mel cepstrum technology to obtain the vocal print characteristic parameter; When the user lands this system for the second time, the input same challenge, obtain voice voiceprint to be certified by feature extraction simultaneously, at last, compare according to the degree of confidence score and the critical value that obtain, the output result be " in person " when less than critical value, otherwise the gained result is " in person non-", thereby finishes verification process.
Described Mel cepstrum technology comprises:
Voice signal carries out Fast Fourier Transform (FFT) (FFT), and voice signal is transformed into frequency domain from time domain;
Before carrying out the V-belt pass filter, common frequencies to be converted into the Mel frequency usually;
The centering frequency, f ( m ) = ( N F s ) F mel - 1 [ f l + m F mel ( f h ) - F mel ( f l ) M + 1 ] , M=0 wherein, 1,2 ... 24; f h, f lBe respectively the highest and low-limit frequency of wave filter range of application; The width of representing FFT with N; F sBe sample frequency (F s=8000Hz); F Mel -1=700 (e H/1125-1) is F MelInverse function;
The H that wave filter is asked m(k) calculate spectrum energy Ek;
Cosine transform is carried out cosine transform with top gained logarithm energy by following formula, and L gets 12 in the formula, is 12 one digit number group Mel cepstrum parameter just thereby obtain length.
Describedly obtain voice voiceprint to be certified by feature extraction and comprise:
The voice password that the speaker is recorded carries out feature extraction
The user imports login voice password, obtains corresponding characteristic parameter by feature extraction equally;
Two groups of argument sequences that obtain are previously mated by the voiceprint matching algorithm.
By method provided by the invention, the present invention provides a kind of voiceprint system on based on the Android smart mobile phone, can make more convenient to operately, quick by this system, can satisfy user's actual demand better.
Description of drawings
In order to be illustrated more clearly in the embodiment of the invention or technical scheme of the prior art, to do to introduce simply to the accompanying drawing of required use in embodiment or the description of the Prior Art below, apparently, accompanying drawing in describing below only is some embodiments of the present invention, for those of ordinary skills, under the prerequisite of not paying creative work, can also obtain other accompanying drawing according to these accompanying drawings.
Fig. 1 is the voiceprint authentication method process flow diagram based on the android platform in the embodiment of the invention;
Fig. 2 is the Mel cepstrum techniqueflow chart in the embodiment of the invention;
Fig. 3 is frequency in the embodiment of the invention and the graph of relation of Mel frequency;
Fig. 4 is the Mel cepstrum coefficient synoptic diagram in the embodiment of the invention.
Embodiment
Below in conjunction with the accompanying drawing in the embodiment of the invention, the technical scheme in the embodiment of the invention is clearly and completely described, obviously, described embodiment only is the present invention's part embodiment, rather than whole embodiment.Based on the embodiment among the present invention, those of ordinary skills belong to the scope of protection of the invention not making all other embodiment that obtain under the creative work prerequisite.
The invention provides a kind of voiceprint system based on the Android platform, different with voiceprint system in the past is, this method with smart mobile phone as operation platform, on dynamical Matching Algorithm, improve simultaneously, reduced the time complexity of algorithm, made system be adapted at more moving above the smart mobile phone.
The operational flow diagram of system is as shown below.At first, the user is by mobile phone input voice password.Then, by feature extraction algorithm speech characteristic parameter is extracted.Native system adopts Mel cepstrum technology to obtain the vocal print characteristic parameter.When the user lands this system for the second time, the input same challenge, obtain voice voiceprint to be certified by feature extraction simultaneously, at last, compare according to the degree of confidence score and the critical value that obtain, the output result be " in person " when less than critical value, otherwise the gained result is " in person non-", thereby finishes verification process.Critical value herein is to obtain by testing repeatedly of back plane system.
The present invention mainly contains two tasks, and one is the extraction of vocal print characteristic parameter; Another is vocal print characteristic parameter coupling.Mel cepstrum technology is adopted in vocal print feature extraction of the present invention.The groundwork of this part is the parameter information that obtains can embody in the voice signal speaker's voice uniqueness.The feature of voiceprint must be the exclusive characteristic of speaker, also be and other people key distinction, and this feature remains unchanged for the speaker, tends towards stability, and can be described as the build-in attribute of people's voice.The concrete flow process of this technology as shown in Figure 2.
The first step: voice signal carries out Fast Fourier Transform (FFT) (FFT), and voice signal is transformed into frequency domain from time domain.
Second step: before carrying out the V-belt pass filter, common frequencies will be converted into the Mel frequency usually.Change by following formula.Mel (f)=2595*log10 (1+f/700) or mel (f)=1125*ln (1+f/700).
The 3rd step: the centering frequency, f ( m ) = ( N F s ) F mel - 1 [ f l + m F mel ( f h ) - F mel ( f l ) M + 1 ] , M=0 wherein, 1,2 ... 24; f h, f lBe respectively the highest and low-limit frequency of wave filter range of application; The width (N gets 512 herein) of representing FFT with N; F sBe sample frequency (F s=8000Hz); F Mel -1=700 (e H/1125-1) is F MelInverse function.
The centre frequency f (m) that obtains is brought into carries out computing in the top wave filter.This wave filter is to be made of 24 V-belt bandpass filters.Filter function is as follows:
H m ( k ) = 0 , k ≤ f ( m - 1 ) , k - f ( m - 1 ) f ( m ) - f ( m - 1 ) , f ( m - 1 ) ≤ k ≤ f ( m ) , f ( m + 1 ) - k f ( m + 1 ) - f ( m ) , f ( m ) ≤ k ≤ f ( m + 1 ) , 0 , k ≥ f ( m + 1 ) ,
The 4th step: the H that wave filter is asked m(k) calculate spectrum energy Ek.Computing formula is as follows:
Ek = ln ( Σ k M - 1 | X ( k ) | 2 H m ( k ) ) , ( 0 ≤ k ≤ M )
Wherein X (k) is tried to achieve frequency spectrum numerical value by the front Fast Fourier Transform (FFT).M represents the actual quantity of Mel frequency filter.The value of M is 24 in this problem.
The 5th step: cosine transform, top gained logarithm energy is carried out cosine transform by following formula, L gets 12 in the formula, is 12 one digit number group Mel cepstrum parameter just thereby obtain length.Formula is as follows.
C m = Σ k = 0 L - 1 cos ( m * ( k + 0.5 ) * π / M ) * Ek , m = 1,2,3 . . . , L
M in the formula represents to be with logical triangular filter number.Adopting the DCT conversion is that expectation can be carried out cepstrum.Finally obtain size and be 12 one-dimension array parameter.Gained Mel cepstrum coefficient as shown in Figure 4.
, by calling the SharedPreferences data storage method data are stored for the vocal print parameter that obtains.Key code in the system is as follows:
SharedPreferences settings=getSharedPreferences (PREFS_NAME, 0); // create object to store data
SharedPreferences.Editor editor=settings.edit (); // obtain the Editor object
SharedPreferences settings=getSharedPreferences (PREFS_NAME, 0); // obtain data
String ban=settings.getString (" moban ", " err "); // obtain template parameter
SharedPreferences?settings1=getSharedPreferences(PREFS_NAME,1);
String biao=settings1.getString (" mubiao ", " ron "); // obtain target component
SharedPreferences object and SQLite database are compared, and operate simplyr, have saved loaded down with trivial details operations such as establishment table.
Vocal print characteristic parameter coupling will be compared to two groups of voiceprint parameters that obtain, and degree of distortion and the degree of confidence score obtained between them compare.
The first step: the voice password that the speaker is recorded carries out feature extraction.The parameter of feature extraction gained is stored in the SD card as Template Information.This argument sequence can be expressed as R={R (1) with mathematic(al) representation, R (2) ... R (n) ..., R (M) }, the n in the formula represents the sequence number of speech frame, wherein n=1 represents that voice extract beginning, represents during n=M that leaching process finishes, thus speech frame add up to M.
Second step: the user imports login voice password, obtains corresponding characteristic parameter by feature extraction equally.Be T={T (1) with the mathematic(al) representation of the argument sequence that obtains, T (2) ... T (m) ..., T (L) }, m represents the sequence number of speech frame in the formula, the eigenvector of T (m) expression m frame.
The 3rd step: two groups of argument sequences that obtain are previously mated by the voiceprint matching algorithm.If the array length of the voiceprint of input voice and the array length of the resulting voice messaging of voice to be certified are M, the minimum distortion degree D that calculates between the two need expend very big cost so, needs to calculate the matrix of M*M.For above reason, can in the M*M matrix, choose some, these points can be expressed as
T (m 1) R (n 1), T (m 2) R (n 2) ...., T (m k) R (n k), satisfy m in the formula 1<m 2<...<m k,
n 1<n 2<...<n k, the bee-line expression formula of every some correspondence is
D(T(T(m 1)R(n 1))),D(T(T(m 2)R(n 2))),…,D(T(T(m k)R(n k)))
Therefore, the accumulation distance that can lose vector in the hope of minimum is
D [ T , R ] = Σ l = 1 j D [ ( T ( m 1 ) R ( m 1 ) ) , ( T ( m l + 1 ) R ( m l + 1 ) ) ]
Wherein j is selected counting.From top formula, be the minimum distortion degree summation of several minor matrixs as can be seen with the matrix abbreviation of M*M, calculated amount is reduced widely, reduced the time cost of system.At last, the vocal print parameter that obtains is carried out template matches and judge speaker ' s identity.
To sum up, the present invention provides a kind of voiceprint system on based on the Android smart mobile phone, can make more convenient to operately, quick by this system, can satisfy user's actual demand better.
One of ordinary skill in the art will appreciate that all or part of step in the whole bag of tricks of above-described embodiment is to instruct relevant hardware to finish by program, this program can be stored in the computer-readable recording medium, storage medium can comprise: ROM (read-only memory) (ROM, Read Only Memory), random access memory (RAM, Random Access Memory), disk or CD etc.
More than a kind of voiceprint authentication method based on the android platform that the embodiment of the invention is provided be described in detail, used specific case herein principle of the present invention and embodiment are set forth, the explanation of above embodiment just is used for helping to understand method of the present invention and core concept thereof; Simultaneously, for one of ordinary skill in the art, according to thought of the present invention, the part that all can change in specific embodiments and applications, in sum, this description should not be construed as limitation of the present invention.

Claims (3)

1. the voiceprint authentication method based on the Android platform is characterized in that, comprises the steps:
The user is by mobile phone input voice password; Adopt Mel cepstrum technology to obtain the vocal print characteristic parameter; When the user lands this system for the second time, the input same challenge, obtain voice voiceprint to be certified by feature extraction simultaneously, at last, compare according to the degree of confidence score and the critical value that obtain, the output result be " in person " when less than critical value, otherwise the gained result is " in person non-", thereby finishes verification process.
2. the voiceprint authentication method based on the android platform as claimed in claim 1 is characterized in that, described Mel cepstrum technology comprises:
Voice signal carries out Fast Fourier Transform (FFT) (FFT), and voice signal is transformed into frequency domain from time domain;
Before carrying out the V-belt pass filter, common frequencies to be converted into the Mel frequency usually;
The centering frequency, f ( m ) = ( N F s ) F mel - 1 [ f l + m F mel ( f h ) - F mel ( f l ) M + 1 ] , M=0 wherein, 1,2 ... 24; f h, f lBe respectively the highest and low-limit frequency of wave filter range of application; The width of representing FFT with N; F sBe sample frequency (F s=8000Hz); F Mel -1=700 (e H/1125-1) is F MelInverse function;
The H that wave filter is asked m(k) calculate spectrum energy Ek;
Cosine transform is carried out cosine transform with top gained logarithm energy by following formula, and L gets 12 in the formula, is 12 one digit number group Mel cepstrum parameter just thereby obtain length.
3. the voiceprint authentication method based on the android platform as claimed in claim 2 describedly obtains voice voiceprint to be certified by feature extraction and comprises:
The voice password that the speaker is recorded carries out feature extraction
The user imports login voice password, obtains corresponding characteristic parameter by feature extraction equally;
Two groups of argument sequences that obtain are previously mated by the voiceprint matching algorithm.
CN2013101828255A 2013-05-16 2013-05-16 Android platform-based voiceprint recognition method Pending CN103280219A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2013101828255A CN103280219A (en) 2013-05-16 2013-05-16 Android platform-based voiceprint recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2013101828255A CN103280219A (en) 2013-05-16 2013-05-16 Android platform-based voiceprint recognition method

Publications (1)

Publication Number Publication Date
CN103280219A true CN103280219A (en) 2013-09-04

Family

ID=49062713

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2013101828255A Pending CN103280219A (en) 2013-05-16 2013-05-16 Android platform-based voiceprint recognition method

Country Status (1)

Country Link
CN (1) CN103280219A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106128465A (en) * 2016-06-23 2016-11-16 成都启英泰伦科技有限公司 A kind of Voiceprint Recognition System and method
CN107666536A (en) * 2016-07-29 2018-02-06 北京搜狗科技发展有限公司 A kind of method and apparatus for finding terminal, a kind of device for being used to find terminal
CN108710788A (en) * 2018-05-22 2018-10-26 上海众人网络安全技术有限公司 A kind of safety certifying method, device, terminal and storage medium
CN110867189A (en) * 2018-08-28 2020-03-06 北京京东尚科信息技术有限公司 Login method and device
CN111311774A (en) * 2019-12-10 2020-06-19 上海明略人工智能(集团)有限公司 Sign-in method and system based on voice recognition

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1746971A (en) * 2004-09-09 2006-03-15 上海优浪信息科技有限公司 Speech key of mobile
JP2011087110A (en) * 2009-10-15 2011-04-28 Nec Corp Portable terminal, method for controlling the same, and program
CN102647521A (en) * 2012-04-05 2012-08-22 福州博远无线网络科技有限公司 Method for removing lock of mobile phone screen based on short voice command and voice-print technology
CN103077721A (en) * 2012-12-25 2013-05-01 百度在线网络技术(北京)有限公司 Voice memorandum method of mobile terminal and mobile terminal

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1746971A (en) * 2004-09-09 2006-03-15 上海优浪信息科技有限公司 Speech key of mobile
JP2011087110A (en) * 2009-10-15 2011-04-28 Nec Corp Portable terminal, method for controlling the same, and program
CN102647521A (en) * 2012-04-05 2012-08-22 福州博远无线网络科技有限公司 Method for removing lock of mobile phone screen based on short voice command and voice-print technology
CN103077721A (en) * 2012-12-25 2013-05-01 百度在线网络技术(北京)有限公司 Voice memorandum method of mobile terminal and mobile terminal

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
尹聪: "基于感知对数面积比系数的说话人确认系统的研究", 《太原理工大学硕士学位论文》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106128465A (en) * 2016-06-23 2016-11-16 成都启英泰伦科技有限公司 A kind of Voiceprint Recognition System and method
CN107666536A (en) * 2016-07-29 2018-02-06 北京搜狗科技发展有限公司 A kind of method and apparatus for finding terminal, a kind of device for being used to find terminal
CN107666536B (en) * 2016-07-29 2021-02-12 北京搜狗科技发展有限公司 Method and device for searching terminal
CN108710788A (en) * 2018-05-22 2018-10-26 上海众人网络安全技术有限公司 A kind of safety certifying method, device, terminal and storage medium
CN110867189A (en) * 2018-08-28 2020-03-06 北京京东尚科信息技术有限公司 Login method and device
CN111311774A (en) * 2019-12-10 2020-06-19 上海明略人工智能(集团)有限公司 Sign-in method and system based on voice recognition

Similar Documents

Publication Publication Date Title
CN110457432B (en) Interview scoring method, interview scoring device, interview scoring equipment and interview scoring storage medium
Lanjewar et al. Implementation and comparison of speech emotion recognition system using Gaussian Mixture Model (GMM) and K-Nearest Neighbor (K-NN) techniques
CN108986835B (en) Based on speech de-noising method, apparatus, equipment and the medium for improving GAN network
CN101980336B (en) Hidden Markov model-based vehicle sound identification method
CN102968986B (en) Overlapped voice and single voice distinguishing method based on long time characteristics and short time characteristics
CN111243602B (en) Voiceprint recognition method based on gender, nationality and emotion information
CN110085251A (en) Voice extracting method, voice extraction element and Related product
CN101923855A (en) Test-irrelevant voice print identifying system
CN107610707A (en) A kind of method for recognizing sound-groove and device
CN103280219A (en) Android platform-based voiceprint recognition method
CN107527620A (en) Electronic installation, the method for authentication and computer-readable recording medium
CN104485102A (en) Voiceprint recognition method and device
CN110265035B (en) Speaker recognition method based on deep learning
CN108198561A (en) A kind of pirate recordings speech detection method based on convolutional neural networks
CN102664010A (en) Robust speaker distinguishing method based on multifactor frequency displacement invariant feature
CN105845139A (en) Off-line speech control method and device
CN105100376A (en) Identity authentication method and apparatus
CN103594084A (en) Voice emotion recognition method and system based on joint penalty sparse representation dictionary learning
CN112786057B (en) Voiceprint recognition method and device, electronic equipment and storage medium
Sangeetha et al. Emotion speech recognition based on adaptive fractional deep belief network and reinforcement learning
CN103456302A (en) Emotion speaker recognition method based on emotion GMM model weight synthesis
CN106548786A (en) A kind of detection method and system of voice data
CN110570870A (en) Text-independent voiceprint recognition method, device and equipment
CN110232928B (en) Text-independent speaker verification method and device
Shen et al. Rars: Recognition of audio recording source based on residual neural network

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C12 Rejection of a patent application after its publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20130904