WO2016054991A1 - 声纹信息管理方法、装置以及身份认证方法、系统 - Google Patents
声纹信息管理方法、装置以及身份认证方法、系统 Download PDFInfo
- Publication number
- WO2016054991A1 WO2016054991A1 PCT/CN2015/091260 CN2015091260W WO2016054991A1 WO 2016054991 A1 WO2016054991 A1 WO 2016054991A1 CN 2015091260 W CN2015091260 W CN 2015091260W WO 2016054991 A1 WO2016054991 A1 WO 2016054991A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- information
- voice
- user
- text
- reference voiceprint
- Prior art date
Links
- 238000007726 management method Methods 0.000 title claims abstract description 55
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000001914 filtration Methods 0.000 claims abstract description 22
- 238000010586 diagram Methods 0.000 description 8
- 238000001228 spectrum Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000001303 quality assessment method Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/02—Preprocessing operations, e.g. segment selection; Pattern representation or modelling, e.g. based on linear discriminant analysis [LDA] or principal components; Feature selection or extraction
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/04—Training, enrolment or model building
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/06—Decision making techniques; Pattern matching strategies
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/003—Changing voice quality, e.g. pitch or formants
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
- G10L17/24—Interactive procedures; Man-machine interfaces the user being prompted to utter a password or a predefined phrase
Definitions
- the present application relates to the field of voiceprint recognition technology, and in particular, to a voiceprint information management method and apparatus, and an identity authentication method and system.
- Voiceprint refers to the spectrum of sound waves carrying speech information displayed by electroacoustic instruments. Different people say the same, the sound waves produced by them are different, and the corresponding sound wave spectrum, that is, the voiceprint information is also different. Therefore, by comparing the voiceprint information, it can be determined whether the corresponding speakers are the same, that is, the voiceprint recognition based identity authentication is implemented; the voiceprint recognition based identity authentication method can be widely applied to various account management systems for securing accounts. Security.
- the user before using the voiceprint recognition technology to implement identity authentication, the user first needs to read the preset text information, collect the voice signal of the user at this time, and analyze and obtain the corresponding voiceprint information as the reference voiceprint information of the user. Deposited into the voiceprint library; when implementing identity authentication, the authenticated person is also required to read the preset text information, collect the voice signal of the authenticated person, analyze and obtain the corresponding voiceprint information, and compare the voiceprint information with the sound
- the reference voiceprint information in the texture library can determine whether the authenticated person is the user himself or herself.
- the text information used for identity authentication has been disclosed when the voiceprint library is established.
- the text information required to be read by the authenticator when performing identity authentication is also known.
- the text message is a sound file, anyone can play the pre-recorded sound file to make the authentication successful. It can be seen that the existing identity authentication method based on voiceprint recognition has serious security risks.
- the present application provides a voiceprint information management method and apparatus, and an identity authentication method and system.
- a first aspect of the present application provides a voiceprint information management method, the method comprising the following steps:
- the voiceprint information management method further includes include:
- the sub-voice information corresponding to each sub-text information is respectively intercepted from the voice information according to the start and end time of the sub-text information.
- the editing the voice information and the corresponding text information into the reference voiceprint information of the first user includes:
- Each pair of sub-speech information and sub-text information is separately edited as a reference voiceprint information of the first user.
- the storing the reference voiceprint information and the identity identifier of the first user includes:
- the corresponding second text information is the same as the first text information in the first reference voiceprint information to be stored, and the corresponding second identity identifier is the first identity identifier corresponding to the first reference voiceprint information.
- the second reference voiceprint information does not exist, directly storing the first reference voiceprint information and the first identity identifier;
- the second reference voiceprint information is present, comparing the quality of the first voice information in the first reference voiceprint information and the second voice information in the second reference voiceprint information, if the first Deleting the first reference voiceprint information when the quality of the voice information is lower than the second voice information;
- the quality of the first voice information is higher than the second voice information, deleting the second reference voiceprint information, and storing the first reference voiceprint information and the first identity identifier.
- a second aspect of the present application provides a voiceprint information management apparatus, the apparatus comprising:
- a voice filter configured to acquire a historical voice file generated by the first user and the second user, and perform filtering processing on the historical voice file to obtain voice information of the first user
- a text identifier configured to perform text recognition processing on the voice information, to obtain text information corresponding to the voice information
- a voiceprint generator for editing the voice information and corresponding text information into reference voiceprint information of the first user, and storing the reference voiceprint information and an identity identifier of the first user.
- the voiceprint information management apparatus further includes:
- a text cutter for dividing the text information into a plurality of sub-text information, and marking a start and end time of each sub-text information
- the voiceprint cutter is configured to separately intercept the sub-voice information corresponding to each sub-text information from the voice information according to the start and end time of the sub-text information.
- the voiceprint generator edits the voice information and corresponding text information into a reference of the first user Voiceprint information, including:
- Each pair of sub-speech information and sub-text information is separately edited as a reference voiceprint information of the first user.
- the voiceprint generator stores the reference voiceprint information and the identity identifier of the first user, including:
- the corresponding second text information is the same as the first text information in the first reference voiceprint information to be stored, and the corresponding second identity identifier is the first identity identifier corresponding to the first reference voiceprint information.
- the second reference voiceprint information does not exist, directly storing the first reference voiceprint information and the first identity identifier;
- the second reference voiceprint information is present, comparing the quality of the first voice information in the first reference voiceprint information and the second voice information in the second reference voiceprint information, if the first Deleting the first reference voiceprint information when the quality of the voice information is lower than the second voice information;
- the quality of the first voice information is higher than the second voice information, deleting the second reference voiceprint information, and storing the first reference voiceprint information and the first identity identifier.
- the third aspect of the present application provides an identity authentication method, where the method includes the following steps:
- the voice information in the obtained reference voiceprint information is matched with the voice information to be authenticated. If the matching succeeds, it is determined that the authentication of the user to be authenticated is successful. If the matching fails, it is determined that the authentication of the user to be authenticated fails.
- the identity authentication method further includes:
- the sub-voice information corresponding to each sub-text information is respectively intercepted from the voice information according to the start and end time of the sub-text information.
- the editing the voice information and the corresponding text information into the reference voiceprint information of the first user includes:
- Each pair of sub-speech information and sub-text information is separately edited as a reference voiceprint information of the first user.
- the storing the reference voiceprint information and the identity identifier of the first user includes:
- the corresponding second text information is the same as the first text information in the first reference voiceprint information to be stored, and the corresponding second identity identifier is the first identity identifier corresponding to the first reference voiceprint information.
- the second reference voiceprint information does not exist, directly storing the first reference voiceprint information and the first identity identifier;
- the second reference voiceprint information is present, comparing the quality of the first voice information in the first reference voiceprint information and the second voice information in the second reference voiceprint information, if the first Deleting the first reference voiceprint information when the quality of the voice information is lower than the second voice information;
- the quality of the first voice information is higher than the second voice information, deleting the second reference voiceprint information, and storing the first reference voiceprint information and the first identity identifier.
- a fourth aspect of the present application provides an identity authentication system; the system includes:
- a voice filter configured to acquire a historical voice file generated by the first user and the second user, and perform filtering processing on the historical voice file to obtain voice information of the first user
- a text identifier configured to perform text recognition processing on the voice information, to obtain text information corresponding to the voice information
- a voiceprint generator configured to edit the voice information and corresponding text information into reference voiceprint information of the first user, and store the reference voiceprint information and an identity identifier of the first user;
- a voiceprint extractor configured to acquire reference voiceprint information corresponding to an identity identifier of a user to be authenticated
- Identifying a front end device configured to output text information in the obtained reference voiceprint information, and receive corresponding voice information to be authenticated
- a voiceprint matching device configured to match the voice information in the obtained reference voiceprint information with the voice information to be authenticated, and if the matching is successful, determine that the authentication of the user to be authenticated is successful, and if the matching fails, determine that the authentication is to be authenticated. User authentication failed.
- the identity authentication system further includes:
- a text cutter for dividing the text information into a plurality of sub-text information, and marking a start and end time of each sub-text information
- the voiceprint cutter is configured to separately intercept the sub-voice information corresponding to each sub-text information from the voice information according to the start and end time of the sub-text information.
- the voiceprint generator edits the voice information and corresponding text information as a reference of the first user Voiceprint information, including:
- Each pair of sub-speech information and sub-text information is separately edited as a reference voiceprint information of the first user.
- the voiceprint generator stores the reference voiceprint information and the identity identifier of the first user, including:
- the corresponding second text information is the same as the first text information in the first reference voiceprint information to be stored, and the corresponding second identity identifier is the first identity identifier corresponding to the first reference voiceprint information.
- the second reference voiceprint information does not exist, directly storing the first reference voiceprint information and the first identity identifier;
- the second reference voiceprint information is present, comparing the quality of the first voice information in the first reference voiceprint information and the second voice information in the second reference voiceprint information, if the first Deleting the first reference voiceprint information when the quality of the voice information is lower than the second voice information;
- the quality of the first voice information is higher than the second voice information, deleting the second reference voiceprint information, and storing the first reference voiceprint information and the first identity identifier.
- the present application obtains the voice information of the first user by filtering the historical voice file stored in the related system, and obtains the text information corresponding to the voice information through text recognition processing, and the voice information and the corresponding voice information.
- the text information is edited as the reference voiceprint information of the first user; since the text information and the voice information in the reference voiceprint information are obtained based on the historical voice file, which is not preset by the related system, that is, it is non-public, Whether the first user, the second user, or any other user cannot predict the specific content of the text information that needs to be read when performing identity authentication, so that the corresponding sound file cannot be recorded in advance, that is, the sound file that is recorded in advance cannot be authenticated.
- the voiceprint information management method provided by the present application performs identity authentication, the authentication result is more accurate, there is no security risk, and the account security is higher.
- FIG. 1 is a flowchart of a method for managing voiceprint information provided by an embodiment of the present application.
- FIG. 2 is a flowchart of another method for managing voiceprint information provided by an embodiment of the present application.
- FIG. 3 is a flowchart of a method for storing reference voiceprint information provided by an embodiment of the present application.
- FIG. 4 is a structural block diagram of a voiceprint information management system provided by an embodiment of the present application.
- FIG. 5 is a structural block diagram of another voiceprint information management system provided by an embodiment of the present application.
- FIG. 6 is a flowchart of an identity authentication method according to an embodiment of the present application.
- FIG. 7 is a flowchart of another identity authentication method provided by an embodiment of the present application.
- FIG. 8 is a structural block diagram of an identity authentication system according to an embodiment of the present application.
- FIG. 9 is a structural block diagram of another identity authentication system according to an embodiment of the present application.
- the voiceprint information management method includes the following steps.
- the first user may be a registered user who has a corresponding private account in the account management system, and correspondingly, the second user may be a service personnel of the account management system.
- the account management system records the voice call process between the registered user and the service personnel and stores the corresponding voice file.
- the embodiment of the present application filters out the machine prompt sound in the historical voice file stored by the account management system, the voice information of the service personnel, and the like, and obtains the voice information of the registered user, and performs text recognition processing on the voice information to obtain
- the text information corresponding to the voice information, the voice information and the corresponding text information can be used as a set of reference voiceprint information of the registered user. Performing the above steps for each registered user separately, the reference voiceprint information corresponding to each registered user can be obtained, and the voiceprint library is created.
- the embodiment of the present application obtains the voice information of the first user by filtering the historical voice file stored in the related system, and obtains the text information corresponding to the voice information through the text recognition process, and the voice information and the corresponding
- the text information is edited as the reference voiceprint information of the first user; since the text information and the voice information in the reference voiceprint information are obtained based on the historical voice file, which is not preset by the relevant system, that is, it is non-public, No matter whether the first user, the second user, or any other user can predict the specific content of the text information that needs to be read when performing identity authentication, the corresponding sound file cannot be recorded in advance, that is, the sound file that is recorded in advance cannot be played.
- the voiceprint information management method provided by the embodiment of the present application performs identity authentication, the authentication result is more accurate, there is no security risk, and the account security is higher.
- a historical voice file corresponding to an arbitrary call process of the first user and the second user may be randomly obtained, so that the identity identifiers in the voiceprint library are in one-to-one correspondence with the reference voiceprint information.
- the specific content of the text information in the obtained reference voiceprint information cannot be predicted because the actual voice file obtained by the historical voice file cannot be predicted. Therefore, performing the identity authentication based on the embodiment can ensure the accuracy of the authentication result. Sex, improve the security of the account.
- all historical voice files corresponding to the first user may also be acquired, and each historical voice file may correspond to at least one set of reference voiceprint information, so that one identity identifier in the voiceprint library may be Corresponding to multiple sets of reference voiceprint information (ie, the first user has multiple sets of reference voiceprint information); correspondingly, any set of reference voiceprint information can be randomly obtained to perform identity authentication. Since the text information in each set of reference voiceprint information is non-public, the reference voiceprint information obtained during identity authentication cannot be predicted, so the specific content of the text information used for performing identity authentication cannot be predicted, and thus cannot be advanced. Recording the corresponding sound file can not achieve the purpose of successful authentication by playing the sound file recorded in advance; therefore, performing identity authentication based on the embodiment can ensure the accuracy of the authentication result and improve the security of the account.
- the voiceprint information management method includes the following steps.
- the text information is divided into a plurality of sub-text information, and the start and end time of each sub-text information is marked.
- the sub-voice information corresponding to each sub-text information is separately intercepted from the voice information according to the start and end time of the sub-text information.
- the filtered voice information includes the plurality of pieces of voice information of the first user, and the corresponding text information obtained by the text recognition includes multiple A sentence or phrase.
- the embodiment of the present application divides the text information into a plurality of sub-text information (each sub-text information may be a sentence, a phrase or a word); at the same time, each sub-text information obtained by the segmentation is marked with a start and end time, according to the start and end time.
- the sub-speech information corresponding to the sub-text information is intercepted in the voice information (that is, the voice information is segmented according to the sub-text information).
- the sentence "My account is locked” in the text message is recognized by the 00:03 to 00:05 period of the voice message, and the "My account is locked” is divided into a sub-text information,
- the start and end time is 00:03 to 00:05.
- the voice information in the 00:03 to 00:05 period of the voice information is intercepted, that is, the subtext information corresponding to "My account is locked” is obtained.
- Sub-voice information By segmenting the text information and the voice information, a plurality of pairs of sub-text information and sub-speech information can be obtained, and are respectively edited into reference voiceprint information according to a predetermined format, thereby obtaining a plurality of reference voiceprint information corresponding to the same user.
- the sub-voice information and the corresponding sub-text information are edited into the reference voiceprint information
- the method includes: processing the sub-voice information into corresponding sub-voice texture information, and setting a file name for the sub-voice texture information,
- the format of the file name may be “soundprint number. file format suffix”, such as 0989X.WAV; storing the sub-voice texture information, and the identity identifier, sub-text information and the like of the first user corresponding to the sub-voice texture information;
- the storage structure of the voiceprint library obtained based on the above voiceprint information management method is as shown in Table 1.
- each row corresponds to a reference voiceprint information in the voiceprint library; the identity identifier (ie, user ID) is used as the primary key for querying and invoking voiceprint information; the user voiceprint number is used to mark the same user ID. The number of corresponding reference voiceprint information.
- the identity identifier ie, user ID
- the user voiceprint number is used to mark the same user ID. The number of corresponding reference voiceprint information.
- the sub-text information “Why is there no refund” is outputted; the voice information to be authenticated obtained by the user to be authenticated is re-read the information of the sub-file, and is processed as the voiceprint information to be authenticated. Comparing the voiceprint information to be authenticated and the child voiceprint information “0389X.WAV” extracted in the voiceprint library, if the two match, determining that the identity authentication is successful, that is, the user to be authenticated is the first corresponding to “139XXXXXXX” User; otherwise, if the two do not match, it is determined that the identity authentication failed.
- the embodiment of the present application obtains the voice information of the first user by filtering the historical voice file stored in the system; and performing text recognition processing on the voice information to obtain corresponding text information;
- the text information is divided into a plurality of sub-text information, and the corresponding sub-speech information is intercepted from the above-mentioned voice information according to the start and end time of each sub-text information, and each pair of sub-text information and sub-voice information is respectively edited into a reference voiceprint information, and the information is saved.
- the voiceprint library is provided so that each of the first users has a plurality of reference voiceprint information; when the identity authentication needs to be performed, one of the plurality of reference voiceprint information corresponding to the identity identifier to be authenticated is randomly selected. Since the reference voiceprint information obtained during the identity authentication is random, the specific content of the text information that needs to be read by the user to be authenticated cannot be predicted. Therefore, the voiceprint library obtained according to the embodiment performs identity authentication, and the authentication result can be guaranteed. Accuracy and improve account security.
- the sub-text information in each reference voiceprint information is short, which can reduce the time required for re-reading the text information, reduce the time consumed by the voiceprint comparison, and improve the authentication efficiency.
- the voiceprint information management method provided by the embodiment of the present application can not only create a new voiceprint library, but also update the created voiceprint library, for example, adding a reference voiceprint information corresponding to a new user, and adding a new voice to the old user.
- Reference voiceprint information For the new user, it is only necessary to obtain the historical voice file corresponding to the new user, and perform the above steps S12 to S4, or steps S22 to S26, to obtain the reference voiceprint information corresponding to the new user. As the historical voice file corresponding to the same user increases with time, the old user can obtain the corresponding new historical voice file and perform the above steps to add a new reference to the old user. Voiceprint information.
- one or more pieces of reference voiceprint information may be set for the first user.
- a plurality of pieces of reference voiceprint information are set for the same first user, it is necessary to ensure that the text information in any two reference voiceprint information corresponding to the first user is different.
- different historical voice files identify the same text information, or the same text information cuts out multiple sub-text information with the same content, so that the same sub-text information corresponds to multiple sub-texts.
- Voice information at this time, the embodiment of the present application uses the method shown in FIG. 3 to complete the storage of the reference voiceprint information.
- the reference voiceprint information to be stored is the first reference voiceprint information composed of the first text information and the first voice information.
- the first reference voiceprint information is stored in the embodiment of the present application. The process includes the following steps:
- step S31 Determine whether there is second reference voiceprint information that satisfies the comparison condition. If yes, execute step S32. Otherwise, step S34 is performed.
- the comparison condition includes: the second text information corresponding to the second reference voiceprint information is the same as the first text information in the first reference voiceprint information, and the second identity identifier corresponding to the second reference voiceprint information is The first identity identifier corresponding to a reference voiceprint information is also the same.
- step S32 Determine whether the quality of the first voice information in the first reference voiceprint information is higher than the quality of the second voice information in the second reference voiceprint information, and if yes, perform step S33, otherwise perform steps S35.
- step S31 it is determined whether the second reference voiceprint information is present, and the search range includes at least the reference voiceprint information that has been stored in the voiceprint library, and may also include the data generated synchronously with the first reference voiceprint information and not yet stored. Reference voiceprint information. If the second reference voiceprint information does not exist, the first reference voiceprint information is directly stored. If the second reference voiceprint information is found, the same first user and the same text information have at least two different voice information.
- the quality of the first voice information in the first reference voiceprint information, and the Comparing the quality of the second voice information in the second reference voiceprint information if the quality of the first voice information is higher than the second voice information, storing the first reference voiceprint information, and deleting the second reference voiceprint information, if If the quality of the voice information is lower than the second voice information, the first reference voiceprint information is directly deleted, that is, only the voice information with the highest quality is retained for the same text information, so as to improve the accuracy of the voice information comparison result in the identity authentication process, Reduce the difficulty of comparison.
- the following three voiceprint library update modes can be implemented: 1) adding reference voiceprint information of the new user; 2) increasing the reference voiceprint information of the text information corresponding to the old user; 3) The reference voiceprint information with lower voice information quality is replaced with the reference voiceprint information with higher voice information quality.
- the new reference voiceprint information obtained by the embodiment of the present application is not directly stored in the voiceprint library, but is first determined whether the text information in the reference voiceprint information and the corresponding text are stored.
- Another reference voiceprint information having the same identity identifier if present, compares the quality of the voice information in the two reference voiceprint information, retains the reference voiceprint information with higher voice information quality, and deletes the voice information with lower quality Reference voiceprint information. Therefore, the embodiment of the present application can ensure that the text information in any two reference voiceprint information corresponding to the same identity identifier (that is, the same first user) is different in the stored reference voiceprint information, and each text can be guaranteed.
- the quality of the voice information corresponding to the text information is the highest; when performing the identity authentication based on the embodiment of the present application, the voiceprint comparison based on the higher quality voice information can ensure the accuracy of the authentication and improve the authentication efficiency.
- FIG. 4 is a structural block diagram of a voiceprint information management system according to an embodiment of the present application; the voiceprint information management system can be applied to an account management system.
- the voiceprint information management system 100 includes a voice filter 110, a text recognizer 120, and a voiceprint generator 130.
- the voice filter 110 is configured to acquire a historical voice file generated by a call between the first user and the second user, and perform filtering processing on the historical voice file to obtain voice information of the first user.
- the text recognizer 120 is configured to perform text recognition processing on the voice information to obtain text information corresponding to the voice information.
- the voiceprint generator 130 is configured to edit the voice information and corresponding text information into reference voiceprint information of the first user, and store the reference voiceprint information and the identity of the first user symbol.
- the historical voice file stored in the related system is filtered to obtain the voice information of the first user, and the text information corresponding to the voice information is obtained through the text recognition process, and the voice information and the corresponding information are obtained.
- the text information is edited as the reference voiceprint information of the first user; since the text information and the voice information in the reference voiceprint information are obtained based on the historical voice file, which is not preset by the relevant system, that is, it is non-public, No matter whether the first user, the second user, or any other user can predict the specific content of the text information that needs to be read when performing identity authentication, the corresponding sound file cannot be recorded in advance, that is, the sound file that is recorded in advance cannot be played.
- the voiceprint information management method provided by the embodiment of the present application performs identity authentication, the authentication result is more accurate, there is no security risk, and the account security is higher.
- FIG. 5 is a structural block diagram of another voiceprint information management system according to an embodiment of the present disclosure; the voiceprint information management system can be applied to an account management system.
- the voiceprint information management system 200 includes a voice filter 210, a text recognizer 220, a text cutter 240, a voiceprint cutter 250, and a voiceprint generator 230.
- the voice filter 210 is configured to acquire a historical voice file generated by a call between the first user and the second user, and perform filtering processing on the historical voice file to obtain voice information of the first user.
- the text recognizer 220 is configured to perform text recognition processing on the voice information to obtain text information corresponding to the voice information.
- the text cutter 240 is configured to slice the text information into a plurality of sub-text information and mark the start and end time of each sub-text information.
- the voiceprint cutter 250 is configured to respectively intercept sub-voice information corresponding to each sub-text information from the voice information according to the start and end time of the sub-text information.
- the voiceprint generator 230 is configured to edit each pair of sub-voice information and sub-text information as a piece of reference voiceprint information of the first user, and store each piece of reference voiceprint information and the identity of the first user. Identifier.
- the embodiment of the present application obtains the voice information of the first user by filtering the historical voice file stored in the system; and performing text recognition processing on the voice information to obtain corresponding text information; and the recognized text
- the information is divided into a plurality of sub-text information, and according to the start and end time of each sub-text information from the above
- the corresponding sub-speech information is intercepted in the voice information, and each pair of sub-text information and sub-speech information is respectively edited into a reference voiceprint information, and stored in the voiceprint library, so that each first user has multiple reference voiceprint information;
- identity authentication needs to be performed, one of the plurality of reference voiceprint information corresponding to the identity identifier to be authenticated may be randomly selected.
- the reference voiceprint information obtained during the identity authentication is random, it is impossible to predict the specific content of the text information that needs to be read by the user to be authenticated, so that the corresponding sound file cannot be recorded in advance, that is, the sound file that cannot be recorded in advance cannot be played.
- the purpose of the authentication success is achieved. Therefore, performing the identity authentication based on the voiceprint library obtained in this embodiment can ensure the accuracy of the authentication result and improve the security of the account.
- the sub-text information in each reference voiceprint information is short, which can reduce the time required for re-reading the text information, reduce the time consumed by the voiceprint comparison, and improve the authentication efficiency.
- the voiceprint generator 130 and the voiceprint generator 230 may be configured. for:
- the corresponding second text information is the same as the first text information in the first reference voiceprint information to be stored, and the corresponding second identity identifier is the first identity identifier corresponding to the first reference voiceprint information.
- the second reference voiceprint information does not exist, directly storing the first reference voiceprint information and the first identity identifier;
- the second reference voiceprint information is present, comparing the quality of the first voice information in the first reference voiceprint information and the second voice information in the second reference voiceprint information, if the first If the quality of the voice information is lower than the second voice information, the first reference voiceprint information is directly deleted;
- the quality of the first voice information is higher than the second voice information, deleting the second reference voiceprint information, and storing the first reference voiceprint information and the first identity identifier.
- the embodiment of the present application can ensure that the text information in any two reference voiceprint information corresponding to the same user is different in the stored reference voiceprint information, and each text can be guaranteed.
- the quality of the voice information corresponding to the information is the highest; thus, when performing the identity authentication based on the embodiment of the present application, the voiceprint comparison based on the higher quality voice information can ensure the accuracy of the authentication and improve the authentication efficiency.
- FIG. 6 is a flowchart of an identity authentication method according to an embodiment of the present application; the identity authentication method may be applied to an account management system. Referring to FIG. 6, the identity authentication method includes the following steps.
- the first user may be a registered user who has a corresponding private account in the account management system.
- the second user may be a service personnel of the account management system.
- the embodiment of the present application obtains the voice information of the first user by filtering the historical voice file stored in the related system, and obtains the text information corresponding to the voice information through the text recognition process, and the voice information and the corresponding
- the text information is edited as the reference voiceprint information of the first user; since the text information and the voice information in the reference voiceprint information are obtained based on the historical voice file, which is not preset by the relevant system, that is, it is non-public, No matter whether the first user, the second user, or any other user can predict the specific content of the text information that needs to be read when performing identity authentication, the corresponding sound file cannot be recorded in advance, that is, the sound file that is recorded in advance cannot be played.
- the voiceprint information management method provided by the embodiment of the present application performs identity authentication, the authentication result is more accurate, there is no security risk, and the account security is higher.
- FIG. 7 is a flowchart of another identity authentication method according to an embodiment of the present application; the identity authentication method may be applied to an account management system.
- the identity authentication method includes the following steps.
- the text information is divided into a plurality of sub-text information, and the start and end time of each sub-text information is marked.
- the sub-voice information corresponding to each sub-text information is separately extracted from the voice information according to the start and end time of the sub-text information.
- the text information in the present application is divided into a plurality of sub-text information, and the corresponding sub-speech information is intercepted according to the start and end time, and each sub-text information and the corresponding sub-voice information are edited into one reference.
- the voiceprint information is such that the first user has a plurality of reference voiceprint information; when the identity authentication needs to be performed, one of the plurality of reference voiceprint information corresponding to the identity identifier to be authenticated is randomly selected. Since the reference voiceprint information obtained during the identity authentication is random, it is impossible to predict the specific content of the text information that needs to be read by the user to be authenticated, so that the corresponding sound file cannot be recorded in advance, that is, the sound file that cannot be recorded in advance cannot be played.
- the identity authentication method provided in this embodiment can ensure the accuracy of the authentication result and improve the security of the account.
- the sub-text information in each reference voiceprint information is short, which can reduce the time required for re-reading the text information, reduce the time consumed by the voiceprint comparison, and improve the authentication efficiency.
- the identity authentication method provided by the embodiment of the present application can also complete the storage of the reference voiceprint information by using the method shown in FIG. 3, and can ensure not only any two reference voiceprint information corresponding to the same user in the stored reference voiceprint information.
- the text information in the text is different, and the quality of the voice information corresponding to each type of text information is also guaranteed to be the highest; when the identity authentication is performed based on the embodiment of the present application, the voiceprint comparison based on the higher quality voice information can ensure the authentication. Accuracy and improve certification efficiency.
- FIG. 8 is a structural block diagram of an identity authentication system according to an embodiment of the present application, where the identity authentication system can be applied to an account management system.
- the identity authentication system 300 includes a voice filter 310, a text recognizer 320, a voiceprint generator 330, a voiceprint extractor 360, a recognition front end 370, and a voiceprint matcher 380.
- the voice filter 310 is configured to acquire a historical voice file generated by the first user and the second user, and perform filtering processing on the historical voice file to obtain voice information of the first user.
- the text recognizer 320 is configured to perform text recognition processing on the voice information to obtain text information corresponding to the voice information.
- the voiceprint generator 330 is configured to edit the voice information and corresponding text information into reference voiceprint information of the first user, and store the reference voiceprint information and the identity of the first user symbol.
- the voiceprint extractor 360 is configured to acquire reference voiceprint information corresponding to an identity identifier of a user to be authenticated.
- the recognition front end 370 is configured to output the text information in the acquired reference voiceprint information and receive the corresponding voice information to be authenticated.
- the voiceprint matcher 380 is configured to match the voice information in the obtained reference voiceprint information with the voice information to be authenticated, and if the matching is successful, determine that the authentication of the user to be authenticated is successful, and if the matching fails, determine The authentication of the user to be authenticated failed.
- the identification pre-amplifier 370 is configured to implement interaction between the identity authentication system and the user to be authenticated; in addition to the text information in the reference voiceprint information acquired by the voiceprint extractor 360, the user input to be authenticated is received.
- the user may also receive an identity authentication request of the user to be authenticated, and after receiving the identity authentication request, trigger the voiceprint extractor 360, and output the authentication result obtained by the voiceprint matcher 380 to the user to be authenticated.
- the historical voice file stored in the related system is filtered to obtain the voice information of the first user, and the text information corresponding to the voice information is obtained through the text recognition process, and the voice information and the corresponding information are obtained.
- the text information is edited as the reference voiceprint information of the first user; since the text information and the voice information in the reference voiceprint information are obtained based on the historical voice file, which is not preset by the relevant system, that is, it is non-public, No matter whether the first user, the second user, or any other user can predict the specific content of the text information that needs to be read when performing identity authentication, the corresponding sound file cannot be recorded in advance, that is, the sound file that is recorded in advance cannot be played.
- the voiceprint information management method provided by the embodiment of the present application performs identity authentication, the authentication result is more accurate, there is no security risk, and the account security is higher.
- FIG. 9 is a structural block diagram of an identity authentication system according to an embodiment of the present application.
- the identity authentication system can be applied to an account management system.
- the identity authentication system 400 includes a voice filter 410, a text recognizer 420, a text cutter 440, a voiceprint cutter 450, a voiceprint generator 430, a voiceprint extractor 460, a recognition front end 470, and Voiceprint matcher 480.
- the voice filter 410 is configured to acquire a historical voice file generated by the first user and the second user, and perform filtering processing on the historical voice file to obtain voice information of the first user.
- the text recognizer 420 is configured to perform text recognition processing on the voice information to obtain text information corresponding to the voice information.
- the text cutter 440 is configured to slice the text information into a plurality of sub-text information and mark the start and end time of each sub-text information.
- the voiceprint cutter 450 is configured to respectively intercept the sub-voice information corresponding to each sub-text information from the voice information according to the start and end time of the sub-text information.
- the voiceprint generator 430 is configured to edit each pair of sub-voice information and sub-text information as a piece of reference voiceprint information of the first user, and store each piece of reference voiceprint information and the identity of the first user. Identifier.
- the voiceprint extractor 460 is configured to acquire reference voiceprint information corresponding to an identity identifier of a user to be authenticated.
- the identification preamplifier 470 is configured to output sub-text information in the acquired reference voiceprint information and receive corresponding to-be-authenticated speech information.
- the voiceprint matcher 480 is configured to: the sub-voice information in the acquired reference voiceprint information and the to-be-recognized If the matching is successful, the authentication of the user to be authenticated is successful. If the matching fails, the authentication of the user to be authenticated is determined to be unsuccessful.
- the identified text information is divided into a plurality of sub-text information, and the corresponding sub-speech information is intercepted according to the start and end time, and each sub-text information and the corresponding sub-speech information are edited into one reference.
- the voiceprint information is such that the first user has a plurality of pieces of reference voiceprint information; when the identity authentication needs to be performed, the corresponding plurality of reference voiceprint information are determined from the identity identifier corresponding to the user to be authenticated, and one of the randomly selected ones is used for This identity certification.
- the identity authentication system provided in this embodiment can ensure the accuracy of the authentication result and improve the security of the account.
- the sub-text information in each reference voiceprint information is short, which can reduce the time required for re-reading the text information, reduce the time consumed by the voiceprint comparison, and improve the authentication efficiency.
- the voiceprint generator 330 and the voiceprint generator 430 may be configured to:
- the corresponding second text information is the same as the first text information in the first reference voiceprint information to be stored, and the corresponding second identity identifier is the first identity identifier corresponding to the first reference voiceprint information.
- the second reference voiceprint information does not exist, directly storing the first reference voiceprint information and the identity identifier of the first user;
- the second reference voiceprint information is present, comparing the quality of the first voice information in the first reference voiceprint information and the second voice information in the second reference voiceprint information, if the first Deleting the first reference voiceprint information when the quality of the voice information is lower than the second voice information;
- the quality of the first voice information is higher than the second voice information, deleting the second reference voiceprint information, and storing the first reference voiceprint information and a corresponding user identity identifier.
- the embodiment of the present application can ensure that the text information in any two reference voiceprint information corresponding to the same identity identifier is different in the stored reference voiceprint information, and each text can be guaranteed.
- the quality of the voice information corresponding to the text information is the highest; when the identity authentication is performed based on the embodiment of the present application, the voiceprint comparison based on the higher quality voice information can ensure the accuracy of the authentication and improve the authentication efficiency.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Quality & Reliability (AREA)
- Theoretical Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- Game Theory and Decision Science (AREA)
- Business, Economics & Management (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Telephonic Communication Services (AREA)
- Storage Device Security (AREA)
Abstract
Description
用户ID | 用户声纹编号 | 子文本信息 | 子声纹信息 |
139XXXXXXXX | 1 | 非常满意 | 0989X.WAV |
139XXXXXXXX | 2 | 为什么还没有退款 | 0389X.WAV |
189XXXXXXXX | 1 | 我很生气 | 0687X.WAV |
189XXXXXXXX | 2 | 账号被锁定 | 0361X.WAV |
Claims (16)
- 一种声纹信息管理方法,其特征在于,包括:获取第一用户与第二用户通话产生的历史语音文件;对所述历史语音文件执行过滤处理,得到所述第一用户的语音信息;对所述语音信息执行文本识别处理,得到所述语音信息对应的文本信息;将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,并存储所述基准声纹信息和所述第一用户的身份标识符。
- 根据权利要求1所述的声纹信息管理方法,其特征在于,还包括:将所述文本信息切分为多个子文本信息,并标记每个子文本信息的起止时间;根据子文本信息的起止时间从所述语音信息中分别截取每个子文本信息对应的子语音信息。
- 根据权利要求2所述的声纹信息管理方法,其特征在于,将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,包括:将每对子语音信息和子文本信息分别编辑为所述第一用户的一条基准声纹信息。
- 根据权利要求1所述的声纹信息管理方法,其特征在于,存储所述基准声纹信息和所述第一用户的身份标识符,包括:判断是否存在对应的第二文本信息与待存储的第一基准声纹信息中的第一文本信息相同,且对应的第二身份标识符与所述第一基准声纹信息对应的第一身份标识符也相同的第二基准声纹信息;如果不存在所述第二基准声纹信息,则直接存储所述第一基准声纹信息和所述第一身份标识符;如果存在所述第二基准声纹信息,则比较所述第一基准声纹信息中的第一语音信息和所述第二基准声纹信息中的第二语音信息的质量,如果所述第一语音信息的质量低于所述第二语音信息,则删除所述第一基准声纹信息;如果所述第一语音信息的质量高于所述第二语音信息,则删除所述第二基准声纹信息,并存储所述第一基准声纹信息和所述第一身份标识符。
- 一种声纹信息管理系统,其特征在于,包括:语音过滤器,用于获取第一用户与第二用户通话产生的历史语音文件,并对所述历史语音文件执行过滤处理,得到所述第一用户的语音信息;文本识别器,用于对所述语音信息执行文本识别处理,得到所述语音信息对应的文本信息;声纹生成器,用于将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,并存储所述基准声纹信息和所述第一用户的身份标识符。
- 根据权利要求5所述的声纹信息管理系统,其特征在于,还包括:文本切割器,用于将所述文本信息切分为多个子文本信息,并标记每个子文本信息的起止时间;声纹切割器,用于根据子文本信息的起止时间从所述语音信息中分别截取每个子文本信息对应的子语音信息。
- 根据权利要求6所述的声纹信息管理系统,其特征在于,所述声纹生成器将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,包括:将每对子语音信息和子文本信息分别编辑为所述第一用户的一条基准声纹信息。
- 根据权利要求5所述的声纹信息管理系统,其特征在于,所述声纹生成器存储所述基准声纹信息和所述第一用户的身份标识符,包括:判断是否存在对应的第二文本信息与待存储的第一基准声纹信息中的第一文本信息相同,且对应的第二身份标识符与所述第一基准声纹信息对应的第一身份标识符也相同的第二基准声纹信息;如果不存在所述第二基准声纹信息,则直接存储所述第一基准声纹信息和所述第一身份标识符;如果存在所述第二基准声纹信息,则比较所述第一基准声纹信息中的第一语音信息和所述第二基准声纹信息中的第二语音信息的质量,如果所述第一语音信息的质量低于所述第二语音信息,则删除所述第一基准声纹信息;如果所述第一语音信息的质量高于所述第二语音信息,则删除所述第二基准声纹信息,并存储所述第一基准声纹信息和所述第一身份标识符。
- 一种身份认证方法,其特征在于,包括:获取第一用户与第二用户通话产生的历史语音文件;对所述历史语音文件执行过滤处理,得到所述第一用户的语音信息;对所述用户语音信息语音信息执行文本识别处理,得到所述用户语音信息语音信息对应的文本信息;将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,并存储基准声纹信息和所述第一用户的身份标识符;获取待认证用户的身份标识符对应的基准声纹信息;输出获取到的基准声纹信息中的文本信息,并接收对应的待认证语音信息;将获取到的基准声纹信息中的语音信息与所述待认证语音信息进行匹配,如果匹配成功,则判定待认证用户认证成功,如果匹配失败,则判定待认证用户认证失败。
- 根据权利要求9所述的身份认证系统,其特征在于,还包括:将所述文本信息切分为多个子文本信息,并标记每个子文本信息的起止时间;根据子文本信息的起止时间从所述语音信息中分别截取每个子文本信息对应的子语音信息。
- 根据权利要求10所述的身份认证系统,其特征在于,将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,包括:将每对子语音信息和子文本信息分别编辑为所述第一用户的一条基准声纹信息。
- 根据权利要求9所述的身份认证系统,其特征在于,存储基准声纹信息和所述第一用户的身份标识符,包括:判断是否存在对应的第二文本信息与待存储的第一基准声纹信息中的第一文本信息相同,且对应的第二身份标识符与所述第一基准声纹信息对应的第一身份标识符也相同的第二基准声纹信息;如果不存在所述第二基准声纹信息,则直接存储所述第一基准声纹信息和所述第一身份标识符;如果存在所述第二基准声纹信息,则比较所述第一基准声纹信息中的第一语音信息和所述第二基准声纹信息中的第二语音信息的质量,如果所述第一语音信息的质量低于所述第二语音信息,则删除所述第一基准声纹信息;如果所述第一语音信息的质量高于所述第二语音信息,则删除所述第二基准声纹信息,并存储所述第一基准声纹信息和所述第一身份标识符。
- 一种身份认证系统,其特征在于,包括:语音过滤器,用于获取第一用户与第二用户通话产生的历史语音文件,并对所述历史语音文件执行过滤处理,得到所述第一用户的语音信息;文本识别器,用于对所述语音信息执行文本识别处理,得到所述语音信息对应的文本信息;声纹生成器,用于将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,并存储基准声纹信息和所述第一用户的身份标识符;声纹提取器,用于获取待认证用户的身份标识符对应的基准声纹信息;识别前置器,用于输出获取到的基准声纹信息中的文本信息,并接收对应的待认证语音信息;声纹匹配器,用于将获取到的基准声纹信息中的语音信息与所述待认证语音信息进行匹配,如果匹配成功,则判定待认证用户认证成功,如果匹配失败,则判定待认证用户认证失败。
- 根据权利要求13所述的身份认证系统,其特征在于,还包括:文本切割器,用于将所述文本信息切分为多个子文本信息,并标记每个子文本信息的起止时间;声纹切割器,用于根据子文本信息的起止时间从所述语音信息中分别截取每个子文本信息对应的子语音信息。
- 根据权利要求14所述的身份认证系统,其特征在于,所述声纹生成器将所述语音信息和对应的文本信息编辑为所述第一用户的基准声纹信息,包括:将每对子语音信息和子文本信息分别编辑为所述第一用户的一条基准声纹信息。
- 根据权利要求13所述的身份认证系统,其特征在于,所述声纹生成器存储基准声纹信息和所述第一用户的身份标识符,包括:判断是否存在对应的第二文本信息与待存储的第一基准声纹信息中的第一文本信息相同,且对应的第二身份标识符与所述第一基准声纹信息对应的第一身份标识符也相同的第二基准声纹信息;如果不存在所述第二基准声纹信息,则直接存储所述第一基准声纹信息和所述第一身份标识符;如果存在所述第二基准声纹信息,则比较所述第一基准声纹信息中的第一语音信息和所述第二基准声纹信息中的第二语音信息的质量,如果所述第一语音信息的质量低于所述第二语音信息,则删除所述第一基准声纹信息;如果所述第一语音信息的质量高于所述第二语音信息,则删除所述第二基准声纹信息,并存储所述第一基准声纹信息和所述第一身份标识符。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017518071A JP6671356B2 (ja) | 2014-10-10 | 2015-09-30 | 声紋情報管理方法および声紋情報管理装置、ならびに本人認証方法および本人認証システム |
KR1020177012683A KR20170069258A (ko) | 2014-10-10 | 2015-09-30 | 성문 정보 관리 방법 및 장치, 및 신원 인증 방법 및 시스템 |
SG11201702919UA SG11201702919UA (en) | 2014-10-10 | 2015-09-30 | Voiceprint information management method and apparatus, and identity authentication method and system |
EP15848463.4A EP3206205B1 (en) | 2014-10-10 | 2015-09-30 | Voiceprint information management method and device as well as identity authentication method and system |
US15/484,082 US10593334B2 (en) | 2014-10-10 | 2017-04-10 | Method and apparatus for generating voiceprint information comprised of reference pieces each used for authentication |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410532530.0A CN105575391B (zh) | 2014-10-10 | 2014-10-10 | 声纹信息管理方法、装置以及身份认证方法、系统 |
CN201410532530.0 | 2014-10-10 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/484,082 Continuation US10593334B2 (en) | 2014-10-10 | 2017-04-10 | Method and apparatus for generating voiceprint information comprised of reference pieces each used for authentication |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016054991A1 true WO2016054991A1 (zh) | 2016-04-14 |
Family
ID=55652587
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2015/091260 WO2016054991A1 (zh) | 2014-10-10 | 2015-09-30 | 声纹信息管理方法、装置以及身份认证方法、系统 |
Country Status (8)
Country | Link |
---|---|
US (1) | US10593334B2 (zh) |
EP (1) | EP3206205B1 (zh) |
JP (1) | JP6671356B2 (zh) |
KR (1) | KR20170069258A (zh) |
CN (1) | CN105575391B (zh) |
HK (1) | HK1224074A1 (zh) |
SG (2) | SG11201702919UA (zh) |
WO (1) | WO2016054991A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10593334B2 (en) | 2014-10-10 | 2020-03-17 | Alibaba Group Holding Limited | Method and apparatus for generating voiceprint information comprised of reference pieces each used for authentication |
EP3611895A4 (en) * | 2017-04-10 | 2020-04-08 | Beijing Orion Star Technology Co., Ltd. | METHOD AND DEVICE FOR USER REGISTRATION AND ELECTRONIC DEVICE |
CN111862933A (zh) * | 2020-07-20 | 2020-10-30 | 北京字节跳动网络技术有限公司 | 用于生成合成语音的方法、装置、设备和介质 |
Families Citing this family (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106156583A (zh) * | 2016-06-03 | 2016-11-23 | 深圳市金立通信设备有限公司 | 一种语音解锁的方法及终端 |
CN106549947A (zh) * | 2016-10-19 | 2017-03-29 | 陆腾蛟 | 一种即时更新的声纹认证方法及系统 |
CN106782564B (zh) * | 2016-11-18 | 2018-09-11 | 百度在线网络技术(北京)有限公司 | 用于处理语音数据的方法和装置 |
US10592649B2 (en) | 2017-08-09 | 2020-03-17 | Nice Ltd. | Authentication via a dynamic passphrase |
CN107564531A (zh) * | 2017-08-25 | 2018-01-09 | 百度在线网络技术(北京)有限公司 | 基于声纹特征的会议记录方法、装置及计算机设备 |
US10490195B1 (en) * | 2017-09-26 | 2019-11-26 | Amazon Technologies, Inc. | Using system command utterances to generate a speaker profile |
CN107863108B (zh) * | 2017-11-16 | 2021-03-23 | 百度在线网络技术(北京)有限公司 | 信息输出方法和装置 |
CN108121210A (zh) * | 2017-11-20 | 2018-06-05 | 珠海格力电器股份有限公司 | 家电设备的权限分配方法和装置、存储介质、处理器 |
CN108257604B (zh) * | 2017-12-08 | 2021-01-08 | 平安普惠企业管理有限公司 | 语音识别方法、终端设备及计算机可读存储介质 |
CN107871236B (zh) * | 2017-12-26 | 2021-05-07 | 广州势必可赢网络科技有限公司 | 一种电子设备声纹支付方法及装置 |
KR102483834B1 (ko) | 2018-01-17 | 2023-01-03 | 삼성전자주식회사 | 음성 명령을 이용한 사용자 인증 방법 및 전자 장치 |
CN111177329A (zh) * | 2018-11-13 | 2020-05-19 | 奇酷互联网络科技(深圳)有限公司 | 一种智能终端的用户交互方法、智能终端及存储介质 |
CN111292733A (zh) * | 2018-12-06 | 2020-06-16 | 阿里巴巴集团控股有限公司 | 一种语音交互方法和装置 |
CN110660398B (zh) * | 2019-09-19 | 2020-11-20 | 北京三快在线科技有限公司 | 声纹特征更新方法、装置、计算机设备及存储介质 |
CN112580390B (zh) * | 2019-09-27 | 2023-10-17 | 百度在线网络技术(北京)有限公司 | 基于智能音箱的安防监控方法、装置、音箱和介质 |
CN110970036B (zh) * | 2019-12-24 | 2022-07-12 | 网易(杭州)网络有限公司 | 声纹识别方法及装置、计算机存储介质、电子设备 |
US11516197B2 (en) * | 2020-04-30 | 2022-11-29 | Capital One Services, Llc | Techniques to provide sensitive information over a voice connection |
CN111785280B (zh) * | 2020-06-10 | 2024-09-10 | 北京三快在线科技有限公司 | 身份认证方法和装置、存储介质和电子设备 |
US11817113B2 (en) | 2020-09-09 | 2023-11-14 | Rovi Guides, Inc. | Systems and methods for filtering unwanted sounds from a conference call |
US11450334B2 (en) * | 2020-09-09 | 2022-09-20 | Rovi Guides, Inc. | Systems and methods for filtering unwanted sounds from a conference call using voice synthesis |
US12008091B2 (en) * | 2020-09-11 | 2024-06-11 | Cisco Technology, Inc. | Single input voice authentication |
US11522994B2 (en) | 2020-11-23 | 2022-12-06 | Bank Of America Corporation | Voice analysis platform for voiceprint tracking and anomaly detection |
CN112565242B (zh) * | 2020-12-02 | 2023-04-07 | 携程计算机技术(上海)有限公司 | 基于声纹识别的远程授权方法、系统、设备及存储介质 |
US12020711B2 (en) * | 2021-02-03 | 2024-06-25 | Nice Ltd. | System and method for detecting fraudsters |
US20240054235A1 (en) * | 2022-08-15 | 2024-02-15 | Bank Of America Corporation | Systems and methods for encrypting dialogue based data in a data storage system |
CN115426632A (zh) * | 2022-08-30 | 2022-12-02 | 上汽通用五菱汽车股份有限公司 | 语音传输方法、装置、车载主机以及存储介质 |
CN115565539B (zh) * | 2022-11-21 | 2023-02-07 | 中网道科技集团股份有限公司 | 一种实现自助矫正终端防伪身份验证的数据处理方法 |
CN117059092B (zh) * | 2023-10-11 | 2024-06-04 | 深圳普一同创科技有限公司 | 基于区块链的智慧医疗交互式智能分诊方法及系统 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1547191A (zh) * | 2003-12-12 | 2004-11-17 | 北京大学 | 结合语义和声纹信息的说话人身份确认系统 |
CN1852354A (zh) * | 2005-10-17 | 2006-10-25 | 华为技术有限公司 | 收集用户行为特征的方法和装置 |
US7158776B1 (en) * | 2001-09-18 | 2007-01-02 | Cisco Technology, Inc. | Techniques for voice-based user authentication for mobile access to network services |
CN102708867A (zh) * | 2012-05-30 | 2012-10-03 | 北京正鹰科技有限责任公司 | 一种基于声纹和语音的防录音假冒身份识别方法及系统 |
CN102760434A (zh) * | 2012-07-09 | 2012-10-31 | 华为终端有限公司 | 一种声纹特征模型更新方法及终端 |
CN103258535A (zh) * | 2013-05-30 | 2013-08-21 | 中国人民财产保险股份有限公司 | 基于声纹识别的身份识别方法及系统 |
Family Cites Families (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11344992A (ja) * | 1998-06-01 | 1999-12-14 | Ntt Data Corp | 音声辞書作成方法、個人認証装置および記録媒体 |
US20040236699A1 (en) | 2001-07-10 | 2004-11-25 | American Express Travel Related Services Company, Inc. | Method and system for hand geometry recognition biometrics on a fob |
IL154733A0 (en) | 2003-03-04 | 2003-10-31 | Financial transaction authorization apparatus and method | |
JP4213716B2 (ja) | 2003-07-31 | 2009-01-21 | 富士通株式会社 | 音声認証システム |
US7386448B1 (en) | 2004-06-24 | 2008-06-10 | T-Netix, Inc. | Biometric voice authentication |
US8014496B2 (en) | 2004-07-28 | 2011-09-06 | Verizon Business Global Llc | Systems and methods for providing network-based voice authentication |
US7536304B2 (en) | 2005-05-27 | 2009-05-19 | Porticus, Inc. | Method and system for bio-metric voice print authentication |
US20060277043A1 (en) | 2005-06-06 | 2006-12-07 | Edward Tomes | Voice authentication system and methods therefor |
JP4755689B2 (ja) * | 2005-07-27 | 2011-08-24 | インターナショナル・ビジネス・マシーンズ・コーポレーション | 正規受信者への安全なファイル配信のためのシステムおよび方法 |
JP4466572B2 (ja) * | 2006-01-16 | 2010-05-26 | コニカミノルタビジネステクノロジーズ株式会社 | 画像形成装置、音声コマンド実行プログラムおよび音声コマンド実行方法 |
CN1808567A (zh) | 2006-01-26 | 2006-07-26 | 覃文华 | 验证真人在场状态的声纹认证设备和其认证方法 |
US8396711B2 (en) * | 2006-05-01 | 2013-03-12 | Microsoft Corporation | Voice authentication system and method |
US20080256613A1 (en) | 2007-03-13 | 2008-10-16 | Grover Noel J | Voice print identification portal |
US8775187B2 (en) | 2008-09-05 | 2014-07-08 | Auraya Pty Ltd | Voice authentication system and methods |
US8537978B2 (en) * | 2008-10-06 | 2013-09-17 | International Business Machines Corporation | Method and system for using conversational biometrics and speaker identification/verification to filter voice streams |
US8655660B2 (en) * | 2008-12-11 | 2014-02-18 | International Business Machines Corporation | Method for dynamic learning of individual voice patterns |
CN102404287A (zh) | 2010-09-14 | 2012-04-04 | 盛乐信息技术(上海)有限公司 | 用数据复用法确定声纹认证阈值的声纹认证系统及方法 |
US9318114B2 (en) * | 2010-11-24 | 2016-04-19 | At&T Intellectual Property I, L.P. | System and method for generating challenge utterances for speaker verification |
CN102222502A (zh) * | 2011-05-16 | 2011-10-19 | 上海先先信息科技有限公司 | 一种汉语随机提示声纹验证的有效方式 |
KR101304112B1 (ko) * | 2011-12-27 | 2013-09-05 | 현대캐피탈 주식회사 | 음성 분리를 이용한 실시간 화자인식 시스템 및 방법 |
US10134401B2 (en) * | 2012-11-21 | 2018-11-20 | Verint Systems Ltd. | Diarization using linguistic labeling |
JP5646675B2 (ja) * | 2013-03-19 | 2014-12-24 | ヤフー株式会社 | 情報処理装置及び方法 |
US20140359736A1 (en) | 2013-05-31 | 2014-12-04 | Deviceauthority, Inc. | Dynamic voiceprint authentication |
CN103679452A (zh) * | 2013-06-20 | 2014-03-26 | 腾讯科技(深圳)有限公司 | 支付验证方法、装置及系统 |
GB2517952B (en) * | 2013-09-05 | 2017-05-31 | Barclays Bank Plc | Biometric verification using predicted signatures |
US8812320B1 (en) * | 2014-04-01 | 2014-08-19 | Google Inc. | Segment-based speaker verification using dynamically generated phrases |
CN105575391B (zh) | 2014-10-10 | 2020-04-03 | 阿里巴巴集团控股有限公司 | 声纹信息管理方法、装置以及身份认证方法、系统 |
-
2014
- 2014-10-10 CN CN201410532530.0A patent/CN105575391B/zh active Active
-
2015
- 2015-09-30 WO PCT/CN2015/091260 patent/WO2016054991A1/zh active Application Filing
- 2015-09-30 EP EP15848463.4A patent/EP3206205B1/en active Active
- 2015-09-30 KR KR1020177012683A patent/KR20170069258A/ko not_active Application Discontinuation
- 2015-09-30 JP JP2017518071A patent/JP6671356B2/ja active Active
- 2015-09-30 SG SG11201702919UA patent/SG11201702919UA/en unknown
- 2015-09-30 SG SG10201903085YA patent/SG10201903085YA/en unknown
-
2016
- 2016-10-26 HK HK16112295.9A patent/HK1224074A1/zh unknown
-
2017
- 2017-04-10 US US15/484,082 patent/US10593334B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7158776B1 (en) * | 2001-09-18 | 2007-01-02 | Cisco Technology, Inc. | Techniques for voice-based user authentication for mobile access to network services |
CN1547191A (zh) * | 2003-12-12 | 2004-11-17 | 北京大学 | 结合语义和声纹信息的说话人身份确认系统 |
CN1852354A (zh) * | 2005-10-17 | 2006-10-25 | 华为技术有限公司 | 收集用户行为特征的方法和装置 |
CN102708867A (zh) * | 2012-05-30 | 2012-10-03 | 北京正鹰科技有限责任公司 | 一种基于声纹和语音的防录音假冒身份识别方法及系统 |
CN102760434A (zh) * | 2012-07-09 | 2012-10-31 | 华为终端有限公司 | 一种声纹特征模型更新方法及终端 |
CN103258535A (zh) * | 2013-05-30 | 2013-08-21 | 中国人民财产保险股份有限公司 | 基于声纹识别的身份识别方法及系统 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10593334B2 (en) | 2014-10-10 | 2020-03-17 | Alibaba Group Holding Limited | Method and apparatus for generating voiceprint information comprised of reference pieces each used for authentication |
EP3611895A4 (en) * | 2017-04-10 | 2020-04-08 | Beijing Orion Star Technology Co., Ltd. | METHOD AND DEVICE FOR USER REGISTRATION AND ELECTRONIC DEVICE |
US11568876B2 (en) | 2017-04-10 | 2023-01-31 | Beijing Orion Star Technology Co., Ltd. | Method and device for user registration, and electronic device |
CN111862933A (zh) * | 2020-07-20 | 2020-10-30 | 北京字节跳动网络技术有限公司 | 用于生成合成语音的方法、装置、设备和介质 |
Also Published As
Publication number | Publication date |
---|---|
CN105575391A (zh) | 2016-05-11 |
EP3206205B1 (en) | 2020-01-15 |
JP6671356B2 (ja) | 2020-03-25 |
EP3206205A1 (en) | 2017-08-16 |
US10593334B2 (en) | 2020-03-17 |
SG10201903085YA (en) | 2019-05-30 |
US20170221488A1 (en) | 2017-08-03 |
KR20170069258A (ko) | 2017-06-20 |
SG11201702919UA (en) | 2017-05-30 |
HK1224074A1 (zh) | 2017-08-11 |
CN105575391B (zh) | 2020-04-03 |
JP2017534905A (ja) | 2017-11-24 |
EP3206205A4 (en) | 2017-11-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2016054991A1 (zh) | 声纹信息管理方法、装置以及身份认证方法、系统 | |
US10685657B2 (en) | Biometrics platform | |
US10135818B2 (en) | User biological feature authentication method and system | |
US10276168B2 (en) | Voiceprint verification method and device | |
CN102985965B (zh) | 声纹标识 | |
CN105069874B (zh) | 一种移动互联网声纹门禁系统及其实现方法 | |
US20160014120A1 (en) | Method, server, client and system for verifying verification codes | |
WO2019127897A1 (zh) | 一种自学习声纹识别的更新方法和装置 | |
CN110533288A (zh) | 业务办理流程检测方法、装置、计算机设备和存储介质 | |
US20130132091A1 (en) | Dynamic Pass Phrase Security System (DPSS) | |
CN109036436A (zh) | 一种声纹数据库建立方法、声纹识别方法、装置及系统 | |
US11076043B2 (en) | Systems and methods of voiceprint generation and use in enforcing compliance policies | |
CN106982344A (zh) | 视频信息处理方法及装置 | |
WO2016107415A1 (zh) | 基于用户网络行为特征的辅助身份验证方法 | |
KR101181060B1 (ko) | 음성 인식 시스템 및 이를 이용한 화자 인증 방법 | |
US20120330663A1 (en) | Identity authentication system and method | |
US20140163986A1 (en) | Voice-based captcha method and apparatus | |
US11705134B2 (en) | Graph-based approach for voice authentication | |
KR102291113B1 (ko) | 회의록 작성 장치 및 방법 | |
KR20220166465A (ko) | 다채널 수신기를 이용한 회의록 생성 시스템 및 방법 | |
US10572636B2 (en) | Authentication by familiar media fragments | |
Yakovlev et al. | LRPD: Large Replay Parallel Dataset | |
Portêlo et al. | Privacy-preserving query-by-example speech search | |
CN114125368B (zh) | 会议音频的参会人关联方法、装置及电子设备 | |
RU2628118C2 (ru) | Способ формирования и использования инвертированного индекса аудиозаписи и машиночитаемый носитель информации |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15848463 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2017518071 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11201702919U Country of ref document: SG |
|
ENP | Entry into the national phase |
Ref document number: 20177012683 Country of ref document: KR Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2015848463 Country of ref document: EP |