WO2019006587A1 - Système de reconnaissance de locuteur, procédé de reconnaissance de locuteur et dispositif intra-auriculaire - Google Patents

Système de reconnaissance de locuteur, procédé de reconnaissance de locuteur et dispositif intra-auriculaire Download PDF

Info

Publication number
WO2019006587A1
WO2019006587A1 PCT/CN2017/091466 CN2017091466W WO2019006587A1 WO 2019006587 A1 WO2019006587 A1 WO 2019006587A1 CN 2017091466 W CN2017091466 W CN 2017091466W WO 2019006587 A1 WO2019006587 A1 WO 2019006587A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
ear
signal
ear canal
voiceprint
Prior art date
Application number
PCT/CN2017/091466
Other languages
English (en)
Chinese (zh)
Inventor
黄彦颖
Original Assignee
深圳市汇顶科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市汇顶科技股份有限公司 filed Critical 深圳市汇顶科技股份有限公司
Priority to CN201780000606.7A priority Critical patent/CN110100278B/zh
Priority to PCT/CN2017/091466 priority patent/WO2019006587A1/fr
Publication of WO2019006587A1 publication Critical patent/WO2019006587A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification

Definitions

  • the present application relates to a speaker recognition system and a speaker recognition method, and more particularly to a speaker recognition system and a speaker recognition method that can avoid being recorded or pirated.
  • Speaker recognition has been widely used in voice security systems or voice authorization systems, and has become an indispensable feature in contemporary technology products.
  • the existing speech recognition system mainly uses the microphone outside the human body to collect the sound, and the sound received by the human body is the sound wave transmitted by the human body through the oral cavity and transmitted through the external air medium, and the existing speaker identification is affected by the person concerned.
  • the risk of recording or pirating In detail, a certain prince can track a certain prince and record the voice of the prince, or eavesdrop on the voice of the prince, and even use the technique of speech synthesis to forge the voice of the singer and store the voice of the prince in advance.
  • the present application provides a speaker recognition system including an in-ear device for inserting an external auditory canal of a user, the in-ear device comprising a sound receiver for receiving from the external auditory canal.
  • An ear canal sound wave to generate an ear canal acoustic signal corresponding to the ear canal sound wave; a frequency processing module, coupled to the sound receiver, for extracting a voiceprint feature corresponding to the user from the ear canal sound signal to generate a voiceprint feature signal; and a terminal device for The voiceprint feature signal determines whether the user is an authenticated user.
  • the in-ear device is a wired or wireless in-ear earphone, an in-ear earphone microphone, an earplug or a hearing aid.
  • the audio processing module performs a speech detection operation and a feature extraction operation on the ear canal acoustic signal to generate the voiceprint feature signal.
  • the audio processing module performs a noise suppression operation on the ear canal acoustic signal.
  • the terminal device is a mobile electronic device, a computer host or an access control system.
  • the terminal device establishes a voiceprint model corresponding to the authenticated user, and receives a voiceprint feature signal from the audio processing module, and compares the voiceprint feature signal according to the voiceprint model to generate a The similarity signal, the terminal device determines, according to the similarity signal, whether the user is the authenticated user.
  • the audio processing module performs a physiological detection operation on the ear canal acoustic signal to generate a physiological detection result
  • the terminal device determines the user according to the voiceprint characteristic signal and the physiological detection result. Whether it is the authenticated user.
  • the physiological detection operation is a respiratory detection operation
  • the physiological detection result is a respiratory detection result
  • the physiological detection operation is a heart rate detection operation
  • the physiological detection result is a heart rate detection result
  • the present application also provides a speaker recognition method, which is applied to a speaker recognition system
  • the speaker recognition system includes an in-ear device and a terminal device
  • the in-ear device includes a microphone and an audio processing module.
  • the in-ear device is placed in an external auditory canal of a user
  • the speaker identification method includes the microphone receiving an ear canal sound wave from the external auditory canal to generate an acoustic wave corresponding to the ear canal An ear canal acoustic signal
  • the audio processing module extracts a voiceprint feature corresponding to the user from the ear canal acoustic signal to generate a voiceprint feature signal
  • the terminal Determining, according to the voiceprint characteristic signal, whether an utterance of the speaker recognition system is the user itself; wherein the utterance of the speaker recognition system is for the speaker recognition system A person or device that performs sound recognition.
  • the present application utilizes an in-ear device for receiving sound to receive the ear canal sound waves from the user's external ear, and uses the audio processing module in the in-ear device to capture the voiceprint characteristics of the user, and uses the terminal device to perform voiceprint comparison to determine Whether the speaker of the speaker recognition system is the user itself.
  • the present application can avoid the risk of being recorded or pirated by a person with a heart.
  • FIG. 1 is a schematic diagram of the appearance of a speaker recognition system according to an embodiment of the present application.
  • FIG. 2 is a functional block diagram of the speaker recognition system of FIG. 1.
  • FIG. 3 is a schematic diagram of a voiceprint identification process according to an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a voiceprint feature extraction process according to an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a voiceprint comparison process according to an embodiment of the present application.
  • FIG. 6 is a schematic diagram of a voiceprint identification process according to an embodiment of the present application.
  • FIG. 7 is a schematic diagram of a voiceprint identification process according to an embodiment of the present application.
  • FIG. 8 is a schematic functional block diagram of a speaker recognition system according to an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a voiceprint identification process according to an embodiment of the present application.
  • the existing speaker recognition mainly receives the sound waves sent by the speaker's mouth, and transmits the sound waves to the microphone outside the human body through the air medium outside the human body, and for the security system that needs to be recognized by the speaker (such as the voice access control system, In the case of voice payment systems, there is a risk that existing speaker recognition has been recorded or stolen by interested people.
  • the sound waves generated by the vocal cord mucosa are transmitted through the Eustachian tube to the Internal Auditory Meatus or even the External Auditory Meatus, and to the external auditory canal.
  • Sound waves (or ear canal sound waves) have different sound characteristics than sound waves received by microphones other than the human body. In other words, even if the speaker is the same person, the ear canal sound waves are different from the side-recorded or stolen sound waves. Sound characteristics.
  • the speaker recognition system of the present application performs sound collection on the external auditory canal of the user, and captures the voiceprint features of the ear canal sound waves, and performs speaker recognition on the voiceprint features of the ear canal sound waves to prevent the user's voice from being affected.
  • FIG. 1 and FIG. 2 are schematic diagrams showing the appearance and functional blocks of the speaker recognition system 10 according to the embodiment of the present application.
  • the speaker recognition system 10 includes an In-Ear device (ie, a Canal-Type Device) 100 and a terminal device 120.
  • the terminal device 120 can be a computer host with computing functions and mobile electronics.
  • the device or the access control system, the in-ear device 100 can be placed into an external ear canal (Canal, ie, External Acoustic Meatus) of the user USR, which can be an earphone, an earphone, and an earplug. (Earplug) or one of the hearing aids (Hearing Aid).
  • the in-ear device 100 can include a radio 102, a speaker 104, and an audio processing module 106.
  • the radio 102 can be a microphone for receiving an ear canal CWV from the external ear canal of the user USR.
  • the ear canal acoustic wave CWV is converted into an ear canal acoustic signal CSg, that is, the radio receiver 102 can generate an ear canal acoustic signal CSg corresponding to the ear canal acoustic wave CWV.
  • the audio processing module 106 is coupled to the radio receiver 102, and extracts a voiceprint feature corresponding to the user USR from the ear canal acoustic signal CSg to generate a voiceprint feature signal VPF, wherein the voiceprint feature signal VPF includes User Voiceprint features of the USR.
  • the in-ear device 100 can transmit the voiceprint feature signal VPF to the terminal device 120 through wired transmission or wireless transmission.
  • the terminal device 120 can determine whether the user USR is an authenticated user or other person according to the voiceprint characteristic signal received by the user, or even a recorder that records the USR sound of the user in advance, wherein the speaker recognition system
  • the utterance of 10 refers to a person or device (such as a recorder or a device having a speech synthesis function) that emits a sound to the speaker recognition system 10 for voiceprint recognition.
  • the terminal device 120 can determine whether the user USR is an authenticated user according to the voiceprint feature signal it receives. In an ideal case, the terminal device 120 receives the voiceprint feature signal VPF generated by the in-ear device 100, and determines that the user USR is indeed an authenticated user based on the voiceprint feature signal VPF.
  • FIG. 3 is a schematic diagram of a voiceprint identification process 30 according to an embodiment of the present application.
  • the voiceprint recognition process 30 can be performed by the speaker recognition system 10, which includes the following steps:
  • Step 302 The tuner 102 of the in-ear device 100 receives the ear canal acoustic wave CWV from the external ear canal of the user USR, and generates an ear canal acoustic signal CSg corresponding to the ear canal acoustic wave CWV.
  • Step 304 The audio processing module 106 of the in-ear device 100 extracts a voiceprint feature corresponding to the user USR from the ear canal acoustic signal CSg, and generates a voiceprint feature signal VPF.
  • Step 306 The terminal device 120 determines whether the user USR is an authenticated user according to the voiceprint feature signal VPF.
  • step 304 the audio processing module 106 extracts the operation details corresponding to the voiceprint feature of the user USR and generates the voiceprint feature signal VPF from the ear canal acoustic signal CSg.
  • FIG. 4 is a voiceprint feature capture.
  • the schematic of the flow 40, the voiceprint feature extraction process 40 is performed by the audio processing module 106 of the in-ear device 100.
  • the audio processing module 106 can perform a voice detection operation, a noise suppression (Noise Suppression) operation, and a feature extraction operation on the ear canal acoustic signal CSg to generate voiceprint features.
  • the signal VPF wherein the speech detection operation, the noise suppression operation, and the feature extraction operation are not limited to being implemented by using a specific algorithm, and the technical details thereof are well known to those skilled in the art, and thus will not be described herein.
  • the sound pattern The speech detection operation, the noise suppression operation, and the feature extraction operation in the enrollment process 40 are all performed by the audio processing module 106 disposed in the in-ear device 100, that is, generated by the audio processing module 106 disposed in the in-ear device 100.
  • Voiceprint feature signal VPF After the audio processing module 106 generates the voiceprint feature signal VPF, the voiceprint feature signal VPF can be transmitted to the terminal device 120 by wire transmission or wireless transmission.
  • step 306 the terminal device 120 determines, according to the voiceprint characteristic signal VPF, whether the user USR is an operation detail of the authenticated user.
  • FIG. 5 is a schematic diagram of a voiceprint comparison process 50, and the voiceprint comparison process 50 is performed by the human body. It is executed by the terminal device 120 other than the terminal device 120. As shown in FIG.
  • the terminal device 120 may first establish a voiceprint model MD corresponding to the authenticated user according to the voiceprint feature signal VPF, and after establishing the voiceprint model MD, compare the voiceprint feature signal VPF with the voiceprint model MD to perform "soundprint matching", and according to the voiceprint matching result, a similarity score SC (Score) is generated, wherein the similarity score SC represents the degree of similarity between the voiceprint characteristic signal VPF and the voiceprint model MD, which may be one Similarity signal.
  • Score similarity score
  • the terminal device 120 may correspond to the established user authentication voiceprint model MD (or the times t 1 to the first audio processing module 106 receives the generated corresponding to the authenticated user a first sound pattern feature in a first time t 1 VPF1 signal, and establishes corresponding to the authenticated user according to the first acoustic model MD voiceprint characteristic signal pattern VPF1, first acoustic signal characteristic pattern voiceprint features VPF1 t represents a first time signal to the VPF 1), to establish a model voiceprint after the MD, the terminal device 120 may be at a second time t 2 receiving the audio signal processing a second acoustic characteristic pattern VPF2 (which represents the second time t voiceprint signal VPF 2) is generated by module 106, the terminal apparatus 120 The second voiceprint characteristic signal VPF2 and the voiceprint model MD may be compared for voiceprint matching, and a similarity score SC is generated according to the voiceprint matching result.
  • the terminal device 120 After the terminal device 120 generates the similarity score SC, it can determine whether the user USR is an authenticated user based on the similarity score SC, that is, the terminal device 120 performs the step of "identifying the identity” in FIG. In an embodiment, when the similarity score SC is greater than a specific value, the terminal device 120 may determine that the user USR is indeed an authenticated user.
  • the steps of "establishing a voiceprint model", “soundprint matching”, and “acquiring a similarity score” in FIG. 5 are not limited to being implemented by using a specific algorithm, and the technical details thereof are well known to those skilled in the art, and thus Let me repeat.
  • the speaker recognition system 10 receives the ear canal sound wave CWV by using the in-ear device 100, and uses the audio processing module 106 to capture the voiceprint feature corresponding to the user USR, and uses the terminal device 120 to determine the use according to the voiceprint characteristic signal VPF. Whether the USR is an authenticated user.
  • a speaker-recognized security system such as a voice access control system, hereinafter referred to as a voice security system
  • a voice authorization system a system that is recognized by a speaker to confirm the identity of a speaker for authorization to proceed to the next step, such as a voice payment system
  • voice transfer transaction system a system that is recognized by a speaker to confirm the identity of a speaker for authorization to proceed to the next step, such as a voice payment system
  • voice transfer transaction system such as a voice credit card transaction system or voice login system, etc.
  • the speaker recognition system 10 performs sound collection on the external auditory canal of the user USR, and performs voiceprint recognition on the voiceprint characteristics of the ear canal sound wave CWV, since the ear canal sound wave is different from the sound wave received by the external microphone.
  • the voice features, and the intent of the person can not crack the voice security system with the speaker recognition system 10 via side recording, piracy or voice synthesis, which can further enhance the security of the voice security system or the voice authorization system.
  • the audio processing module 106 can determine whether the ear canal acoustic wave CWV has a respiratory sound wave by the ear canal acoustic signal CSg, that is, perform a physiological detection operation on the ear canal acoustic signal CSg to confirm that the speech end of the speaker recognition system 10 has physiological characteristics.
  • a natural person rather than a device such as a tape recorder or a speech synthesizer, wherein the physiological detection operation can be a breathing detection operation or even a heart rate detection operation.
  • FIG. 6 is a schematic diagram of a voiceprint identification process 60 according to an embodiment of the present application.
  • the voiceprint recognition process 60 can be performed by the speaker recognition system 10, which includes the following steps:
  • Step 602 The tuner 102 of the in-ear device 100 receives the ear canal acoustic wave CWV from the external ear canal of the user USR, and generates an ear canal acoustic signal CSg corresponding to the ear canal acoustic wave CWV.
  • Step 603 The audio processing module 106 of the in-ear device 100 performs a physiological detection operation on the ear canal acoustic signal CSg to generate a physiological detection result Bio.
  • Step 604 The audio processing module 106 of the in-ear device 100 extracts the voiceprint feature corresponding to the user USR from the ear canal acoustic signal CSg, and generates a voiceprint feature signal VPF.
  • Step 606 The terminal device 120 determines, according to the voiceprint characteristic signal VPF and the physiological detection result Bio, whether the user USR is the authenticated user itself.
  • the voiceprint recognition process 60 is similar to the voiceprint recognition process 30. Unlike the voiceprint recognition process 30, the voiceprint recognition process 60 further includes a step 603.
  • the audio processing module 106 is not limited to performing a breath detection operation on the ear canal acoustic signal CSg by using a specific algorithm. For example, the audio processing module 106 can detect whether the ear canal sound wave CWV has a specific respiratory frequency according to the ear canal acoustic signal CSg. Breathing sound waves, not limited to this. The technical details of the breath detection operation are well known to those skilled in the art and will not be described herein.
  • the physiological test result Bio can be a binary value (Binary Value), which represents the detection of "having breathing” or “no breathing”, when the physiological test result Bio indicates that "there is In the case of "breathing", the speaking end of the representative speaker recognition system 10 is a natural person.
  • the physiological detection result Bio may also be a non-binary value such as a gray level value, and the representative representative detects that there is “breathing” (or The Confidence Level of the "no breath” is detected, or the specific respiratory rate and characteristics of the user USR.
  • the terminal device 120 determines whether the speech end of the speaker recognition system 10 is the user USR itself based on the voiceprint characteristic signal VPF and the physiological detection result Bio. In one embodiment, when the physiological detection result Bio indicates that "breathing" is detected and the similarity score SC is greater than a specific value, the terminal device 120 may determine that the user USR is indeed an authenticated user.
  • the voice security system or voice authorization system usually has a question-and-answer dialogue situation.
  • the bank or credit card center, payment system center, hereinafter referred to as the customer service
  • the customer service may ask in the voice call: May I ask your account?" and the user may answer: "123456789", in which the customer service question can be sent to the external auditory canal of the user USR through the speaker 104.
  • the ear canal sound wave CWV can include the client's question sound wave.
  • the audio processing module 106 in the in-ear device 100 can determine whether the ear canal sound wave CWV has a reflected sound wave to the question sound wave by the ear canal sound signal CSg to generate a reflected wave detection result.
  • the acoustic wave CWV When the reflected wave detection result shows the ear canal
  • the speech end of the speaker recognition system 10 is a natural person, rather than a device such as a recorder or a speech synthesizer, thereby eliminating the possibility that the speech end of the speaker recognition system 10 is a device.
  • the question sound wave can be broadly regarded as the prompt sound wave.
  • the user USR After the prompt sound wave ends, the user USR can start to speak.
  • the customer service terminal may say in the voice call: "Please hear the buzzer and read you.
  • the account number/password ie, the prompt statement
  • the prompt sound wave may include a sound wave or the click sound related to the prompt statement.
  • FIG. 7 is a schematic diagram of a voiceprint identification process 70 according to an embodiment of the present application.
  • the voiceprint recognition process 70 can be performed by the speaker recognition system 10, which includes the following steps:
  • Step 701 The speaker 104 sends a prompt sound wave to the user's USR external auditory canal.
  • Step 702 The tuner 102 of the in-ear device 100 receives the ear canal acoustic wave CWV from the external ear canal of the user USR, and generates an ear canal acoustic signal CSg corresponding to the ear canal acoustic wave CWV.
  • Step 703 The audio processing module 106 of the in-ear device 100 determines whether there is a reflected sound wave corresponding to the prompt sound wave in the ear canal sound wave CWV according to the ear canal sound signal CSg to generate a reflected wave detection result Rf.
  • Step 704 The audio processing module 106 of the in-ear device 100 extracts the voiceprint feature corresponding to the user USR from the ear canal acoustic signal CSg, and generates a voiceprint feature signal VPF.
  • Step 706 The terminal device 120 determines whether the user USR is an authenticated user based on the voiceprint characteristic signal VPF and the reflected wave detection result Rf.
  • the voiceprint recognition process 70 is similar to the voiceprint recognition process 30. Different from the voiceprint recognition process 30, the voiceprint recognition process 70 further includes steps 701 and 703.
  • the audio processing module 106 is not limited to determining whether there is a reflected sound wave corresponding to the prompt sound wave in the ear canal sound wave CWV by using a specific algorithm. For example, since the outer ear prop of the human body has an ear canal length range, the audio processing module 106 can According to the length range of the ear canal, it is judged whether there is a reflected sound wave corresponding to the prompt sound wave in the ear canal sound wave CWV.
  • the technical details of the physiological detection operation (such as the breathing detection operation or the heart rate detection operation) in the ear canal are well known to those skilled in the art, and thus will not be described herein.
  • the reflected wave detection result Rf can be a binary number The value represents "reflected wave” or "no reflected wave”. When the reflected wave detection result Rf indicates "reflected wave”, the caller of the speaker recognition system 10 is a natural person.
  • the terminal device 120 determines whether the originating end of the speaker recognition system 10 is the user USR itself based on the voiceprint characteristic signal VPF and the reflected wave detection result Rf. In one embodiment, when the reflected wave detection result Rf indicates "reflected wave” and the similarity score SC is greater than a specific value, the terminal device 120 may determine that the user USR is indeed an authenticated user.
  • the speaker recognition system of the present application can perform a voice change on the voiceprint characteristic signal VPF generated by the in-ear device 100 by using a personal electronic device such as a smart phone (Voice)
  • the terminal device 120 performs speaker recognition based on the voiced characteristic signal after the voice change, that is, whether the voice end of the speaker recognition system is the user USR according to the voice change characteristic signal after the voice change.
  • the user USR can further increase the security of the voice security system or the voice authorization system only through the authentication process that the terminal device 120 recognizes the speaker when holding the personal electronic device.
  • FIG. 8 is a functional block diagram of a speaker recognition system 80 according to an embodiment of the present application.
  • the speaker recognition system 80 is similar to the speaker recognition system 10.
  • the speaker recognition system 80 further includes a personal electronic device 800, which may be a smart wearable device, a smart phone, or a tablet computer.
  • the personal electronic device 800 receives the voiceprint characteristic signal VPF generated by the in-ear device 100, and performs a voice-changing operation on the voiceprint characteristic signal VPF to generate a voice-changing voiceprint characteristic signal VPF', and
  • the voice-changing voiceprint characteristic signal VPF' is transmitted to the terminal device 120, and the terminal device 120 performs speaker recognition based on the voice-changing voiceprint characteristic signal VPF'.
  • FIG. 9 is a schematic diagram of a voiceprint identification process 90 according to an embodiment of the present application.
  • the voiceprint recognition process 90 can be performed by the speaker recognition system 80, which includes the following steps:
  • Step 902 The tuner 102 of the in-ear device 100 receives the ear canal acoustic wave CWV from the external ear canal of the user USR, and generates an ear canal acoustic signal CSg corresponding to the ear canal acoustic wave CWV.
  • Step 904 The audio processing module 106 of the in-ear device 100 extracts a voiceprint feature corresponding to the user USR from the ear canal acoustic signal CSg, and generates a voiceprint feature signal VPF.
  • Step 905 The personal electronic device 800 performs a voice-changing operation on the voiceprint feature signal VPF to generate a voice-changing voiceprint feature signal VPF'.
  • Step 906 The terminal device 120 determines whether the user USR is an authenticated user based on the voiced characteristic signal VPF'.
  • the voiceprint recognition process 90 is similar to the voiceprint recognition process 30. Unlike the voiceprint recognition process 30, the voiceprint recognition process 90 further includes a step 905. In step 905, the personal electronic device 800 is not limited to performing a voice-changing operation on the voiceprint feature signal VPF by using a specific algorithm to generate a voice-changing voiceprint feature signal VPF' to encrypt the information/voiceprint feature signal VPF. It is well known to those skilled in the art, and thus will not be described herein.
  • the terminal device 120 may first establish a voiceprint model MD' corresponding to the user USR and the personal electronic device 800 according to the voice-changing voiceprint characteristic signal VPF', after establishing the voiceprint model MD', and then comparing the voice changes.
  • the voiceprint characteristic signal VPF' and the voiceprint model MD' are used for "soundprint matching", and according to the voiceprint matching result, a similarity score SC' is generated, and the similarity score SC' represents the voiceprint characteristic signal VPF' after the voice change.
  • the degree of similarity to the voiceprint model MD' For details of the remaining operations, refer to the aforementioned related paragraphs, and details are not described herein again.
  • the terminal device 120 is not limited to being a computer host, as long as the terminal device 120 is an electronic device (such as a cloud server) that can perform the voiceprint comparison process 50 shown in FIG. 5, or even a mobile electronic device (such as a mobile phone, a tablet computer, etc.). All of them meet the requirements of this application and fall within the scope of this application.
  • the audio processing module is not limited to being disposed in the in-ear device, and the audio processing module may also be disposed in the terminal device.
  • the in-ear device only needs to send the ear canal acoustic signal to the terminal device, and the audio processing module in the terminal device is It is within the scope of the present application to extract the voiceprint features corresponding to the user USR in the ear canal acoustic signal, which also meets the requirements of the present application.
  • the speaker recognition system of the present application uses an in-ear device to receive sound to receive the ear canal sound waves from the user's external ear, and utilizes the audio processing module in the in-ear device to capture the voiceprint characteristics of the user and utilize The terminal device performs voiceprint comparison based on the voiceprint characteristic signal to determine whether the speech end of the speaker recognition system is the user itself.
  • the present application can avoid the risk of being recorded or pirated by a person with a heart.

Abstract

La présente invention concerne un système de reconnaissance de locuteur (10) qui comprend un dispositif intra-auriculaire (100) pour positionnement dans un conduit auditif d'un utilisateur ; et un dispositif de terminal (120). Le dispositif intra-auriculaire (100) comprend : un récepteur de son (102) utilisé pour recevoir une onde acoustique de conduit auditif depuis le conduit auditif de façon à générer un signal acoustique de conduit auditif correspondant à l'onde acoustique de conduit auditif ; et un module de traitement audio (106) connecté au récepteur de son (102) et utilisé pour extraire, à partir du signal acoustique de conduit auditif, une caractéristique d'empreinte vocale correspondant à l'utilisateur de façon à générer un signal de caractéristique d'empreinte vocale. Le dispositif de terminal (120) est utilisé pour déterminer, en fonction du signal de caractéristique d'empreinte vocale, si l'utilisateur est un utilisateur authentifié.
PCT/CN2017/091466 2017-07-03 2017-07-03 Système de reconnaissance de locuteur, procédé de reconnaissance de locuteur et dispositif intra-auriculaire WO2019006587A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201780000606.7A CN110100278B (zh) 2017-07-03 2017-07-03 说话者识别系统及说话者识别方法及入耳式装置
PCT/CN2017/091466 WO2019006587A1 (fr) 2017-07-03 2017-07-03 Système de reconnaissance de locuteur, procédé de reconnaissance de locuteur et dispositif intra-auriculaire

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/091466 WO2019006587A1 (fr) 2017-07-03 2017-07-03 Système de reconnaissance de locuteur, procédé de reconnaissance de locuteur et dispositif intra-auriculaire

Publications (1)

Publication Number Publication Date
WO2019006587A1 true WO2019006587A1 (fr) 2019-01-10

Family

ID=64949595

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/091466 WO2019006587A1 (fr) 2017-07-03 2017-07-03 Système de reconnaissance de locuteur, procédé de reconnaissance de locuteur et dispositif intra-auriculaire

Country Status (2)

Country Link
CN (1) CN110100278B (fr)
WO (1) WO2019006587A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113643707A (zh) * 2020-04-23 2021-11-12 华为技术有限公司 一种身份验证方法、装置和电子设备

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2071856U (zh) * 1990-05-22 1991-02-20 查苕章 耳塞式送受话器
JP2003058190A (ja) * 2001-08-09 2003-02-28 Mitsubishi Heavy Ind Ltd 個人認証方式
CN101042869A (zh) * 2006-03-24 2007-09-26 致胜科技股份有限公司 鼻骨传导活体声纹辨识装置
CN101541238A (zh) * 2007-01-24 2009-09-23 松下电器产业株式会社 活体信息测定装置及其控制方法
JP2010086328A (ja) * 2008-09-30 2010-04-15 Yamaha Corp 認証装置および携帯電話機
CN203984682U (zh) * 2013-11-29 2014-12-03 华北电力大学 一种针对特定对象的助听装置

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1933705A1 (fr) * 2005-10-07 2008-06-25 Koninklijke Philips Electronics N.V. Thermometre auriculaire a identification auriculaire
JP4937661B2 (ja) * 2006-07-31 2012-05-23 ナップエンタープライズ株式会社 携帯個人認証方法及び電子商取引方法
US8622919B2 (en) * 2008-11-17 2014-01-07 Sony Corporation Apparatus, method, and computer program for detecting a physiological measurement from a physiological sound signal
CN102142254A (zh) * 2011-03-25 2011-08-03 北京得意音通技术有限责任公司 基于声纹识别和语音识别的防录音假冒的身份确认方法
US10154818B2 (en) * 2014-12-24 2018-12-18 Samsung Electronics Co., Ltd. Biometric authentication method and apparatus

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2071856U (zh) * 1990-05-22 1991-02-20 查苕章 耳塞式送受话器
JP2003058190A (ja) * 2001-08-09 2003-02-28 Mitsubishi Heavy Ind Ltd 個人認証方式
CN101042869A (zh) * 2006-03-24 2007-09-26 致胜科技股份有限公司 鼻骨传导活体声纹辨识装置
CN101541238A (zh) * 2007-01-24 2009-09-23 松下电器产业株式会社 活体信息测定装置及其控制方法
JP2010086328A (ja) * 2008-09-30 2010-04-15 Yamaha Corp 認証装置および携帯電話機
CN203984682U (zh) * 2013-11-29 2014-12-03 华北电力大学 一种针对特定对象的助听装置

Also Published As

Publication number Publication date
CN110100278B (zh) 2023-09-22
CN110100278A (zh) 2019-08-06

Similar Documents

Publication Publication Date Title
US20210165866A1 (en) Methods, apparatus and systems for authentication
JP5015939B2 (ja) 音響的外耳特徴付けのための方法及び装置
KR102626752B1 (ko) 귀 근접성 검출
US8589167B2 (en) Speaker liveness detection
CN112585676A (zh) 生物特征认证
WO2017069118A1 (fr) Dispositif d'authentification personnelle, procédé d'authentification personnelle, et programme d'authentification personnelle
TW200820218A (en) Portable personal authentication method and electronic business transaction method
TW202141469A (zh) 用於語音使用者介面的入耳式活性偵測
US10896682B1 (en) Speaker recognition based on an inside microphone of a headphone
US11900730B2 (en) Biometric identification
US20230020631A1 (en) Ear canal deformation based continuous user identification system using ear wearables
US20240013789A1 (en) Voice control method and apparatus
JP7120313B2 (ja) 生体認証装置、生体認証方法およびプログラム
Shang et al. Voice liveness detection for voice assistants using ear canal pressure
US11799657B2 (en) System and method for performing biometric authentication
US11069363B2 (en) Methods, systems and apparatus for managing voice-based commands
CN107533415B (zh) 声纹检测的方法和装置
CN110100278B (zh) 说话者识别系统及说话者识别方法及入耳式装置
US20220272131A1 (en) Method, electronic device and system for generating record of telemedicine service
Zhang et al. A phoneme localization based liveness detection for text-independent speaker verification
JP7019765B2 (ja) 個人認証装置、個人認証方法および個人認証プログラム
Zhang et al. A Continuous Liveness Detection System for Text-independent Speaker Verification
Zhang Towards Enhanced Mobile Voice Authentication
FI127920B (en) Online multimodal information transfer method, associated system and device
JP2022070872A (ja) 個人認証装置、個人認証方法および個人認証プログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17916959

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17916959

Country of ref document: EP

Kind code of ref document: A1