EP2620940A1 - Procédé et dispositif d'écoute pour l'évaluation d'un composant de la voix - Google Patents

Procédé et dispositif d'écoute pour l'évaluation d'un composant de la voix Download PDF

Info

Publication number
EP2620940A1
EP2620940A1 EP12196705.3A EP12196705A EP2620940A1 EP 2620940 A1 EP2620940 A1 EP 2620940A1 EP 12196705 A EP12196705 A EP 12196705A EP 2620940 A1 EP2620940 A1 EP 2620940A1
Authority
EP
European Patent Office
Prior art keywords
microphone
hearing
ear
signals
phase difference
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP12196705.3A
Other languages
German (de)
English (en)
Inventor
Vaclav Bouse
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sivantos Pte Ltd
Original Assignee
Siemens Medical Instruments Pte Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Medical Instruments Pte Ltd filed Critical Siemens Medical Instruments Pte Ltd
Publication of EP2620940A1 publication Critical patent/EP2620940A1/fr
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/30Monitoring or testing of hearing aids, e.g. functioning, settings, battery power
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/43Electronic input selection or mixing based on input signal analysis, e.g. mixing or selection between microphone and telecoil or between microphones with different directivity characteristics

Definitions

  • the present invention relates to a method of estimating a constituent of a listener's own voice. Moreover, the present invention relates to a hearing device in which a corresponding method is implemented. Furthermore, the present invention relates to a hearing device having a filter, which was created according to the above method.
  • a hearing device is here understood to mean any device which can be worn in or on the ear and produces a sound stimulus, in particular a hearing device, a headset, headphones and the like.
  • Hearing aids are portable hearing aids that are used to care for the hearing impaired.
  • different types of hearing aids such as behind-the-ear hearing aids (BTE), hearing aid with external receiver (RIC: receiver in the canal) and in-the-ear hearing aids (ITE), e.g. Concha hearing aids or canal hearing aids (ITE, CIC).
  • BTE behind-the-ear hearing aids
  • RIC hearing aid with external receiver
  • ITE in-the-ear hearing aids
  • ITE in-the-ear hearing aids
  • ITE in-the-ear hearing aids
  • ITE concha hearing aids or canal hearing aids
  • the hearing aids listed by way of example are worn on the outer ear or in the ear canal.
  • bone conduction hearing aids, implantable or vibrotactile hearing aids are also available on the market. The stimulation of the damaged hearing takes place either mechanically or electrically.
  • Hearing aids have in principle as essential components an input transducer, an amplifier and an output transducer.
  • the input transducer is usually a sound receiver, z. As a microphone, and / or an electromagnetic receiver, for. B. an induction coil.
  • the output transducer is usually used as an electroacoustic transducer, z. As miniature speaker, or as an electromechanical transducer, z. B. bone conduction, realized.
  • the amplifier is usually integrated in a signal processing unit. This basic structure is in FIG. 1 shown using the example of a behind-the-ear hearing aid. In a hearing aid housing 1 for carrying behind the ear, one or more microphones 2 for receiving the sound from the environment are installed.
  • a signal processing unit 3 which is also integrated in the hearing aid housing 1, processes the microphone signals and amplifies them.
  • the output signal of the signal processing unit 3 is transmitted to a loudspeaker or earpiece 4, which outputs an acoustic signal.
  • the sound is optionally transmitted via a sound tube, which is fixed with an earmold in the ear canal, to the eardrum of the device carrier.
  • the power supply of the hearing device and in particular the signal processing unit 3 is effected by a likewise integrated into the hearing aid housing 1 battery. 5
  • CASA Computer Auditory Scene Analysis
  • This CASA principle is therefore based on the fact that the current hearing situation is analyzed mathematically.
  • the CASA principle is based on the ASA principle, its main achievements in the work of Bregman, AS (1994): "Auditory Scene Analysis: The Perceptual Organization of Sound", Bradford Books summarized , The state of developments regarding CASA is reproduced in the article Wang, D., Brown, GJ (2006): Computational Auditory Scene Analysis: Principals, Algorithms and Applications, John Wiley & Sons Publishing, ISBN 978-0-471-74109-1 ,
  • Monaural CASA algorithms work on a single signal channel and try to separate the sources. At least, language should be separated. In most cases, they are based on very strict conditions with regard to the sound sources. One of these conditions relates, for example, to the fundamental frequency estimate. In addition, the monaural CASA algorithms are in principle unable to use the spatial information of a signal.
  • Multi-channel algorithms attempt to separate the signals based on the spatial locations of the sources.
  • the configuration of the microphones is essential. For example, in binaural configuration, i. H. if the microphones are on both sides of the head but no reliable source separation is done with these algorithms.
  • the object of the present invention is to be able to recognize the own voice of a wearer of a hearing device more reliably.
  • a hearing device for carrying out the above method, wherein the hearing device has the two microphones and a signal processing device for transforming, segmenting and grouping.
  • two microphones are placed very special.
  • the second microphone is placed in the ear canal, while the first microphone is placed essentially at the ear canal exit or outside the ear canal (eg in the Concha or on the Pinna).
  • the microphone arranged in the auditory canal can record significantly more portions of sound which enter the auditory canal via bone conduction than the external microphone.
  • characteristic information based on one's own voice can be obtained.
  • a CASA algorithm can then own voice, so the voice of the wearer of the hearing in which the CASA algorithm is running, reliably estimate or extract.
  • At least one further feature different from the phase difference and the level difference is obtained for each of the microphone signals and used for the segmentation and / or grouping.
  • grouping is possible in principle only on the basis of the phase difference and the level difference is, it is beneficial to use at least one additional feature for grouping. In principle, other features may be better suited for segmentation.
  • the further feature may relate to a change or a rate of change in the spectrum of the microphone signals.
  • This has the advantage that, for example, rapid level increases (ON-SETs) can be well recognized at certain frequencies. Such signal edges are suitable for segmentation.
  • the further feature may also include the harmonic (degree of the acoustic periodicity) or the correlation of the two microphone signals. Harmonicity makes it easier to directly recognize language components.
  • the correlation has the advantage that a correlate between the externally audible language and the language transmitted via bone conduction can additionally be used for the reliable determination of one's own language.
  • the hearing apparatus adapted to estimate a constituent of the own language may comprise a filter controlled by the grouping or corresponding grouping information of the signal processing means.
  • the regions in the time-frequency plane defined by the grouping are then used to extract corresponding signal portions, which are then presumably derived from their own voice.
  • the segmentation and grouping technique may be used as needed, e.g. B. be repeated every time the hearing aid. This has the advantage that the filter can then be constantly adapted to the current conditions (eg seat of the hearing aid in or on the ear).
  • a hearing device which has a filter which is used for extracting the own voice of a wearer of the hearing device and filters out signal components belonging to the group of regions fall, which were obtained in advance by a method described above.
  • the difference to the previous hearing device is therefore that the filter no longer has to be variable and thus is less expensive to produce.
  • the hearing device can be embodied as an in-the-ear hearing device.
  • the hearing device may also be embodied as a behind-the-ear hearing device which has a hearing aid housing for carrying behind the ear and an external earpiece for carrying in the auditory canal or a sound tube for transmitting sound from the hearing device housing into the auditory canal
  • Microphone is arranged on the external handset or the sound tube and the first microphone in the hearing aid housing.
  • FIG. 2 schematically shows an ear canal 10 with eardrum 11, wherein an ITE hearing aid 12 is inserted into the ear canal 10.
  • an ITE hearing aid 12 is inserted into the ear canal 10.
  • the outer ear 13 At the exit of the ear canal 10 is the outer ear 13, which is not completely drawn here.
  • the hearing device 12 has a side 14 facing the eardrum 11 and an outwardly directed side 15 facing away from the eardrum.
  • the hearing device 12 has a first microphone 16 on the outwardly facing side 15. This microphone 16 is only symbolically drawn outside the hearing device 12. In fact, however, the microphone is regularly in the hearing aid or at least on the surface of the hearing aid.
  • the first microphone 16 supplies a microphone signal m 1 .
  • This first microphone signal m 1 is used for the CASA algorithm described below.
  • it is also the usual signal processing device 17 of the hearing aid 12 is provided.
  • This conventional signal processing device 17 often includes an amplifier.
  • the output signal of the signal processing device 17 is forwarded to a loudspeaker or receiver 18 which is arranged on the side 14 of the hearing device 12 facing the eardrum 11. Also, he is symbolically located outside of the hearing aid 12 here and is mostly in the hearing aid housing.
  • the hearing device or the hearing device 12 has here in addition to the first microphone 16, a second microphone 19.
  • This second microphone 19 is also located on the eardrum 11 facing side 14 of the hearing aid 12. It thus absorbs sound, which is in the room between the hearing aid 12, the eardrum 11 and the wall of the ear canal 10 results. In this often closed room especially the sound of one's own voice becomes over bone conduction entered.
  • the second microphone 19 picks up this sound and provides a second microphone signal m 2 in the hearing device 12.
  • This second microphone 19 may be referred to as an in-channel microphone.
  • the CASA system 20 provides an estimate ⁇ for a component of its own language.
  • FIG. 4 is the CASA system 20 of FIG. 3 shown in detail.
  • the two microphone signals m 1 and m 2 are supplied to an analysis unit 21.
  • This analysis unit 21 examines each of the microphone signals m 1 and m 2 for specific characteristics. For this purpose, the time signals m 1 and m 2 are transformed into the time-frequency range, resulting in so-called "tf signals", which can also be referred to as short-time spectra. The transformation can be performed by a high-resolution filter bank.
  • features are then extracted for each frequency channel of each of the two microphone signals m 1 and m 2 .
  • phase difference and the level difference between the two microphone signals m 1 and m 2 ie in particular the phase and level difference in each point of the tf-level of the tf signals.
  • further features can be extracted from the microphone signals m 1 and m 2 by the analysis device 21.
  • One of these additional features may relate to so-called "on-sets". This includes, for example, rapid changes in a spectrum, which typically arise at the beginning of a vowel. Such on-sets usually represent steep edges in a tf diagram and are suitable for segmenting the tf signals.
  • Another feature extracted by analyzer 21 may be the harmonics, which is understood to mean the degree of acoustic periodicity.
  • the harmonicity is often used to recognize language.
  • the correlation of the microphone signals m 1 and m 2 can be examined. In particular, for example, the correlation between the sound transmitted via the bone conduction into the auditory canal and the sound brought in from the outside to the ear can be analyzed. This also gives hints to your own language.
  • the analysis device 21 is followed on the output side by a segmentation device 22.
  • the segmentation device 22 calculates limits around signal components in the tf plane in such a way that according to FIG FIG. 5 Regions 24 are defined.
  • These regions 24 are tf signal components of a single sound source.
  • the regions 24 in the tf plane for individual sources can be calculated in several known ways. Regions that can be assigned to a defined source therefore contain a source sound component 25. Outside of these regions 24 there are interference sound components 26 that can not be assigned to a specific source. At the time of the segmentation, however, it is not yet known which region 24 belongs to which specific source.
  • regions 24 in the tf plane are formed for both microphone signals m 1 and m 2 .
  • the segmentation device 22 is followed by a grouping device 23.
  • the segmented signal components ie the signal components 25 in the regions 24, are organized into signal streams which are assigned to the different sound sources. In the present case, only those signal components are synthesized into a signal stream which belong to the own language of the hearing device wearer.
  • arbitrary regions 24 of the tf-plane can be combined with each other.
  • the information of the phase difference and the level difference is used.
  • the phase difference and the level difference of the two microphone signals in a model must be estimated beforehand. Based on these estimates, it can then be determined whether or not one of the segmented regions belongs to one's own vote. If the determined phase and level difference lie within a predetermined tolerance range around the estimated phase and level difference, then the region in question is counted to its own voice.
  • a region 24 is grouped with one or more multiple regions 24 is dependent on the phase difference and the level difference between the two microphone signals m 1 and m 2 .
  • the additional features listed above can also be used. So a group thus formed represents all parts of a short-term spectrum, which are to be joined together in order to gain only the own language or own voice from the variety of sound components. The remaining signal components in the short-term spectrum should namely be suppressed.
  • the grouping device 23 forwards the corresponding grouping information to a filter 27 in the CASA system 20.
  • the filter 27 is thus controlled or parameterized with the grouping information.
  • the filter 27 receives the temporal microphone signals m 1 and m 2 , filters the two signals and derives an estimate for their own voice or a component ⁇ of their own voice.
  • the filter can use the signal components of the regions 24, unfortunately, tf signals of the two microphones or only those of one of the tf signals from a microphone to reconstruct their own voice.
  • the system according to the invention typically has a processing delay of a few 100 ms. This delay is necessary to extract the features and group the regions. However, this delay is not a problem in practice.
  • FIG. 5 is a further embodiment with respect to the hardware structure of a hearing aid according to the invention shown.
  • the hearing aid is a BTE hearing aid whose main component 28 is worn behind the ear, in particular behind the pinna 29.
  • This BTE hearing aid has at the main component 28, a first microphone 30.
  • the hearing aid here has a so-called external handset, which is mounted in the ear canal 32. Together with this external handset 31, a second microphone 33 is also mounted in the ear canal 32.
  • the inventive extracting or estimating a component of one's own voice can also be used in a BTE hearing aid.
  • the CASA principle can also be used for the first time to register or extract one's own voice, since due to the special placement of the microphones, sufficient spatial information of the signals is now available. From this spatial information corresponding grouping information can be obtained, so that ultimately can be dispensed with complicated language models.

Landscapes

  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Neurosurgery (AREA)
  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Headphones And Earphones (AREA)
EP12196705.3A 2012-01-19 2012-12-12 Procédé et dispositif d'écoute pour l'évaluation d'un composant de la voix Withdrawn EP2620940A1 (fr)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
DE102012200745.8A DE102012200745B4 (de) 2012-01-19 2012-01-19 Verfahren und Hörvorrichtung zum Schätzen eines Bestandteils der eigenen Stimme

Publications (1)

Publication Number Publication Date
EP2620940A1 true EP2620940A1 (fr) 2013-07-31

Family

ID=47594360

Family Applications (1)

Application Number Title Priority Date Filing Date
EP12196705.3A Withdrawn EP2620940A1 (fr) 2012-01-19 2012-12-12 Procédé et dispositif d'écoute pour l'évaluation d'un composant de la voix

Country Status (3)

Country Link
US (1) US20130188816A1 (fr)
EP (1) EP2620940A1 (fr)
DE (1) DE102012200745B4 (fr)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102060949B1 (ko) * 2013-08-09 2020-01-02 삼성전자주식회사 청각 기기의 저전력 운용 방법 및 장치
EP3522569A1 (fr) * 2014-05-20 2019-08-07 Oticon A/s Dispositif auditif
US9843873B2 (en) 2014-05-20 2017-12-12 Oticon A/S Hearing device
US10299049B2 (en) 2014-05-20 2019-05-21 Oticon A/S Hearing device
US9897378B2 (en) 2015-10-08 2018-02-20 Nyc Designed Inspirations Llc Cosmetic makeup sponge/blender container
DK3550858T3 (da) 2015-12-30 2023-06-12 Gn Hearing As Et på hovedet bærbart høreapparat
WO2017147428A1 (fr) 2016-02-25 2017-08-31 Dolby Laboratories Licensing Corporation Capture et extraction de propre signal vocal
SE542485C2 (en) 2017-01-03 2020-05-19 Earin Ab Charging and storage of wireless earbuds
WO2018128577A2 (fr) * 2017-01-03 2018-07-12 Earin Ab Écouteurs sans fil, et capsule de stockage et de charge associée

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2405673A1 (fr) * 2010-07-07 2012-01-11 Siemens Medical Instruments Pte. Ltd. Procédé de localisation d'un source audio et d'un système auditif à plusieurs canaux

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7099821B2 (en) * 2003-09-12 2006-08-29 Softmax, Inc. Separation of target acoustic signals in a multi-transducer arrangement
DE102005032274B4 (de) * 2005-07-11 2007-05-10 Siemens Audiologische Technik Gmbh Hörvorrichtung und entsprechendes Verfahren zur Eigenstimmendetektion
US8175291B2 (en) * 2007-12-19 2012-05-08 Qualcomm Incorporated Systems, methods, and apparatus for multi-microphone based speech enhancement

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2405673A1 (fr) * 2010-07-07 2012-01-11 Siemens Medical Instruments Pte. Ltd. Procédé de localisation d'un source audio et d'un système auditif à plusieurs canaux

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ARTIKEL WANG, D.; BROWN, G. J.: "Computational Auditory Scene Analysis: Principals, Algorithms and Applications", 2006, JOHN WILEY & SONS-VERLAG
BREGMAN, A. S.: "Auditory Scene Analysis: The Perceptual Organisation of Sound", BRADFORD BOOKS ZUSAMMENGEFASST. DER STAND DER ENTWICKLUNGEN BZGL, 1994
SHAO Y ET AL: "A computational auditory scene analysis system for speech segregation and robust speech recognition", COMPUTER SPEECH AND LANGUAGE, ELSEVIER, LONDON, GB, vol. 24, no. 1, 1 January 2010 (2010-01-01), pages 77 - 93, XP026545650, ISSN: 0885-2308, [retrieved on 20080328], DOI: 10.1016/J.CSL.2008.03.004 *
VACLAV BOUSE ET AL: "A model-based auditory scene analysis approach and its application to speech source localization", ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2011 IEEE INTERNATIONAL CONFERENCE ON, IEEE, 22 May 2011 (2011-05-22), pages 2624 - 2627, XP032001350, ISBN: 978-1-4577-0538-0, DOI: 10.1109/ICASSP.2011.5947023 *

Also Published As

Publication number Publication date
DE102012200745A1 (de) 2013-07-25
DE102012200745B4 (de) 2014-05-28
US20130188816A1 (en) 2013-07-25

Similar Documents

Publication Publication Date Title
DE102012200745B4 (de) Verfahren und Hörvorrichtung zum Schätzen eines Bestandteils der eigenen Stimme
EP3451705B1 (fr) Procédé et dispositif de reconnaissance rapide de voix propre
EP2603018B1 (fr) Dispositif auditif avec détection d'activité de locuteur et procédé de fonctionnement d'un dispositif auditif
EP2405673B1 (fr) Procédé de localisation d'un source audio et système auditif à plusieurs canaux
DE102010026884B4 (de) Verfahren zum Betreiben einer Hörvorrichtung mit zweistufiger Transformation
EP2645743B1 (fr) Appareil auditif pour un traitement binaural et procédé destiné à préparer un traitement binaural
EP1962556A2 (fr) Procédé d'amélioration de la prise de conscience spatiale et dispositif auditif correspondant
EP1983801A2 (fr) Procédé d'adaptation d'un système d'appareil auditif binaural
DE102013215131A1 (de) Verfahren zur Verfolgung einer Schallquelle
DE102007035171A1 (de) Verfahren zum Anpassen eines Hörgeräts mit Hilfe eines perzeptiven Modells
EP2109329A2 (fr) Procédé d'évaluation en plusieurs étapes destiné à la réduction des bruits parasites et dispositif auditif
EP2495724B1 (fr) Procédé et dispositif destinés à l'évaluation d'un bruit parasite
EP1912470B1 (fr) Procédé pour la compression dynamique d'un signal audio et appareil auditif correspondant
EP2584795B1 (fr) Procédé de détermination d'une ligne caractéristique de compression
EP1962554A2 (fr) Appareil auditif doté d'une séparation de signal d'erreur et procédé correspondant
EP2658289B1 (fr) Procédé de commande d'une caractéristique de guidage et système auditif
EP2434781A1 (fr) Procédé de reconstruction d'un signal vocal et dispositif auditif
EP2219389B1 (fr) Dispositif et procédé d'estimation des bruits parasites dans l'arrivée d'un appareil auditif binauréal
EP2023667A2 (fr) Procédé de réglage d'un système auditif doté d'un modèle perceptif pour oreilles binaurales et système auditif correspondant
EP2982136B1 (fr) Procédé d'estimation d'un signal utile et dispositif auditif
DE102013207080B4 (de) Binaurale Mikrofonanpassung mittels der eigenen Stimme
DE102010011730A1 (de) Hörvorrichtung und Verfahren zum Erzeugen einer omnidirektionalen Richtcharakteristik
EP2622879B1 (fr) Procédé et dispositif de compression de fréquence
EP1971186A2 (fr) Procédé de réduction de bruits parasites à l'aide de modèles pouvant être entraînés

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

17P Request for examination filed

Effective date: 20130719

RBV Designated contracting states (corrected)

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20140201