EP3479373A1 - Systèmes et procédés d'amélioration de la capacité auditive - Google Patents

Systèmes et procédés d'amélioration de la capacité auditive

Info

Publication number
EP3479373A1
EP3479373A1 EP17820919.3A EP17820919A EP3479373A1 EP 3479373 A1 EP3479373 A1 EP 3479373A1 EP 17820919 A EP17820919 A EP 17820919A EP 3479373 A1 EP3479373 A1 EP 3479373A1
Authority
EP
European Patent Office
Prior art keywords
hearing assistance
assistance device
audio signals
signal
hearing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP17820919.3A
Other languages
German (de)
English (en)
Other versions
EP3479373A4 (fr
Inventor
Steven MANSER
Daniel S. Keller
Drew DUNDAS
Edward V. Bacho
Rodney C. Perkins
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
K/S Himpp
Original Assignee
K/S Himpp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US15/432,830 external-priority patent/US10631108B2/en
Application filed by K/S Himpp filed Critical K/S Himpp
Publication of EP3479373A1 publication Critical patent/EP3479373A1/fr
Publication of EP3479373A4 publication Critical patent/EP3479373A4/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/55Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
    • H04R25/554Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired using a wireless connection, e.g. between microphone and amplifier or using Tcoils
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K11/00Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
    • G10K11/18Methods or devices for transmitting, conducting or directing sound
    • G10K11/26Sound-focusing or directing, e.g. scanning
    • G10K11/34Sound-focusing or directing, e.g. scanning using electrical steering of transducer arrays, e.g. beam steering
    • G10K11/341Circuits therefor
    • G10K11/346Circuits therefor using phase variation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K15/00Acoustics not otherwise provided for
    • G10K15/02Synthesis of acoustic waves
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/43Electronic input selection or mixing based on input signal analysis, e.g. mixing or selection between microphone and telecoil or between microphones with different directivity characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/39Aspects relating to automatic logging of sound environment parameters and the performance of the hearing aid during use, e.g. histogram logging, or of user selected programs or settings in the hearing aid, e.g. usage logging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/41Detection or adaptation of hearing aid parameters or programs to listening situation, e.g. pub, forest
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/55Communication between hearing aids and external devices via a network for data exchange
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/07Use of position data from wide-area or local-area positioning systems in hearing devices, e.g. program or information selection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/55Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
    • H04R25/558Remote control, e.g. of amplification, frequency

Definitions

  • the present disclosure relates generally to personalized sound delivery and hearing systems, and more specifically, relates to a method and an active, adaptive system for the enhancement of speech clarity at a distance or in environments with interfering background noise through the combination of audio inputs from more than one hearing device.
  • Hearing assistance devices such as hearing aids
  • hearing assistance devices include, but are not limited to, devices for use in the ear, in the ear canal, completely in the canal, and behind the ear.
  • Such devices have been developed to ameliorate the effects of hearing losses in individuals.
  • Hearing deficiencies can range from deafness to hearing losses where the individual has impairment responding to different frequencies of sound or to being able to differentiate sounds occurring simultaneously.
  • the hearing assistance device in its most elementary form usually provides for auditory correction through the amplification and filtering of sound provided in the environment with the intent that previously inaudible sounds become audible while maintaining comfort for more intense sounds while the device is worn.
  • Hearing aids employ different forms of amplification to achieve improved hearing.
  • ANC active noise cancellation
  • US2012/0250916 and WO2005052911 both relate to a hearing aid which can perform active noise cancellation.
  • the hearing aid includes a signal processor which produces a compensation/cancellation signal that can attenuate acoustic signals that bypasses the signal path of the hearing aid and enters the ear canal.
  • the hearing aid includes two microphones and a control unit provided for adjusting a time delay of the two microphone signals.
  • WO06003618 relates to an earplug with a circuit for active noise cancellation.
  • a noise signal is received in the earplug, a cancelling signal is processed by means of the circuit to cancel the noise signal.
  • U.S. Pat. No. 6,567,524 teaches a hearing protective earplug with an audio communication terminal for obtaining speech signals of high quality while attenuating noise. The earplug performs noise attenuation automatically adapted to the noise conditions and communication modes.
  • U.S. Pat. No. 6, 181,801 and U.S. Pat. No. 6,021,207 relate to a communications earpiece which receives audio signals, wired and wireless, respectively, sent from an external device such as a mobile phone. Ambient sounds are used for noise cancellation.
  • the communications earpiece can be used by both hearing impaired and non-hearing impaired users.
  • ANC anti-phasic, but equal amplitude sound is generated in the ear canal in an effort to cause destructive interference and thereby negate the presence of specific sounds.
  • An example would be found in noise cancelling earphones, wherein low frequency, periodic noises like multi-talker babble or aircraft engine noise can be identified, their level and spectral content in the ear canal (or under a headphone) can be captured and analyzed.
  • the phase path to the eardrum predicted, and a signal with equal intensity and frequency composition but 180° out of phase with the ambient sound is added, causing a reduction or elimination of the undesirable sound at the eardrum.
  • ANCs typically require a microphone located in the ear canal to capture and analyze the magnitude and phase of the signal, and an algorithm present in the system to allow for the process.
  • Various systems and methods are disclosed herein to combine electric audio signals from more than one hearing assistance device in order to enhance the ability to communicate effectively under adverse listening conditions through a process of adaptive constructive interference.
  • multiple users equipped with their respective hearing assistance devices may have the ability to connect to the individual devices of other local users in order to create a local area network of users who can then share acoustic information via a radio link for easier communication.
  • the hearing assistance devices of both users collectively receive more electric audio signals than a single hearing assistance device receives. Therefore, incorporating the electric audio signals with the acoustic signal from the vicinity of the user's device allows for enhancement of the desired sounds such as conversational speech relative to the background of less desirable sound.
  • electric audio signals received generated at the second hearing assistance device are transmitted to the first hearing assistance device and encoded as a replacement for input to a microphone channel located at the first hearing assistance device and combined with the signal captured by the second microphone on the first hearing device.
  • the combined electric audio signals are processed through a digital signal processor in the local hearing assistance device and adaptively time-shifted in order to enhance the amplitude of the desired signal via constructive interference.
  • the combined signal is then routed through a multi-channel compressor in the local hearing assistance device to apply gain and frequency shaping as appropriate for the hearing profile of the listener, converted to an analog signal in a digital-analog converter, and then transduced and output to the user' s ear.
  • a third hearing assistance device may be configured to receive electric audio signals from both microphones located at the first and second hearing assistance devices.
  • the electric audio signal from both microphones located at the first and second hearing assistance devices is combined and encoded as a replacement for input to a microphone channel located at the third hearing assistance device.
  • the electric audio signal received from the microphone at the third hearing assistance device may be combined with the electric audio signals received from both microphones located at the first and second hearing assistance devices.
  • electric audio signals from the second and third hearing assistance devices can be transmitted to the first hearing assistance device.
  • electric audio signals from the first and third hearing assistance devices can be transmitted to the second hearing assistance device. This approach can be expanded to encompass larger numbers of devices.
  • each hearing assistance device employs the other's microphone
  • each hearing assistance device is enabled to exploit the spatial differences between the two hearing assistance devices and environment noise.
  • this technique can be employed in non-acoustic hearing enhancement devices such as cochlear implants, bone anchored hearing devices, contact hearing devices, implantable hearing devices and in other methods of stimulating the auditory system.
  • a system may include a hearing assistance device (e.g. hearing aid) and an integrated control, for instance a mobile phone or computing device that is wirelessly linked to the sound delivery device.
  • the microphones on the sound delivery device and/or the associated mobile device may then detect sound levels at different locations at different times.
  • Significant data can be collected from different users to be aggregated and uploaded to a server for analysis to determine the current or average sound levels at particular establishments.
  • a database can be created that includes sound level and characteristic information for different cities, restaurants, sporting venues, public transportation, and others places.
  • sound level ratings may then be aggregated by a server in a database and accessed by users through an application on their mobile device, website, or through integration into websites that combine feedback on establishments.
  • GPS data can be tagged to the sound level data recorded through the microphones to identify the location.
  • signal-to-noise ratio, and time-stamp data maybe tagged to the sound level.
  • users can tag or indicate which establishment they are attending after receiving notifications on their mobile device or any combination of the tagging methods for confirmation.
  • Figure 1 is an overview of an exemplary hearing assistance device in accordance with the present disclosure.
  • Figure 2 is a schematic block diagram of an exemplary hearing assistance device in accordance with the present disclosure.
  • Figure 3 is a flow chart illustrating a process for combining electric audio signals from more than one hearing assistance device that incorporates an active signal enhancement system in accordance with the present disclosure.
  • Figure 4 is an overview of an example hearing system in accordance with the present disclosure.
  • Figure 5 is an overview of an example hearing system where multiple users and hearing assistance devices are employed in accordance with the present disclosure.
  • FIG. 1 is a schematic block diagram of an exemplary hearing assistance device 100 that includes a hearing aid circuitry 101, an active signal enhancement system 102, and a wireless transmitter and receiver 103.
  • the wireless transmitter and receiver 103 is connected to an interface module 104.
  • the hearing assistance device 100 is connected to at least one other hearing assistance device to establish a secure, bidirectional, full duplex connection via the wireless transmitter and receiver 103.
  • the wireless transmitter and receiver 103 is configured to send and receive data via Near Field Communication (NFC), WiFi®, Bluetooth®, another wireless technology to other hearing assistance devices, mobile devices, tablets, laptops, computers, etc. Connection between the hearing assistance device 100 and other hearing assistance devices can be facilitated at the interface module 104.
  • the interface module 104 may include for example, mobile devices, tablets, laptops, computers, etc., equipped with a user interface.
  • the hearing assistance device 100 may connect to another hearing device via Bluetooth® pairing, facilitated by a user located at a mobile device, also paired with the hearing assistance device 100.
  • the wireless transmitter and receiver 103 may include a short-range wireless communication system, commonly known as Near Field Communication (NFC).
  • NFC Near Field Communication
  • the NFC circuit is a two-way communication circuit comprising both a transmitter and a receiver.
  • NFC is a mainly inductive communication system, which has a very short effective transmission range, such as approximately 5-6 centimeters. The protocol used for wireless near field communication via NFC is well described.
  • NFC enables a user to simplify the otherwise cumbersome Bluetooth® pairing procedure by temporarily bringing a mobile device comprising both Bluetooth® and NFC circuitry within the effective NFC transmission range of 5-6 centimeters of another device, which also comprises both Bluetooth® and NFC circuitry, e.g. an after-market headset, and then let the NFC circuits automatically exchange information between the devices, in order to perform the Bluetooth® pairing procedure of the two devices.
  • the two paired devices are separated again but will now be able to communicate via Bluetooth® or other method of transmitting and receiving digital audio signals, e.g. streaming audio to the head-set from the mobile phone and vice versa, as long as they are within the Bluetooth® communication range.
  • the NFC is used where the two devices are close together for exchange of critical pairing information, and after separation of the devices, it is no longer used for any communication between the devices.
  • the hearing aid circuitry 101 can include a signal path comprising one input transducer 203, e.g. a microphone.
  • the input transducer 203 can be pointed outward towards the ambient space surrounding the hearing device user.
  • the input transducer 203 can convert an ambient sound entering the ear of the user from the ambient space to an electric signal. Even though one input transducer is shown in the figure, it is understood that there can be more than one input transducer and more than one signal path.
  • the input transducer 203 generates electric audio signals, and sends the electric audio signals to a wireless transmitter and receiver 103 to send to a second hearing aid (not pictured).
  • the second hearing aid (not pictured) can send electric audio signals received from its input transducer to the current wireless transmitter and receiver 103.
  • the electric signal is communicated to the wireless transmitter and receiver 103, where it can be sent to another hearing assistance device.
  • electric audio signals can be received from the wireless transmitter and receiver 103.
  • electric audio signals from the transducer 203 can be sent to a gain stage 204 in which the electric audio signals are amplified. From the gain stage 204, the signal is communicated to an analog-to-digital (A/D) converter 205, which converts the amplified analog electric signal to a digital signal. The signal is then processed through an active signal enhancement unit 212.
  • A/D analog-to-digital
  • DSP digital signal-processing
  • the digital electric signal is communicated to a multi-channel compressor 202 in the local hearing assistance device to reduce the digital noise.
  • the digital noise reduction processing is applied to avoid amplifying undesirable, random environmental noise.
  • the enhancement of desirable signals occurs by the combination of the signal obtained at a remote location with the signal obtained at the local location. If the signals are appropriately time shifted to align the peaks of the signal, the result is a more robust representation of the desired signal, and theoretically no enhancement of the less desirable competing noise.
  • the combined signal is then routed to a digital signal-processing (DSP) unit 206 to apply gain and frequency shaping as appropriate for the hearing profile of the listener, converted to an analog signal in a digital-analog converter 207, and then transduced and output to the user.
  • DSP digital signal-processing
  • the digital signal-processing (DSP) unit 206 is adapted to process the digital electric signal in accordance with a desired correction of the hearing loss specific for the user of the hearing device.
  • the digital electric signal is communicated to a digital-to-analogue (D/A) converter 207, which converts the digital electric signal to an analog pulse density modulated (PDM) electric signal.
  • the analog electric signal is communicated to a multiplexer 208, and then to a low output impedance output driver 209.
  • the analog PDM electric signal can be communicated to an output transducer 210, e.g. a speaker within the ear canal, which converts the electric signal to a sound pressure signal affecting the tympanic membrane in the residual volume of the ear canal (not shown).
  • Figure 2 also exemplifies the active signal enhancement unit 212 of the hearing assistance device 100.
  • the signal enhancement system 102 improves signal to noise ratio by increasing the amplitude of the desirable signal via proximity and constructive interference while relatively preserving the undesired background noise (which is typically assumed to be diffuse and effectively random). Digital noise reduction processing is applied later in the chain to attempt to avoid amplifying undesirable, random environmental noise.
  • the enhancement of desirable signals occurs by the combination of the signal obtained at a remote location with the signal obtained at the local location. If the signals are appropriately time shifted to align the peaks of the signal, the result is a more robust representation of the desired signal, and theoretically no enhancement of the less desirable competing noise.
  • the signal enhancement unit 212 may be configured to enhance speech clarity at a distance or in environments with interfering background noise through the combination of audio inputs from more than one hearing device. Noise may be unwanted audio signals that disturb the hearing device user.
  • the remote microphone is able to capture the desirable signals and the effect of time aligning peaks in the two or more signals results in an enhancement of signal magnitude. Therefore, the speech intelligibility of sound sources that are at a distance from the talker and the listener (e.g., in between or to the side of the two) is enhanced, as the signal is enhanced via constructive interference and proximity effects, while the noise, being random, is not.
  • both the electric audio signals from the microphone located at the first hearing assistance device and the electric audio signals received from the connected one or more hearing assistance devices are combined.
  • the combined electric audio signals are processed through a digital signal processor in the hearing assistance device to enhance the amplitude of the desired signal via constructive interference, at step 308.
  • the magnitude of the desirable signal is increased via proximity and constructive interference while relatively preserving the undesired background noise.
  • the combined signal is then routed through a multi-channel compressor in the local hearing assistance device to reduce the digital noise.
  • the digital noise reduction processing is applied to avoid amplifying undesirable, random environmental noise.
  • the enhancement of desirable signals occurs by the combination of the signal obtained at a remote location with the signal obtained at the local location.
  • the combined signal is then routed to a digital signal-processing (DSP) unit to apply gain and frequency shaping as appropriate for the hearing profile of the listener, converted to an analog signal in a digital-analog converter, and then transduced and output to the user.
  • DSP digital signal-processing
  • FIG. 4 illustrates an overview of an example hearing system 300 according to the present disclosure.
  • the system 300 may include a first hearing assistance device 100 associated with a first user 305 and connected to a first mobile device 310 linked with the hearing assistance device 100 using antennas 315, and a personal profile 125 associated with the first user 305 may be stored optionally on the mobile device 310 or elsewhere (e.g., server 330 via connection over the Internet or other communications network).
  • the hearing assistance device 100 may include a charging case that can store the audio data, and the audio data may be uploaded to a computer (for instance for users without a mobile device) which could then upload data over the network 355 to a server 330.
  • a network 355 may also link the mobile device 310 and/or hearing assistance device 100 to a server 330 and database 350 that stores personal profiles, including software for analysis of sound data and performing other functions as disclosed herein.
  • other users 307 with one or more other hearing assistance devices 100 may also be linked to the network 355 and server 330 and sound/hearing data from the other users 307 may be aggregated and stored in the database 350.
  • a clinician 340 may be connected to the network 355 via a computing device 335 to allow the user to diagnose and make changes to the settings of the hearing assistance devices 100. The changes made by clinicians 340 may also be stored in the database 350 for separate or combined reference.
  • hearing assistance device 100 associated with first user 305 may be in communication, such as via radio link, with one or more other hearing assistance devices 100 associated with one or more other users 307 and related hearing/audio signal data may be shared among the devices.
  • Other users 307 also may have other mobile devices, such as smart phones, connected respectively to associated hearing assistance device 100.
  • the hearing system 300 efficiently optimizes the hearing assistance device 100 in certain environments based on an accumulation of data from both the user 305 and other users 307, and in some cases the clinician 340. This accumulated data can be utilized to present the user options or automatically set the audio settings on a user's hearing assistance device 100. Furthermore, this data may be utilized by clinicians to evaluate certain settings and improve their recommended settings for a given user 305 and noise environment.
  • a database could be created that includes sound level and characteristic information for different cities, restaurants, sporting venues, public transportation, and others places. These sound level ratings may then be aggregated by a server in a database and accessed by users through an application on their mobile device, website, or through integration into websites that combine feedback on establishments.
  • GPS data can be tagged to the sound level data recorded through the microphones to identify the location.
  • signal-to-noise ratio, and time- stamp data maybe tagged to the sound level.
  • users can tag or indicate which establishment they are attending after receiving notifications on their mobile device or any combination of the tagging methods for confirmation.
  • FIG. 5 illustrates an overview of an example hearing system 400 where multiple users and hearing assistance devices are employed.
  • user 305 equipped with a hearing assistance device 100 may be configured to receive electric audio signals from both microphones 502 and 503 located at users 308 and 309 hearing assistance devices 100.
  • the electric audio signal from both microphones 502 and 503 located at the users 308 and 309 hearing assistance devices 100 are encoded as a replacement for input to a microphone 501 channel located at the hearing assistance device 100 associated with user 305.
  • the electric audio signal received from microphone 501 at the hearing assistance device 100 associated with user 305 may be combined with the electric audio signals received from microphones 502 and 503 located at the hearing assistance devices 100 associated with users 308 and 309.
  • the disclosure herein may be implemented with any type of hardware and/or software, and may be a pre-programmed general purpose computing device.
  • the system may be implemented using a server, a personal computer, a portable computer, a thin client, or any suitable device or devices.
  • the disclosure and/or components thereof may be a single device at a single location, or multiple devices at a single, or multiple, locations that are connected together using any appropriate communication protocols over any communication medium such as electric cable, fiber optic cable, or in a wireless manner.
  • the computing system can include clients and servers.
  • a client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
  • a server transmits data (e.g., an HTML page) to a client device (e.g., for purposes of displaying data to and receiving user input from a user interacting with the client device).
  • client device e.g., for purposes of displaying data to and receiving user input from a user interacting with the client device.
  • Data generated at the client device e.g., a result of the user interaction
  • Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back-end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front-end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components.
  • the components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network ("LAN”) and a wide area network (“WAN”), an internetwork (e.g., the Internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks).
  • LAN local area network
  • WAN wide area network
  • Internet internetwork
  • peer-to-peer networks e
  • Implementations of the subj ect matter and the operations described in this specification can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them.
  • Implementations of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions, encoded on computer storage medium for execution by, or to control the operation of, data processing apparatus.
  • the program instructions can be encoded on an artificially-generated propagated signal, e.g., a machine- generated electrical, optical, or electromagnetic signal that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus.
  • a computer storage medium can be, or be included in, a computer-readable storage device, a computer-readable storage substrate, a random or serial access memory array or device, or a combination of one or more of them.
  • a computer storage medium is not a propagated signal, a computer storage medium can be a source or destination of computer program instructions encoded in an artificially-generated propagated signal.
  • the computer storage medium can also be, or be included in, one or more separate physical components or media (e.g., multiple CDs, disks, or other storage devices).
  • the term "data processing apparatus” encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, a system on a chip, or multiple ones, or combinations, of the foregoing apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
  • the apparatus can also include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, a cross-platform runtime environment, a virtual machine, or a combination of one or more of them.
  • the apparatus and execution environment can realize various different computing model infrastructures, such as web services, distributed computing and grid computing infrastructures.
  • a computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, object, or other unit suitable for use in a computing environment.
  • a computer program may, but need not, correspond to a file in a file system.
  • a program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub-programs, or portions of code).
  • a computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
  • the processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform actions by operating on input data and generating output.
  • the processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
  • processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer.
  • a processor will receive instructions and data from read-only memory or random access memory or both.
  • the essential elements of a computer are a processor for performing actions in accordance with instructions and one or more memory devices for storing instructions and data.
  • a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks.
  • mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks.
  • a computer need not have such devices.
  • a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device (e.g., a universal serial bus (USB) flash drive), to name just a few.
  • Devices suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
  • the processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.

Abstract

L'invention concerne divers systèmes et procédés pour augmenter la qualité du son et l'intelligibilité de la parole délivrée à un utilisateur en combinant des signaux audio électriques provenant de plus d'un dispositif d'aide auditive comportant un système d'amélioration de signal actif. Le procédé consiste à recevoir des signaux audio électriques au niveau d'un premier dispositif d'aide auditive, et à envoyer des signaux audio électriques à un second dispositif d'aide auditive. Le signal audio électrique est codé en tant que remplacement destiné à être entré dans un canal de microphone situé au niveau du second dispositif d'aide auditive. Les signaux audio électriques provenant du microphone situé au niveau du premier dispositif d'aide auditive et du microphone situé au niveau du second dispositif d'aide auditive sont combinés.
EP17820919.3A 2016-07-01 2017-06-19 Systèmes et procédés d'amélioration de la capacité auditive Withdrawn EP3479373A4 (fr)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201662357469P 2016-07-01 2016-07-01
US15/432,830 US10631108B2 (en) 2016-02-08 2017-02-14 Hearing augmentation systems and methods
PCT/US2017/038075 WO2018005140A1 (fr) 2016-07-01 2017-06-19 Systèmes et procédés d'amélioration de la capacité auditive

Publications (2)

Publication Number Publication Date
EP3479373A1 true EP3479373A1 (fr) 2019-05-08
EP3479373A4 EP3479373A4 (fr) 2020-02-12

Family

ID=60785531

Family Applications (1)

Application Number Title Priority Date Filing Date
EP17820919.3A Withdrawn EP3479373A4 (fr) 2016-07-01 2017-06-19 Systèmes et procédés d'amélioration de la capacité auditive

Country Status (2)

Country Link
EP (1) EP3479373A4 (fr)
WO (1) WO2018005140A1 (fr)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10284998B2 (en) 2016-02-08 2019-05-07 K/S Himpp Hearing augmentation systems and methods
US10390155B2 (en) 2016-02-08 2019-08-20 K/S Himpp Hearing augmentation systems and methods
US10631108B2 (en) 2016-02-08 2020-04-21 K/S Himpp Hearing augmentation systems and methods
US10341791B2 (en) 2016-02-08 2019-07-02 K/S Himpp Hearing augmentation systems and methods
US10433074B2 (en) 2016-02-08 2019-10-01 K/S Himpp Hearing augmentation systems and methods

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5479522A (en) * 1993-09-17 1995-12-26 Audiologic, Inc. Binaural hearing aid
US7120258B1 (en) * 1999-10-05 2006-10-10 Able Planet, Inc. Apparatus and methods for mitigating impairments due to central auditory nervous system binaural phase-time asynchrony
AU2001278418A1 (en) * 2000-07-14 2002-01-30 Gn Resound A/S A synchronised binaural hearing system
EP1448022A1 (fr) * 2003-02-14 2004-08-18 GN ReSound A/S Compression dynamique dans une prothèse auditive
DE102004047759B3 (de) * 2004-09-30 2006-06-01 Siemens Audiologische Technik Gmbh Verwendung eines Hörhilfegerätesystems mit wenigstens zwei Hörhilfegeräten
US7688991B2 (en) * 2006-05-24 2010-03-30 Phonak Ag Hearing assistance system and method of operating the same
US8515109B2 (en) * 2009-11-19 2013-08-20 Gn Resound A/S Hearing aid with beamforming capability
DK2629551T3 (en) * 2009-12-29 2015-03-02 Gn Resound As Binaural hearing aid system
US8379871B2 (en) * 2010-05-12 2013-02-19 Sound Id Personalized hearing profile generation with real-time feedback
EP2744226A1 (fr) * 2012-12-17 2014-06-18 Oticon A/s Instrument auditif
EP2871857B1 (fr) * 2013-11-07 2020-06-17 Oticon A/s Système d'assistance auditive biauriculaire comprenant deux interfaces sans fil
US9641945B2 (en) * 2014-09-28 2017-05-02 Ubdevice Corp. Bluetooth hearing aid system
WO2015001135A2 (fr) * 2014-11-03 2015-01-08 Phonak Ag Procédé d'assistance auditive utilisant un flux de diffusion audio

Also Published As

Publication number Publication date
EP3479373A4 (fr) 2020-02-12
WO2018005140A1 (fr) 2018-01-04

Similar Documents

Publication Publication Date Title
US10631108B2 (en) Hearing augmentation systems and methods
US9949040B2 (en) Peer to peer hearing system
US10327071B2 (en) Head-wearable hearing device
EP3039882B1 (fr) Aide à la conversation
US8675884B2 (en) Method and a system for processing signals
EP3479373A1 (fr) Systèmes et procédés d'amélioration de la capacité auditive
EP3337190B1 (fr) Procédé de réduction de bruit dans un dispositif de traitement audio
US8144891B2 (en) Earphone set
DK2835986T3 (en) Hearing aid with input transducer and wireless receiver
EP3386216B1 (fr) Système auditif comprenant un estimateur de niveau binaural et/ou de gain, et une méthode correspondante
CN101437190B (zh) 具有助听功能的耳机
US10529358B2 (en) Method and system for reducing background sounds in a noisy environment
US20230254649A1 (en) Method of detecting a sudden change in a feedback/echo path of a hearing aid
CN101437189A (zh) 具有助听功能的无线耳机
EP3072314B1 (fr) Un procédé pour l'exploitation d'un system auditif pour l'établissement de coups de télépone ainsi qu'un system auditif correspondant
US8824668B2 (en) Communication system comprising a telephone and a listening device, and transmission method
KR20170040137A (ko) 환경 적응 청각 신호의 수신을 위한 청력 보완 기기 및 환경 적응 청각 신호의 수신 방법
Choi et al. Implementation of Digital Hearing Aid Using Bluetooth Audio Digital Signal Processor
KR20160023078A (ko) 환경 적응 청각 신호의 수신을 위한 청력 보완 기기 및 환경 적응 청각 신호의 수신 방법
Luo et al. Benefits of compression amplification in telephone bluetooth-assistive listening devices for people with hearing loss
CN112911480A (zh) 一种助听器声音放大方法及助听器
Lee et al. Recent trends in hearing aid technologies

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20180915

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
A4 Supplementary search report drawn up and despatched

Effective date: 20200115

RIC1 Information provided on ipc code assigned before grant

Ipc: H04R 25/02 20060101ALI20200109BHEP

Ipc: H04R 25/00 20060101ALI20200109BHEP

Ipc: G10K 11/175 20060101AFI20200109BHEP

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20201125

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20210407