EP2747458A1 - Enhanced dynamics processing of streaming audio by source separation and remixing - Google Patents

Enhanced dynamics processing of streaming audio by source separation and remixing Download PDF

Info

Publication number
EP2747458A1
EP2747458A1 EP20130198739 EP13198739A EP2747458A1 EP 2747458 A1 EP2747458 A1 EP 2747458A1 EP 20130198739 EP20130198739 EP 20130198739 EP 13198739 A EP13198739 A EP 13198739A EP 2747458 A1 EP2747458 A1 EP 2747458A1
Authority
EP
European Patent Office
Prior art keywords
audio signal
sound source
source components
hearing
processing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP20130198739
Other languages
German (de)
French (fr)
Other versions
EP2747458B1 (en
Inventor
Kelly Fitz
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Starkey Laboratories Inc
Original Assignee
Starkey Laboratories Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/725,443 external-priority patent/US9485589B2/en
Application filed by Starkey Laboratories Inc filed Critical Starkey Laboratories Inc
Publication of EP2747458A1 publication Critical patent/EP2747458A1/en
Application granted granted Critical
Publication of EP2747458B1 publication Critical patent/EP2747458B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/35Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using translation techniques
    • H04R25/356Amplitude, e.g. amplitude shift or compression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/50Customised settings for obtaining desired overall acoustical characteristics
    • H04R25/505Customised settings for obtaining desired overall acoustical characteristics using digital signal processing

Definitions

  • the present application is a Continuation-in-Part (CIP) of and claims the benefit of priority under 35 U.S.C. ⁇ 120 to U.S. Application Serial No. 12/474,881, filed May 29, 2009 , and titled COMPRESSION AND MIXING FOR HEARING ASSISTANCE DEVICES, which claims the benefit of priority under 35 U.S.C. ⁇ 119(e) to U.S. Provisional Patent Application Serial No. 61/058,101, filed on June 2, 2008 , the benefit of priority of each of which is claimed hereby, and each of which are incorporated by reference herein in its entirety.
  • the present application is related to U.S. Application Serial No. 13/568,618, filed August 7, 2012 , and titled COMPRESSION OF SPACED SOURCES FOR HEARING ASSISTANCE DEVICES, which is incorporated by reference herein in its entirety.
  • This patent application pertains to apparatus and processes enhanced dynamics processing of streaming audio by source separation and remixing for hearing assistance devices.
  • Hearing assistance devices such as hearing aids, include electronic instruments worn in or around the ear that compensate for hearing losses by amplifying and processing sound.
  • the electronic circuitry of the device is contained within a housing that is commonly either placed in the external ear canal and/or behind the ear.
  • Transducers for converting sound to an electrical signal and vice-versa may be integrated into the housing or external to it.
  • Hearing aids may be designed to compensate for such hearing deficits by amplifying received sound in a frequency-specific manner, thus acting as a kind of acoustic equalizer that compensates for the abnormal frequency response of the impaired ear. Adjusting a hearing aid's frequency specific amplification characteristics to achieve a desired level of compensation for an individual patient is referred to as fitting the hearing aid.
  • One common way of fitting a hearing aid is to measure hearing loss, apply a fitting algorithm, and fine-tune the hearing aid parameters.
  • Hearing assistance devices also use a dynamic range adjustment, called dynamic range compression, which controls the level of sound sent to the ear of the patient to normalize the loudness of sound in specific frequency regions.
  • the gain that is provided at a given frequency is controlled by the level of sound in that frequency region (the amount of frequency specificity is determined by the filters in the multiband compression design).
  • compression adjusts the level of a sound at a given frequency such that its loudness is similar to that for a normal hearing person without a hearing aid.
  • the prescription can break down when there are two or more simultaneous sounds in the same frequency region.
  • the two sounds may be at two different levels, and therefore each should receive different gain for each to be perceived at their own necessary loudness. Because only one gain value can be prescribed by the hearing aid, however, at most one sound can receive the appropriate gain, providing the second sound with the less than desired sound level and resulting loudness.
  • FIG. 1 shows the levels of two different sounds out of a filter centered at 1 kHz-in this example, the two sounds are two different speech samples.
  • the samples are overlaid on FIG. 1 and one is in a thick dark line 1 and the second is in a thin line 2.
  • FIG. 2 shows the gains that would be applied to those two different sounds at 1 kHz if they were to be presented to a hypothetical multiband dynamic range compressor. Notice that the ideal gain for each speech sample is different. Again, the samples from the thick dark line 1 are shown in comparison to those of the thin line 2.
  • FIG. 3 shows the two gains from FIG. 1 represented by the thick dark line 1 and the thin line 2, but with a line of intermediate thickness 3 which shows the gain that is applied when the two sounds are mixed together before being sent to the multiband compressor. Notice that when the two sounds are mixed together, neither receives the exact gain that should be prescribed for each separately; in fact, there are times when the gain should be high for one speech sample, but it is low because the gain is controlled by the level of the mix of the two sounds, not the level of each sound individually. This can cause artificial envelope fluctuations in each sound, described as comodulation or cross modulation by Stone and Moore ( Stone, M. A., and Moore, B. C. (2008). "Effects of spectro-temporal modulation changes produced by multi-channel compression on intelligibility in a competing-speech task," J Acoust Soc Am 123, 1063-1076 .)
  • the impact is three-fold: the loudness of that instrument is not normal for the hearing aid listener (it may be too soft, for example), distortion to the temporal envelope of that instrument can occur, and interaural-level difference (ILD) cues for sound source localization and segregation can be distorted, making the perceived auditory image of that instrument fluctuate in a way that was not in the original recording.
  • ILD interaural-level difference
  • an external streaming audio device applies compression or other processing to sources isolated from an audio signal using source separation, and mixes the resulting signals back into the unprocessed audio signal to enhance individual sources while minimizing audible artifacts.
  • Variations of the present system use source separation in a side chain to guide processing of a composite audio signal.
  • Hearing assistance devices include the capability to receive audio from a variety of sources.
  • a hearing assistance device may receive audio or data from a transmitter or streamer from an external device, such as an assistive listening device (ALD).
  • ALD assistive listening device
  • Data such as configuration parameters and telemetry information can be downloaded and/or uploaded to the instruments for the purpose of programming, control and data logging.
  • Audio information can be digitized, packetized and transferred as digital packets to and from the hearing instruments for the purpose of streaming entertainment, carrying on phone conversations, playing announcements, alarms and reminders.
  • music is streamed from an external device to a hearing assistance device using a wireless transmission.
  • Types of wireless transmissions include, but are not limited to, 802.11 (WIFI), Bluetooth or other means of wireless communication with a hearing instrument.
  • Streaming entertainment audio like music and movies can be acoustically dense, with many simultaneous sources and a relatively high degree of dynamic range compression.
  • Conventional hearing aid signal processing may not be able to improve the clarity, intelligibility or sound quality of these signals, and may in fact degrade them by introducing significant cross-source modulation in which strong source drive the compression of weaker sources.
  • Previous solutions to this problem include using more compression channels to reduce the amount of cross-source modulation by reducing the number of frequency components in each compression channel, thereby reducing the likelihood that components from two separate sources would be processed in the same channel.
  • independent processing of components from a single source can impair perceptual fusion by reducing the amount of within-source co-modulation, or common modulation, which promotes perceptual fusion across frequency.
  • an external streaming audio device applies processing (such as compression, in an embodiment) to sources isolated from an audio signal using source separation, and mixes the resulting processed signals back into the unprocessed audio signal to enhance individual sources while minimizing audible artifacts.
  • processing such as compression, in an embodiment
  • Variations of the present system use source separation in a side chain to guide processing of a composite audio signal.
  • Various aspects of the present subject matter apply musical source separation to isolate individual voices and instruments in a mix and apply optimal source-specific gain processing before remixing.
  • a remix is automatically provided that is customized to compensate for the hearing loss of the wearer of a hearing assistance device.
  • each source in a mix receives optimal gain and compression, in a way that is not possible when compression is applied to the entire mixture.
  • the hearing impaired listener is presented with a new mix that is optimized to compensate for their impairment. Because the sources are processed independently, degradations due to cross source modulation are minimized.
  • each source in the mixture is compressed independently, such that each source receives gain that is optimal and appropriate without interference or corruption from other components of the mixture.
  • the present subject matter applies compression to sources isolated from a mixture using source separate techniques, and mixes the compressed sources back into the unprocessed signal to enhance individual sources while minimizing audible artifacts.
  • Various techniques can be used for audio source separation, as shown in the filed of computational auditory scene analysis (CASA).
  • CASA computational auditory scene analysis
  • a method using non-negative matrix factorization is used for source separation.
  • Other methods can be used without departing from the scope of the present subject matter.
  • Available source separation techniques have problems in that they require latency and the sound quality of separated signals is degraded by artifacts.
  • Source separation techniques operate outside of real time, but near enough to real time to run in a streaming device with acceptable latency.
  • the resulting individual sources can be mixed back in with the original signal to mask artifacts and add enhancement without signal degradation or unnatural sounding artifacts.
  • FIG. 9 shows a flow diagram for a streaming audio system, in which an audio signal is separated into component sound sources and compressed before being mixed with the unprocessed audio signal and streamed to a hearing assistance device, according to one embodiment of the present subject matter.
  • Source separation 910 is applied to an incoming signal mixture 902 to obtain separate individual sound source components 904, 906, 908.
  • the separate source components 904, 906, 908 are individually compressed 920 to obtain compressed to obtain compressed source components 924, 926, 928.
  • the compressed sound source components are mixed 930 with the audio signal (incoming signal mixture 902) to produce a mixed audio signal 932.
  • the mixed audio signal 932 is streamed to a hearing assistance device worn by a wearer.
  • the mixed audio signal provides a mix with the isolated sources appropriately compressed or enhanced, while artifacts due to imperfect source separation are masked, according to various embodiments.
  • the processing applied to the isolated source can be conventional hearing aid processing or other processing type.
  • the audio signal 902 can be additionally processed in parallel with the isolated source before remixing, in an embodiment.
  • the audio signal 902 is delayed to compensate for latency in source separation, in various embodiments.
  • source separation can be used in a side chain to guide processing of the composite audio signal 902.
  • the isolated sound sources or characteristics of the isolated sound sources
  • Other types of content- or context-specific processing can be guided by analysis performed on the segregated components, according to various embodiments.
  • This enhancement mitigates artifacts due to imperfect source separation, since the isolated source would be used only for analysis, and would not be mixed back into the processed audio stream.
  • the present subject matter provides improved clarity and sound quality in streamed music and audio, in various embodiments.
  • the audio signals can be mono, stereo or multi-channel in various embodiments.
  • the present subject matter need not be limited to music or streaming audio.
  • this technique can be applied to streamed audio for movies and television, and can leverage multichannel (e.g. 5.1) mixing strategies, such as the mixing of speech to the center channel, to improve the source separation in various embodiments.
  • Other signals can benefit from the present methods without departing from the scope of the present subject matter.
  • FIG. 4 illustrates a system for processing left and right stereo signals from a plurality of sound sources in order to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices.
  • the figure shows separate left 410 and right 420 channels where a plurality of left sound sources 1L, 2L, ... , NL are mixed by mixer 411 to make a composite signal that is compressed using compressor 412 to produce the left output signal LO.
  • FIG. 4 also shows in the right channel 420 a plurality of right sound sources 1R, 2R, ... , NR that are mixed by mixer 421 to make a composite right signal that is compressed by compressor 422 to produce a right signal RO.
  • the separate sound sources can be right and left tracks of individual instruments.
  • the tracks include vocals or other sounds.
  • the system provides compression after the mixing which can result in over-attenuation of desired sounds, which is an undesired side effect of the signal processing.
  • track 1 included bass guitar and track 2 included a lead guitar
  • the louder instrument would dominate the signal strength in the channel at any given time and may result in over-attenuation of the weaker signal when compression is applied to the composite signal.
  • left and right signals are compressed independently, level difference between the left and right output signals LO and RO are compressed, i.e., ILD cues are reduced.
  • FIG. 5 illustrates a system for processing left and right stereo signals from a plurality of sound sources by applying compression before mixing to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices, according to one embodiment of the present subject matter.
  • This embodiment applies compression (512 for the left channel 510 and 522 for the right channel 520) to each signal independently to assist in preserving the ability to mix each signal accordingly (using mixers 510 and 521, respectively).
  • This approach allows each sound source 1L, 2L, ... , NL and 1R, 2R, ... , NL to be added to the composite signal as desired.
  • two or more sound sources are input into the mixer. These may be right and left components of an instrumental input, vocal input, or other sound input.
  • Level difference between the left and right output signals LO and RO are compressed, i.e., ILD cues are reduced, because left and right signals are compressed independently.
  • FIG. 6 shows one embodiment of a signal processor that includes a surround sound synthesizer for producing the surround sound signals from the left and right stereo signals where compression is applied the surround sound signals before mixing to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices according to one embodiment of the present subject matter.
  • a surround sound synthesizer 601 receives a right stereo signal SR and a left stereo signal SL and converts the signals into LS, L, C, R, and RS signals.
  • the HRTFs are not used and the signal passes from the surround sound synthesizer 601 to the compression stages 610R and 610L before being sent to the mixers 611 R and 611L.
  • the signals are processed by right and left head-related transfer functions (HRTFs) 608R and 608L.
  • HRTFs head-related transfer functions
  • the resulting signals are then sent through compression stages 610R and 610L before being sent through mixers 611R and 611L.
  • the resulting outputs RO and LO are used by the hearing assistance device to provide stereo sound reception.
  • Level difference between the left and right output signals LO and RO are compressed, i.e., ILD cues are reduced, because left and right signals are compressed independently.
  • surround sound systems include, but are not limited to Dolby 5.1, 6.1, and 7.1 systems, and the application of HRTFs is optional.
  • the examples provided herein are intended to be demonstrative and not limiting, exclusive, or exhaustive.
  • One advantage of the system of FIG. 6 is that the center channel, which frequently is dominated by vocals, can be separated compressed from the other channels, which are largely dominated by the music. Such compression and mixing avoids cross modulation of gain.
  • the level of compression is commensurate with that found in hearing assistance devices, such as hearing aids. Other levels of compression are possible without departing from the scope of the present subject matter.
  • FIG. 7 shows one embodiment for separating a stereo signal into three channels for a more source-specific compression.
  • the signal for the singer is equally applied to both the left and right channel, centering the perceptual image of the singer.
  • FIG. 7 is one example of how to combine the original channels before compression and how to mix the post-compressed signals back into a stereo signal, but other approaches exist.
  • FIG. 7 shows the left (A+S) signal 701 and the right (B+S) signal 702 applied to multipliers (which multiply by 1 ⁇ 2) and summed by summers to create the CA, CB, and 2CS signals.
  • the CS signal is obtained using multiplier 705.
  • the CA, CB and CS signals are compressed by compressors 706, 708, and 707, respectively, and summed by summers 710 and 712.
  • the resulting outputs are multiplied by 2/3 by multipliers 714 and 715 to provide the compressed left and compressed right signals, as shown in FIG. 7 .
  • multipliers 714 and 715 are multiplied by 2/3 by multipliers 714 and 715 to provide the compressed left and compressed right signals, as shown in FIG. 7 .
  • FIG. 8 represents a general way of isolating a stereo signal into individual components that can then be separately compressed and recombined to create a stereo signal.
  • There are known ways of taking a stereo signal and extracting the center channel in a more complex way than shown in FIG. 8 e.g., U.S. Pat. No. 6,405,163 , and U.S. Patent Application Publication Number 2007/0076902 ).
  • Techniques can also be applied to monaural signals to separate the signal into individual instruments. With either approach, the sounds are separated into individual sound source signals, and each source is compressed; the individually compressed sources are then combined to create either the monaural or stereo signal for listening by the hearing impaired listener.
  • Left stereo signal 801 and right stereo signal 802 are sent through a process 803 that separates individual sound sources. Each source is sent to a compressor 804 and then mixed with mixer 806 to provide left 807 and right 808 stereo signals according to one embodiment of the present subject matter.
  • the present subject matter can be embodied in a number of different applications.
  • the mixing can be performed in a computer programmed to mix the tracks and perform compression as set forth herein.
  • the mixing is done in a fitting system.
  • fitting systems include, but are not limited to, the fitting systems set forth in U.S. Patent Application Ser. No. 11/935,935, filed Nov. 6, 2007 , and entitled: SIMULATED SURROUND SOUND HEARING AID FITTING SYSTEM, the entire specification of which is hereby incorporated by reference in its entirety.
  • the wireless communications can include standard or nonstandard communications.
  • standard wireless communications include link protocols including, but not limited to, BluetoothTM, IEEE 802.11(wireless LANs), 802.15 (WPANs), 802.16 (WiMAX), cellular protocols including, but not limited to CDMA and GSM, ZigBee, and ultra-wideband (UWB) technologies.
  • Such protocols support radio frequency communications and some support infrared communications.
  • the present system is demonstrated as a radio system, it is possible that other forms of wireless communications can be used such as ultrasonic, optical, and others.
  • the standards which can be used include past and present standards. It is also contemplated that future versions of these standards and new future standards may be employed without departing from the scope of the present subject matter.
  • the wireless communications support a connection from other devices.
  • Such connections include, but are not limited to, one or more mono or stereo connections or digital connections having link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface.
  • link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface.
  • link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface.
  • such connections include all past and present link protocols. It is also contemplated that future versions of these protocols and new future standards may be employed without departing from the scope of the present subject matter.
  • Hearing assistance devices typically include an enclosure or housing, a microphone, hearing assistance device electronics including processing electronics, and a speaker or receiver. It is understood that in various embodiments the microphone is optional. It is understood that in various embodiments the receiver is optional. Antenna configurations may vary and may be included within an enclosure for the electronics or be external to an enclosure for the electronics. Thus, the examples set forth herein are intended to be demonstrative and not a limiting or exhaustive depiction of variations.
  • the mixing is done using the processor of the hearing assistance device.
  • that processing can be done by the digital signal processor of the hearing aid or by another set of logic programmed to perform the mixing function provided herein.
  • Other applications and processes are possible without departing from the scope of the present subject matter.
  • the hearing aids referenced in this patent application include a processor.
  • the processor may be a digital signal processor (DSP), microprocessor, microcontroller, other digital logic, or combinations thereof.
  • DSP digital signal processor
  • the processing of signals referenced in this application can be performed using the processor. Processing may be done in the digital domain, the analog domain, or combinations thereof. Processing may be done using subband processing techniques. Processing may be done with frequency domain or time domain approaches. Some processing may involve both frequency and time domain aspects. For brevity, in some examples drawings may omit certain blocks that perform frequency synthesis, frequency analysis, analog-to-digital conversion, digital-to-analog conversion, amplification, and certain types of filtering and processing.
  • the processor is adapted to perform instructions stored in memory which may or may not be explicitly shown.
  • Various types of memory may be used, including volatile and nonvolatile forms of memory.
  • instructions are performed by the processor to perform a number of signal processing tasks.
  • analog components are in communication with the processor to perform signal tasks, such as microphone reception, or receiver sound embodiments (i.e., in applications where such transducers are used).
  • signal tasks such as microphone reception, or receiver sound embodiments (i.e., in applications where such transducers are used).
  • different realizations of the block diagrams, circuits, and processes set forth herein may occur without departing from the scope of the present subject matter.
  • hearing assistance devices including hearing aids, including but not limited to, behind-the-ear (BTE), in-the-ear (ITE), in-the-canal (ITC), receiver-in-canal (RIC), or completely-in-the-canal (CIC) type hearing aids.
  • BTE behind-the-ear
  • ITE in-the-ear
  • ITC in-the-canal
  • RIC receiver-in-canal
  • CIC completely-in-the-canal
  • hearing assistance devices including but not limited to, behind-the-ear (BTE), in-the-ear (ITE), in-the-canal (ITC), receiver-in-canal (RIC), or completely-in-the-canal (CIC) type hearing aids.
  • BTE behind-the-ear
  • ITE in-the-ear
  • ITC in-the-canal
  • RIC receiver-in-canal
  • CIC completely-in-the-canal
  • hearing assistance devices including but not limited to, behind-the-ear (BTE), in
  • the present subject matter can also be used in hearing assistance devices generally, such as cochlear implant type hearing devices and such as deep insertion devices having a transducer, such as a receiver or microphone, whether custom fitted, standard, open fitted or occlusive fitted. It is understood that other hearing assistance devices not expressly stated herein may be used in conjunction with the present subject matter.
  • the apparatus and processes set forth herein may be embodied in digital hardware, analog hardware, and/or combinations thereof. It is also understood that in various embodiments, the apparatus and processes set forth herein may be embodied in hardware, software, firmware, and/or combinations thereof.

Abstract

This application relates to a systems and methods for enhanced dynamics processing of streaming audio by source separation and remixing for hearing assistance devices, according to one example. In one embodiment, an external streaming audio device processes sources isolated from an audio signal using source separation, and mixes the resulting signals back into the unprocessed audio signal to enhance individual sources while minimizing audible artifacts. Variations of the present system use source separation in a side chain to guide processing of a composite audio signal.

Description

    CLAIM OF PRIORITY
  • The present application is a Continuation-in-Part (CIP) of and claims the benefit of priority under 35 U.S.C. § 120 to U.S. Application Serial No. 12/474,881, filed May 29, 2009 , and titled COMPRESSION AND MIXING FOR HEARING ASSISTANCE DEVICES, which claims the benefit of priority under 35 U.S.C. § 119(e) to U.S. Provisional Patent Application Serial No. 61/058,101, filed on June 2, 2008 , the benefit of priority of each of which is claimed hereby, and each of which are incorporated by reference herein in its entirety. The present application is related to U.S. Application Serial No. 13/568,618, filed August 7, 2012 , and titled COMPRESSION OF SPACED SOURCES FOR HEARING ASSISTANCE DEVICES, which is incorporated by reference herein in its entirety.
  • FIELD OF THE INVENTION
  • This patent application pertains to apparatus and processes enhanced dynamics processing of streaming audio by source separation and remixing for hearing assistance devices.
  • BACKGROUND
  • Hearing assistance devices, such as hearing aids, include electronic instruments worn in or around the ear that compensate for hearing losses by amplifying and processing sound. The electronic circuitry of the device is contained within a housing that is commonly either placed in the external ear canal and/or behind the ear. Transducers for converting sound to an electrical signal and vice-versa may be integrated into the housing or external to it.
  • Whether due to a conduction deficit or sensorineural damage, hearing loss in most patients occurs non-uniformly over the audio frequency range, most commonly at high frequencies. Hearing aids may be designed to compensate for such hearing deficits by amplifying received sound in a frequency-specific manner, thus acting as a kind of acoustic equalizer that compensates for the abnormal frequency response of the impaired ear. Adjusting a hearing aid's frequency specific amplification characteristics to achieve a desired level of compensation for an individual patient is referred to as fitting the hearing aid. One common way of fitting a hearing aid is to measure hearing loss, apply a fitting algorithm, and fine-tune the hearing aid parameters.
  • Hearing assistance devices also use a dynamic range adjustment, called dynamic range compression, which controls the level of sound sent to the ear of the patient to normalize the loudness of sound in specific frequency regions. The gain that is provided at a given frequency is controlled by the level of sound in that frequency region (the amount of frequency specificity is determined by the filters in the multiband compression design). When properly used, compression adjusts the level of a sound at a given frequency such that its loudness is similar to that for a normal hearing person without a hearing aid. There are other fitting philosophies, but they all prescribe a certain gain for a certain input level at each frequency. It is well known that the application of the prescribed gain for a given input level is affected by time constants of the compressor. What is less well understood is that the prescription can break down when there are two or more simultaneous sounds in the same frequency region. The two sounds may be at two different levels, and therefore each should receive different gain for each to be perceived at their own necessary loudness. Because only one gain value can be prescribed by the hearing aid, however, at most one sound can receive the appropriate gain, providing the second sound with the less than desired sound level and resulting loudness.
  • This phenomenon is illustrated in the following figures. FIG. 1 shows the levels of two different sounds out of a filter centered at 1 kHz-in this example, the two sounds are two different speech samples. The samples are overlaid on FIG. 1 and one is in a thick dark line 1 and the second is in a thin line 2.
  • FIG. 2 shows the gains that would be applied to those two different sounds at 1 kHz if they were to be presented to a hypothetical multiband dynamic range compressor. Notice that the ideal gain for each speech sample is different. Again, the samples from the thick dark line 1 are shown in comparison to those of the thin line 2.
  • FIG. 3 shows the two gains from FIG. 1 represented by the thick dark line 1 and the thin line 2, but with a line of intermediate thickness 3 which shows the gain that is applied when the two sounds are mixed together before being sent to the multiband compressor. Notice that when the two sounds are mixed together, neither receives the exact gain that should be prescribed for each separately; in fact, there are times when the gain should be high for one speech sample, but it is low because the gain is controlled by the level of the mix of the two sounds, not the level of each sound individually. This can cause artificial envelope fluctuations in each sound, described as comodulation or cross modulation by Stone and Moore (Stone, M. A., and Moore, B. C. (2008). "Effects of spectro-temporal modulation changes produced by multi-channel compression on intelligibility in a competing-speech task," J Acoust Soc Am 123, 1063-1076.)
  • This could be particularly problematic with music and other acoustic sound mixes such as the soundtrack to a Dolby 5.1 movie, where signals of significantly different levels are mixed together with the goal of provided a specific aural experience. If the mix is sent to a compressor and improper gains are applied to the different sounds, then the auditory experience is negatively affected and is not the experience intended by the produce of the sound. In the case of music, the gain for each musical instrument is not correct, and the gain to one instrument might be quite different than it would be if the instrument were played in isolation. The impact is three-fold: the loudness of that instrument is not normal for the hearing aid listener (it may be too soft, for example), distortion to the temporal envelope of that instrument can occur, and interaural-level difference (ILD) cues for sound source localization and segregation can be distorted, making the perceived auditory image of that instrument fluctuate in a way that was not in the original recording..
  • Another example is when the accompanying instrumental tracks in a movie soundtrack have substantial energy then compression can overly reduce the overall level and distort the ILD of the simultaneous vocal tracks, diminishing the ability of the wearer to enjoy the mix of instrumental and vocal sound and even to hear and understand the vocal track. Thus, there is a need in the art for improved compression and mixing systems for hearing assistance devices and for external devices that stream audio to hearing assistance devices.
  • SUMMARY
  • This application relates to a systems and methods for enhanced dynamics processing of streaming audio by source separation and remixing for hearing assistance devices, according to one example. In one embodiment, an external streaming audio device applies compression or other processing to sources isolated from an audio signal using source separation, and mixes the resulting signals back into the unprocessed audio signal to enhance individual sources while minimizing audible artifacts. Variations of the present system use source separation in a side chain to guide processing of a composite audio signal.
  • This Summary is an overview of some of the teachings of the present application and is not intended to be an exclusive or exhaustive treatment of the present subject matter. Further details about the present subject matter are found in the detailed description and the appended claims. The scope of the present invention is defined by the appended claims and their legal equivalents.
  • BRIEF DESCRIPTION OF THE DRAWINGS
    • FIG. 1 shows the levels of two different sounds out of a filter centered at 1 kHz.
    • FIG. 2 shows the gains that would be applied to those two different sounds of FIG. 1 at 1 kHz if they were to be presented to a hypothetical multiband dynamic range compressor.
    • FIG. 3 shows the two gains from FIG. 1 represented by the thick line and the thinner line, but with a line of intermediate thickness which shows the gain that is applied when the two sounds are mixed together before being sent to the multiband compressor.
    • FIG. 4 illustrates a system for processing left and right stereo signals from a plurality of sound sources in order to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices.
    • FIG. 5 illustrates a system for processing left and right stereo signals from a plurality of sound sources by applying compression before mixing to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices according to one embodiment of the present subject matter.
    • FIG. 6 shows one embodiment of a signal processor that includes a surround sound synthesizer for producing the surround sound signals from the left and right stereo signals where compression is applied the surround sound signals before mixing to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices according to one embodiment of the present subject matter.
    • FIG. 7 shows an embodiment where a stereo music signal is processed to separate the center signal from the left-dominant and right-dominant signals in order to compress the center signal separately from the left-dominant and right-dominant signals, according to one embodiment of the present subject matter.
    • FIG. 8 shows an embodiment for separating sounds into component sound sources and compressing each individual sound source before being remixed into the original number of channels, according to one embodiment of the present subject matter.
    • FIG. 9 shows a flow diagram for a streaming audio system, in which an audio signal is separated into component sound sources and compressed before being mixed with the unprocessed audio signal and streamed to a hearing assistance device, according to one embodiment of the present subject matter.
    DETAILED DESCRIPTION
  • The following detailed description of the present invention refers to subject matter in the accompanying drawings which show, by way of illustration, specific aspects and embodiments in which the present subject matter may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the present subject matter. References to "an", "one", or "various" embodiments in this disclosure are not necessarily to the same embodiment, and such references contemplate more than one embodiment. The following detailed description is, therefore, not to be taken in a limiting sense, and the scope is defined only by the appended claims, along with the full scope of legal equivalents to which such claims are entitled.
  • Hearing assistance devices include the capability to receive audio from a variety of sources. For example, a hearing assistance device may receive audio or data from a transmitter or streamer from an external device, such as an assistive listening device (ALD). Data such as configuration parameters and telemetry information can be downloaded and/or uploaded to the instruments for the purpose of programming, control and data logging. Audio information can be digitized, packetized and transferred as digital packets to and from the hearing instruments for the purpose of streaming entertainment, carrying on phone conversations, playing announcements, alarms and reminders. In one embodiment, music is streamed from an external device to a hearing assistance device using a wireless transmission. Types of wireless transmissions include, but are not limited to, 802.11 (WIFI), Bluetooth or other means of wireless communication with a hearing instrument.
  • Streaming entertainment audio like music and movies can be acoustically dense, with many simultaneous sources and a relatively high degree of dynamic range compression. Conventional hearing aid signal processing may not be able to improve the clarity, intelligibility or sound quality of these signals, and may in fact degrade them by introducing significant cross-source modulation in which strong source drive the compression of weaker sources. Previous solutions to this problem include using more compression channels to reduce the amount of cross-source modulation by reducing the number of frequency components in each compression channel, thereby reducing the likelihood that components from two separate sources would be processed in the same channel. However, independent processing of components from a single source can impair perceptual fusion by reducing the amount of within-source co-modulation, or common modulation, which promotes perceptual fusion across frequency. This may facilitate component-specific processing, but not source-specific processing. Moreover, especially in music signals, it is common for several consonant (as opposed to dissonant) source to produce components that are very close in frequency and not resolvable even with a large number of compression channels.
  • The present subject matter relates to a systems and methods for enhanced dynamics processing of streaming audio by source separation and remixing for hearing assistance devices, according to one example. In one embodiment, an external streaming audio device applies processing (such as compression, in an embodiment) to sources isolated from an audio signal using source separation, and mixes the resulting processed signals back into the unprocessed audio signal to enhance individual sources while minimizing audible artifacts. Variations of the present system use source separation in a side chain to guide processing of a composite audio signal.
  • Various aspects of the present subject matter apply musical source separation to isolate individual voices and instruments in a mix and apply optimal source-specific gain processing before remixing. In various embodiments, a remix is automatically provided that is customized to compensate for the hearing loss of the wearer of a hearing assistance device. In one embodiment, each source in a mix receives optimal gain and compression, in a way that is not possible when compression is applied to the entire mixture. The hearing impaired listener is presented with a new mix that is optimized to compensate for their impairment. Because the sources are processed independently, degradations due to cross source modulation are minimized.
  • When applying compression to an audio mixture (or audio signal), each source in the mixture is compressed independently, such that each source receives gain that is optimal and appropriate without interference or corruption from other components of the mixture. The present subject matter applies compression to sources isolated from a mixture using source separate techniques, and mixes the compressed sources back into the unprocessed signal to enhance individual sources while minimizing audible artifacts. Various techniques can be used for audio source separation, as shown in the filed of computational auditory scene analysis (CASA). In one embodiment, a method using non-negative matrix factorization is used for source separation. Other methods can be used without departing from the scope of the present subject matter. Available source separation techniques have problems in that they require latency and the sound quality of separated signals is degraded by artifacts. However, in the case of streamed music, latency constraints are relaxed, and thus signal processing can be done on the external streaming device. Source separation techniques operate outside of real time, but near enough to real time to run in a streaming device with acceptable latency. The resulting individual sources can be mixed back in with the original signal to mask artifacts and add enhancement without signal degradation or unnatural sounding artifacts.
  • FIG. 9 shows a flow diagram for a streaming audio system, in which an audio signal is separated into component sound sources and compressed before being mixed with the unprocessed audio signal and streamed to a hearing assistance device, according to one embodiment of the present subject matter. Source separation 910 is applied to an incoming signal mixture 902 to obtain separate individual sound source components 904, 906, 908. The separate source components 904, 906, 908 are individually compressed 920 to obtain compressed to obtain compressed source components 924, 926, 928. After compressing the components, the compressed sound source components are mixed 930 with the audio signal (incoming signal mixture 902) to produce a mixed audio signal 932. According to various embodiments, the mixed audio signal 932 is streamed to a hearing assistance device worn by a wearer. The mixed audio signal provides a mix with the isolated sources appropriately compressed or enhanced, while artifacts due to imperfect source separation are masked, according to various embodiments. In various embodiments, the processing applied to the isolated source can be conventional hearing aid processing or other processing type. The audio signal 902 can be additionally processed in parallel with the isolated source before remixing, in an embodiment. The audio signal 902 is delayed to compensate for latency in source separation, in various embodiments.
  • According to other embodiments, source separation can be used in a side chain to guide processing of the composite audio signal 902. For example, the isolated sound sources (or characteristics of the isolated sound sources) can guide the tuning of a bank of resonant filters to enhance individual components in the composition signal. Other types of content- or context-specific processing can be guided by analysis performed on the segregated components, according to various embodiments. This enhancement mitigates artifacts due to imperfect source separation, since the isolated source would be used only for analysis, and would not be mixed back into the processed audio stream. The present subject matter provides improved clarity and sound quality in streamed music and audio, in various embodiments. The audio signals can be mono, stereo or multi-channel in various embodiments.
  • The present subject matter need not be limited to music or streaming audio. When combined with appropriate video buffering technology, this technique can be applied to streamed audio for movies and television, and can leverage multichannel (e.g. 5.1) mixing strategies, such as the mixing of speech to the center channel, to improve the source separation in various embodiments. Other signals can benefit from the present methods without departing from the scope of the present subject matter.
  • FIG. 4 illustrates a system for processing left and right stereo signals from a plurality of sound sources in order to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices. The figure shows separate left 410 and right 420 channels where a plurality of left sound sources 1L, 2L, ... , NL are mixed by mixer 411 to make a composite signal that is compressed using compressor 412 to produce the left output signal LO. FIG. 4 also shows in the right channel 420 a plurality of right sound sources 1R, 2R, ... , NR that are mixed by mixer 421 to make a composite right signal that is compressed by compressor 422 to produce a right signal RO. It is understood that the separate sound sources can be right and left tracks of individual instruments. It is also possible that the tracks include vocals or other sounds. The system provides compression after the mixing which can result in over-attenuation of desired sounds, which is an undesired side effect of the signal processing. For example, if track 1 included bass guitar and track 2 included a lead guitar, it is possible that the louder instrument would dominate the signal strength in the channel at any given time and may result in over-attenuation of the weaker signal when compression is applied to the composite signal. Furthermore, because left and right signals are compressed independently, level difference between the left and right output signals LO and RO are compressed, i.e., ILD cues are reduced.
  • FIG. 5 illustrates a system for processing left and right stereo signals from a plurality of sound sources by applying compression before mixing to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices, according to one embodiment of the present subject matter. This embodiment applies compression (512 for the left channel 510 and 522 for the right channel 520) to each signal independently to assist in preserving the ability to mix each signal accordingly (using mixers 510 and 521, respectively). This approach allows each sound source 1L, 2L, ... , NL and 1R, 2R, ... , NL to be added to the composite signal as desired. It is understood that to provide a plurality of sound sources two or more sound sources are input into the mixer. These may be right and left components of an instrumental input, vocal input, or other sound input. Level difference between the left and right output signals LO and RO are compressed, i.e., ILD cues are reduced, because left and right signals are compressed independently.
  • FIG. 6 shows one embodiment of a signal processor that includes a surround sound synthesizer for producing the surround sound signals from the left and right stereo signals where compression is applied the surround sound signals before mixing to produce mixed left and right sound output signals that can be used by left and right hearing assistance devices according to one embodiment of the present subject matter. A surround sound synthesizer 601 receives a right stereo signal SR and a left stereo signal SL and converts the signals into LS, L, C, R, and RS signals. In various embodiments, the HRTFs are not used and the signal passes from the surround sound synthesizer 601 to the compression stages 610R and 610L before being sent to the mixers 611 R and 611L. In various embodiments, the signals are processed by right and left head-related transfer functions (HRTFs) 608R and 608L. The resulting signals are then sent through compression stages 610R and 610L before being sent through mixers 611R and 611L. The resulting outputs RO and LO are used by the hearing assistance device to provide stereo sound reception. Level difference between the left and right output signals LO and RO are compressed, i.e., ILD cues are reduced, because left and right signals are compressed independently. It is understood that other surround sound systems may be employed without departing from the scope of the present subject matter. For example, surround sound systems include, but are not limited to Dolby 5.1, 6.1, and 7.1 systems, and the application of HRTFs is optional. Thus, the examples provided herein are intended to be demonstrative and not limiting, exclusive, or exhaustive.
  • One advantage of the system of FIG. 6 is that the center channel, which frequently is dominated by vocals, can be separated compressed from the other channels, which are largely dominated by the music. Such compression and mixing avoids cross modulation of gain. In various embodiments, the level of compression is commensurate with that found in hearing assistance devices, such as hearing aids. Other levels of compression are possible without departing from the scope of the present subject matter.
  • FIG. 7 shows one embodiment for separating a stereo signal into three channels for a more source-specific compression. Often in music, the signal for the singer is equally applied to both the left and right channel, centering the perceptual image of the singer. Consider the simple example of a stereo music signal with a singer S that is equally in the left and right channel, instrument A that is predominantly in the left channel, and instrument B that is predominantly in the right channel. Then, the left L and right R channels can be described as: L = A + S
    Figure imgb0001
    R = B + S
    Figure imgb0002
  • Then, one can remove the singer from the instruments by subtracting the left from the right channels, and create a signal that is dominated by the singer by adding the left and right channels: L - R = A + S - B + S = A - B
    Figure imgb0003
    L + R = A + S + B + S = A + B + 2 * S
    Figure imgb0004
    CS = L + R / 2 = S + A + B / 2
    Figure imgb0005
  • Thus, one can compress the (L+R)/2 mix to the compressor so that the gain is primarily that for the singer. To get a signal that is primarily instrument A and one that is primarily instrument B: CA = L - R / 2 = A + S - B + S / 2 = A - B - S / 2
    Figure imgb0006
    CB = R - L / 2 = B + S - A + S / 2 = B - A - S / 2
    Figure imgb0007
  • After CS, CL and CR have been individually compressed, they are mixed together to create a stereo channel again: CL = 2 * CS + CA / 3
    Figure imgb0008
    CR = 2 * CS + CB / 3
    Figure imgb0009
  • FIG. 7 is one example of how to combine the original channels before compression and how to mix the post-compressed signals back into a stereo signal, but other approaches exist. FIG. 7 shows the left (A+S) signal 701 and the right (B+S) signal 702 applied to multipliers (which multiply by ½) and summed by summers to create the CA, CB, and 2CS signals. The CS signal is obtained using multiplier 705. The CA, CB and CS signals are compressed by compressors 706, 708, and 707, respectively, and summed by summers 710 and 712. The resulting outputs are multiplied by 2/3 by multipliers 714 and 715 to provide the compressed left and compressed right signals, as shown in FIG. 7. It is understood that this is one example of how to process the signals and that other variations are possible without departing from the scope of the present subject matter. Thus, the system set forth in FIG. 7 is intended to be demonstrative and not exhaustive or exclusive.
  • FIG. 8 represents a general way of isolating a stereo signal into individual components that can then be separately compressed and recombined to create a stereo signal. There are known ways of taking a stereo signal and extracting the center channel in a more complex way than shown in FIG. 8 (e.g., U.S. Pat. No. 6,405,163 , and U.S. Patent Application Publication Number 2007/0076902 ). Techniques can also be applied to monaural signals to separate the signal into individual instruments. With either approach, the sounds are separated into individual sound source signals, and each source is compressed; the individually compressed sources are then combined to create either the monaural or stereo signal for listening by the hearing impaired listener.
  • Left stereo signal 801 and right stereo signal 802 are sent through a process 803 that separates individual sound sources. Each source is sent to a compressor 804 and then mixed with mixer 806 to provide left 807 and right 808 stereo signals according to one embodiment of the present subject matter.
  • It is understood that the present subject matter can be embodied in a number of different applications. In applications involving mixing of music to generate hearing assistance device-compatible stereo signals, the mixing can be performed in a computer programmed to mix the tracks and perform compression as set forth herein. In various embodiments, the mixing is done in a fitting system. Such fitting systems include, but are not limited to, the fitting systems set forth in U.S. Patent Application Ser. No. 11/935,935, filed Nov. 6, 2007 , and entitled: SIMULATED SURROUND SOUND HEARING AID FITTING SYSTEM, the entire specification of which is hereby incorporated by reference in its entirety.
  • Various embodiments of the present subject matter support wireless communications with a hearing assistance device. In various embodiments the wireless communications can include standard or nonstandard communications. Some examples of standard wireless communications include link protocols including, but not limited to, Bluetooth™, IEEE 802.11(wireless LANs), 802.15 (WPANs), 802.16 (WiMAX), cellular protocols including, but not limited to CDMA and GSM, ZigBee, and ultra-wideband (UWB) technologies. Such protocols support radio frequency communications and some support infrared communications. Although the present system is demonstrated as a radio system, it is possible that other forms of wireless communications can be used such as ultrasonic, optical, and others. It is understood that the standards which can be used include past and present standards. It is also contemplated that future versions of these standards and new future standards may be employed without departing from the scope of the present subject matter.
  • The wireless communications support a connection from other devices. Such connections include, but are not limited to, one or more mono or stereo connections or digital connections having link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface. In various embodiments, such connections include all past and present link protocols. It is also contemplated that future versions of these protocols and new future standards may be employed without departing from the scope of the present subject matter.
  • It is understood that variations in communications protocols, antenna configurations, and combinations of components may be employed without departing from the scope of the present subject matter. Hearing assistance devices typically include an enclosure or housing, a microphone, hearing assistance device electronics including processing electronics, and a speaker or receiver. It is understood that in various embodiments the microphone is optional. It is understood that in various embodiments the receiver is optional. Antenna configurations may vary and may be included within an enclosure for the electronics or be external to an enclosure for the electronics. Thus, the examples set forth herein are intended to be demonstrative and not a limiting or exhaustive depiction of variations.
  • In various embodiments, the mixing is done using the processor of the hearing assistance device. In cases where such devices are hearing aids, that processing can be done by the digital signal processor of the hearing aid or by another set of logic programmed to perform the mixing function provided herein. Other applications and processes are possible without departing from the scope of the present subject matter.
  • It is understood that the hearing aids referenced in this patent application include a processor. The processor may be a digital signal processor (DSP), microprocessor, microcontroller, other digital logic, or combinations thereof. The processing of signals referenced in this application can be performed using the processor. Processing may be done in the digital domain, the analog domain, or combinations thereof. Processing may be done using subband processing techniques. Processing may be done with frequency domain or time domain approaches. Some processing may involve both frequency and time domain aspects. For brevity, in some examples drawings may omit certain blocks that perform frequency synthesis, frequency analysis, analog-to-digital conversion, digital-to-analog conversion, amplification, and certain types of filtering and processing. In various embodiments the processor is adapted to perform instructions stored in memory which may or may not be explicitly shown. Various types of memory may be used, including volatile and nonvolatile forms of memory. In various embodiments, instructions are performed by the processor to perform a number of signal processing tasks. In such embodiments, analog components are in communication with the processor to perform signal tasks, such as microphone reception, or receiver sound embodiments (i.e., in applications where such transducers are used). In various embodiments, different realizations of the block diagrams, circuits, and processes set forth herein may occur without departing from the scope of the present subject matter.
  • The present subject matter is demonstrated for hearing assistance devices, including hearing aids, including but not limited to, behind-the-ear (BTE), in-the-ear (ITE), in-the-canal (ITC), receiver-in-canal (RIC), or completely-in-the-canal (CIC) type hearing aids. It is understood that behind-the-ear type hearing aids may include devices that reside substantially behind the ear or over the ear. Such devices may include hearing aids with receivers associated with the electronics portion of the behind-the-ear device, or hearing aids of the type having receivers in the ear canal of the user, including but not limited to receiver-in-canal (RIC) or receiver-in-the-ear (RITE) designs. The present subject matter can also be used in hearing assistance devices generally, such as cochlear implant type hearing devices and such as deep insertion devices having a transducer, such as a receiver or microphone, whether custom fitted, standard, open fitted or occlusive fitted. It is understood that other hearing assistance devices not expressly stated herein may be used in conjunction with the present subject matter.
  • It is understood that in various embodiments, the apparatus and processes set forth herein may be embodied in digital hardware, analog hardware, and/or combinations thereof. It is also understood that in various embodiments, the apparatus and processes set forth herein may be embodied in hardware, software, firmware, and/or combinations thereof.
  • This application is intended to cover adaptations and variations of the present subject matter. It is to be understood that the above description is intended to be illustrative, and not restrictive. The scope of the present subject matter should be determined with reference to the appended claim, along with the full scope of legal equivalents to which the claims are entitled.

Claims (15)

  1. A method, comprising:
    isolating individual sound source components from a composite audio signal;
    independently processing the individual sound source components;
    providing delay to the composite audio signal to compensate for latency in source separation;
    mixing the processed sound source components with the delayed composite audio signal to produce a mixed audio signal to minimize audible artifacts; and
    providing the mixed audio signal to a wearer using a receiver in the hearing assistance device worn by the wearer.
  2. The method of claim 1, wherein independently processing the individual sound source components includes independently compressing the individual sound source components.
  3. The method of claim 1, wherein independently processing the individual sound source components includes applying source-specific gain processing to the individual sound source components.
  4. The method of any of the preceding claims, wherein mixing the compressed sound source components includes providing a mix that is customized to compensate for the wearer's hearing loss.
  5. The method of any of the preceding claims, wherein isolating individual sound source components from an audio signal includes using non-negative matrix factorization.
  6. The method of any of the preceding claims, further comprising processing the audio signal in parallel with the isolated individual source components before mixing.
  7. The method of claim 6, wherein processing the audio signal includes compensating for the wearer's hearing loss.
  8. The method of any of the preceding claims, wherein the processing includes using an external device and wherein providing the mixed audio signal includes streaming the mixed audio signal to the hearing assistance device.
  9. A system, comprising:
    at least one hearing assistance device adapted to receive a streaming input;
    and
    an external device, including:
    a processor configured to:
    process a composite audio signal to isolate individual sound source components;
    independently process the individual sound source components; and
    mix the processed sound source components with the composite audio signal to produce a mixed audio signal to minimize audible artifacts due to source separation; and
    a wireless transmitter connected to the processor, the wireless transmitter configured to steam the mixed audio signal to the at least one hearing assistance device.
  10. The system of claim 9, wherein the hearing assistance device includes a signal processor.
  11. The system of claim 10, wherein the signal processor includes a digital signal processor (DSP).
  12. The system of any of claim 9 through claim 11, wherein the at least one hearing assistance device includes a cochlear implant.
  13. The system of any of claim 9 through claim 11, wherein the at least one hearing assistance device includes a hearing aid.
  14. The system of claim 13, wherein the hearing aid includes an in-the-ear (ITE) hearing aid.
  15. The system of claim 13, wherein the hearing aid includes a behind-the-ear (BTE) hearing aid.
EP13198739.8A 2012-12-21 2013-12-20 Enhanced dynamics processing of streaming audio by source separation and remixing Active EP2747458B1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/725,443 US9485589B2 (en) 2008-06-02 2012-12-21 Enhanced dynamics processing of streaming audio by source separation and remixing

Publications (2)

Publication Number Publication Date
EP2747458A1 true EP2747458A1 (en) 2014-06-25
EP2747458B1 EP2747458B1 (en) 2015-09-16

Family

ID=50070269

Family Applications (1)

Application Number Title Priority Date Filing Date
EP13198739.8A Active EP2747458B1 (en) 2012-12-21 2013-12-20 Enhanced dynamics processing of streaming audio by source separation and remixing

Country Status (2)

Country Link
EP (1) EP2747458B1 (en)
DK (1) DK2747458T3 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105489227A (en) * 2014-10-06 2016-04-13 奥迪康有限公司 Hearing device comprising a low-latency sound source separation unit
WO2017017612A1 (en) * 2015-07-29 2017-02-02 Cochlear Limited Wireless communication in an implantable medical device system
EP3599775A1 (en) * 2018-07-27 2020-01-29 Mimi Hearing Technologies GmbH Systems and methods for processing an audio signal for replay on stereo and multi-channel audio devices

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102470429B1 (en) * 2019-03-14 2022-11-23 붐클라우드 360 인코포레이티드 Spatial-Aware Multi-Band Compression System by Priority

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6405163B1 (en) 1999-09-27 2002-06-11 Creative Technology Ltd. Process for removing voice from stereo recordings
US20020078817A1 (en) * 2000-12-06 2002-06-27 Toshihiko Date Music-signal compressing/decompressing apparatus
US20070076902A1 (en) 2005-09-30 2007-04-05 Aaron Master Method and Apparatus for Removing or Isolating Voice or Instruments on Stereo Recordings
WO2011100802A1 (en) * 2010-02-19 2011-08-25 The Bionic Ear Institute Hearing apparatus and method of modifying or improving hearing

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6405163B1 (en) 1999-09-27 2002-06-11 Creative Technology Ltd. Process for removing voice from stereo recordings
US20020078817A1 (en) * 2000-12-06 2002-06-27 Toshihiko Date Music-signal compressing/decompressing apparatus
US20070076902A1 (en) 2005-09-30 2007-04-05 Aaron Master Method and Apparatus for Removing or Isolating Voice or Instruments on Stereo Recordings
WO2011100802A1 (en) * 2010-02-19 2011-08-25 The Bionic Ear Institute Hearing apparatus and method of modifying or improving hearing

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
BÜCHLER M ET AL: "Sound classification in hearing aids inspired by auditory scene analysis", EURASIP JOURNAL OF APPLIED SIGNAL PROCESSING, HINDAWI PUBLISHING CO., CUYAHOGA FALLS, OH, US, vol. 2005, no. 18, 15 October 2005 (2005-10-15), pages 2991 - 3002, XP002610407, ISSN: 1110-8657, [retrieved on 20101119], DOI: 10.1155/ASP.2005.2991 *
STONE, M. A.; MOORE, B. C.: "Effects of spectro-temporal modulation changes produced by multi-channel compression on intelligibility in a competing-speech task", J ACOUST SOC AM, vol. 123, 2008, pages 1063 - 1076

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105489227A (en) * 2014-10-06 2016-04-13 奥迪康有限公司 Hearing device comprising a low-latency sound source separation unit
EP3007467A1 (en) * 2014-10-06 2016-04-13 Oticon A/s A hearing device comprising a low-latency sound source separation unit
US10341785B2 (en) 2014-10-06 2019-07-02 Oticon A/S Hearing device comprising a low-latency sound source separation unit
WO2017017612A1 (en) * 2015-07-29 2017-02-02 Cochlear Limited Wireless communication in an implantable medical device system
CN107921267A (en) * 2015-07-29 2018-04-17 科利耳有限公司 Wireless communication in implanted medical equipment system
US10207117B2 (en) 2015-07-29 2019-02-19 Cochlear Limited Wireless communication in an implantable medical device system
US10857365B2 (en) 2015-07-29 2020-12-08 Cochlear Limited Wireless communication in an implantable medical device system
CN107921267B (en) * 2015-07-29 2021-12-21 科利耳有限公司 Wireless communication in implantable medical device system
US11433245B2 (en) 2015-07-29 2022-09-06 Cochlear Limited Wireless communication in an implantable medical device system
US11951315B2 (en) 2015-07-29 2024-04-09 Cochlear Limited Wireless communication in an implantable medical device system
EP3599775A1 (en) * 2018-07-27 2020-01-29 Mimi Hearing Technologies GmbH Systems and methods for processing an audio signal for replay on stereo and multi-channel audio devices

Also Published As

Publication number Publication date
EP2747458B1 (en) 2015-09-16
DK2747458T3 (en) 2015-11-09

Similar Documents

Publication Publication Date Title
US9924283B2 (en) Enhanced dynamics processing of streaming audio by source separation and remixing
US9185500B2 (en) Compression of spaced sources for hearing assistance devices
EP2131610B1 (en) Compression and mixing for hearing assistance devices
US9161137B2 (en) Mixing of in-the-ear microphone and outside-the-ear microphone signals to enhance spatial perception
US10880659B2 (en) Providing and transmitting audio signal
EP3220661A1 (en) A method for predicting the intelligibility of noisy and/or enhanced speech and a binaural hearing system
US20080205659A1 (en) Method for improving spatial perception and corresponding hearing apparatus
EP1566077A1 (en) Equalisation of the output in a stereo widening network
EP2858382A1 (en) System and method for selective harmonic enhancement for hearing assistance devices
EP3255902A1 (en) Method and apparatus for improving speech intelligibility in hearing devices using remote microphone
EP2747458B1 (en) Enhanced dynamics processing of streaming audio by source separation and remixing
US10313805B2 (en) Binaurally coordinated frequency translation in hearing assistance devices
WO2010148169A1 (en) Spatial audio object coding (saoc) decoder and postprocessor for hearing aids
EP2696599B1 (en) Compression of spaced sources for hearing assistance devices
EP3016408B1 (en) Compressor architecture for avoidance of cross-modulation in remote microphones
WO2022043906A1 (en) Assistive listening system and method
DK2437521T4 (en) Method for frequency compression with harmonic correction and corresponding apparatus

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20131220

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

INTG Intention to grant announced

Effective date: 20150430

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

Ref country code: CH

Ref legal event code: NV

Representative=s name: R.A. EGLI AND CO, PATENTANWAELTE, CH

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 750612

Country of ref document: AT

Kind code of ref document: T

Effective date: 20151015

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602013003117

Country of ref document: DE

REG Reference to a national code

Ref country code: DK

Ref legal event code: T3

Effective date: 20151102

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 3

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20151217

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20151216

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 750612

Country of ref document: AT

Kind code of ref document: T

Effective date: 20150916

REG Reference to a national code

Ref country code: NL

Ref legal event code: FP

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160116

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20151231

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160118

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602013003117

Country of ref document: DE

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20151220

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

26N No opposition filed

Effective date: 20160617

REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20151220

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 4

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20161111

Year of fee payment: 4

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20131220

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

REG Reference to a national code

Ref country code: NL

Ref legal event code: MM

Effective date: 20170101

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20170101

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20171220

REG Reference to a national code

Ref country code: FR

Ref legal event code: ST

Effective date: 20180831

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180102

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20150916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20171220

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: CH

Payment date: 20211125

Year of fee payment: 9

P01 Opt-out of the competence of the unified patent court (upc) registered

Effective date: 20230610

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20221231

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20221231

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DK

Payment date: 20231208

Year of fee payment: 11

Ref country code: DE

Payment date: 20231128

Year of fee payment: 11