EP0555278A1 - Zweifach-sprachprozessor - Google Patents

Zweifach-sprachprozessor

Info

Publication number
EP0555278A1
EP0555278A1 EP91918663A EP91918663A EP0555278A1 EP 0555278 A1 EP0555278 A1 EP 0555278A1 EP 91918663 A EP91918663 A EP 91918663A EP 91918663 A EP91918663 A EP 91918663A EP 0555278 A1 EP0555278 A1 EP 0555278A1
Authority
EP
European Patent Office
Prior art keywords
aid
processor
speech
signal
hearing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP91918663A
Other languages
English (en)
French (fr)
Other versions
EP0555278A4 (en
Inventor
Gary John Dooley
Peter John Blamey
Graeme Milbourne Clark
Peter Misha Seligman
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Melbourne
Cochlear Ltd
Original Assignee
University of Melbourne
Cochlear Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Melbourne, Cochlear Ltd filed Critical University of Melbourne
Publication of EP0555278A1 publication Critical patent/EP0555278A1/de
Publication of EP0555278A4 publication Critical patent/EP0555278A4/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/55Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
    • H04R25/552Binaural
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61NELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
    • A61N1/00Electrotherapy; Circuits therefor
    • A61N1/18Applying electric currents by contact electrodes
    • A61N1/32Applying electric currents by contact electrodes alternating or intermittent currents
    • A61N1/36Applying electric currents by contact electrodes alternating or intermittent currents for stimulation
    • A61N1/36036Applying electric currents by contact electrodes alternating or intermittent currents for stimulation of the outer, middle or inner ear
    • A61N1/36038Cochlear stimulation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61NELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
    • A61N1/00Electrotherapy; Circuits therefor
    • A61N1/02Details
    • A61N1/04Electrodes
    • A61N1/05Electrodes for implantation or insertion into the body, e.g. heart electrode
    • A61N1/0526Head electrodes
    • A61N1/0541Cochlear electrodes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/43Signal processing in hearing aids to enhance the speech intelligibility
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/35Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using translation techniques
    • H04R25/356Amplitude, e.g. amplitude shift or compression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/60Mounting or interconnection of hearing aid parts, e.g. inside tips, housings or to ossicles
    • H04R25/604Mounting or interconnection of hearing aid parts, e.g. inside tips, housings or to ossicles of acoustic or vibrational transducers
    • H04R25/606Mounting or interconnection of hearing aid parts, e.g. inside tips, housings or to ossicles of acoustic or vibrational transducers acting directly on the eardrum, the ossicles or the skull, e.g. mastoid, tooth, maxillary or mandibular bone, or mechanically stimulating the cochlea, e.g. at the oval window

Definitions

  • the present invention relates to improvements in the processing of sound for the purposes of supplying an information signal to either an acoustic hearing aid, a cochlear implant aid device or both so as to improve the quality of hearing of a patient.
  • an acoustic hearing aid is reference to an aid of the type adapted to fit in or adjacent an ear of a patient and which provides an acoustic output suitable to at least partially compensate for hearing deficiencies of the patient.
  • a cochlear implant aid will refer to a device which includes components which are fitted within the body of a patient and which are adapted to electrically stimulate the nervous system of a patient in order to at least partially compensate for usually profound hearing loss of the patient.
  • cochlear implants to patients with some residual hearing in the contralateral ear. Many patients recognise speech better using conventional acoustic hearing aids together with the cochlear implant than they do using either device alone but find the use of the' combination unacceptable.
  • acoustic hearing aid or the cochlear implant aid but not both devices together. It is an object of the present invention to provide a bimodal aid device which can drive both an acoustic hearing aid and a cochlear implant aid which thereby improves the quality of binaural information received by a patient.
  • Two further problems experienced in the prior art of hearing aids are (1) quickly and easily measuring the nature and degree of hearing impairment of a client for the purposes of providing an appropriate hearing aid and (2) the difficulty in matching appropriate hearing aid qualities and capabilities to the specific requirements of the user.
  • a bimodal aid for the hearing impaired which " includes processing means adapted to receive and process audio information received from a microphone; said processing means supplying processed information derived from said audio information to an implant aid adapted to be implanted in a first ear of a patient and to an acoustic aid adapted to be worn in or adjacent a second ear of said patient whereby binaural information is provided to said patient.
  • a bimodal aid for the hearing impaired comprising a sound/speech processor electrically connected to a hearing aid transducer adapted to be worn adjacent to or in an ear of a patient and electrically connected to an electrical signal transducer adapted to be worn in an ear of a patient; aid speech processor receiving and processing audio input information so as to produce an acoustic signal from said hearing aid transducer and an electrical signal from said electrical transducer whereby coherent binaural information is provided to said patient.
  • an electronically configurable sound/speech processor for the hearing impaired; said processor including configuration means and signal processing means; said processor adapted to receive audio information and to process said audio information by said signal processing means in accordance with parameters set by said configuration means so as to produce an output signal adapted to stimulate a hearing aid transducer; said configuration means adapted to receive one or more of electronic signal input or software input for the purpose of modifying said parameters.
  • said sound/speech processor utilises the speech features FO, Fl, F2, A0, Al, A2, A3, A4, A5 and voiced/voiceless sound decisions to produce said output signal adapted to stimulate a Rearing aid transducer, wherein said features are defined as follows:- FO is the fundamental frequency, Fl is the first formant frequency, F2 is the second formant frequency, A0 is the amplitude of FO, Al is the amplitude of Fl,
  • A2 is the amplitude of F2
  • said signal processing means includes means for dynamically changing the amplitude of different frequency bands in the defined speech spectrum as a function of the speech features Al, A2, A3, A4, and A5 parameters so that the loudness in these bands is appropriately scaled between the threshold and maximum comfortable levels of the hearing aid user at the Fl and F2 frequencies and in the higher frequency bands.
  • said signal processing means includes filter means whose settings are dynamically varied according to speech parameters extracted from said signal processing means whereby said filter means dynamically adapts said output signal to overcome the effects of noise and/or particular deficiencies in the hearing of a user.
  • said signal processing means within said processor includes means for reconstructing signals in real time whereby the amplitude and/or frequency characteristics of said output signal can be controlled optimally so as to enhance speech recognition in a user.
  • said filter means set by said configuration means are provided based on measurements set by an audiologist during a hearing aid fitting procedure. The filter settings remain fixed after completion of the procedure and remain fixed thereafter.
  • said processor includes filter means whose parameters are changed dynamically while said processor is in use providing said output signal to said hearing aid transducer in accordance with information provided to said configuration means by speech parameter extraction means acting on said audio information.
  • said output signal is synthesised by said signal processing means utilising only speech parameters.
  • Fig. 1 is a schematic representation of a bimodal aid according to a first embodiment of the invention
  • Fig. 2 is a schematic representation of the main functional components which comprise the entire bimodal aid
  • Fig. 3 is a schematic block diagram of the implant processing circuitry together with the acoustic processor circuitry
  • Fig. 4 is a chart showing an example of the pattern of electrical stimulation of the implant electrodes for various steady state phonemes using the multi-peak coding strategy
  • Fig. 5 is a graph showing the standard loudness growth function for the speech processor portion when driving the implant
  • Fig. 6 is a schematic block diagram of the functional components of the bimodal processor which drive the acoustic aid
  • Fig. 7 is a schematic block diagram of the components comprising the acoustic processor which processes the speech signal in accordance with particular forms of the invention to drive the acoustic aid,
  • Fig. 8 is a component block diagram of the acoustic processor
  • Fig. 9 is a component block diagram of a Biquad filter as utilised in the acoustic processor of Fig. 8
  • Fig. 10 is a component block diagram of the output driver for the acoustic aid
  • Fig. 11 is a schematic representation of preferred modes of operation of the acoustic processor for voiced vowel input
  • Fig. 12 shows a sound intensity against frequency for an acoustic aid operating according to mode 1 ,
  • Fig. 13 shows comparative plots of sound intensity against frequency in relation to an example of modes one and three operation of the acoustic processor
  • Fig. 14 outlines a fitting strategy for the bimodal aid
  • Fig. 15 is a flowchart of a fitting strategy for the acoustic aid portion of the device in mode 1
  • Fig. 16 is a schematic block diagram of the utilisation of the diagnostic and programming unit for use in association with the bimodal aid. Description of Preferred Embodiments
  • the bimodal aid is a hearing aid device which has the capability to provide information through a cochlear implant aid in one ear and a speech processing acoustic hearing aid in the other ear of a patient. Both the implant aid and the acoustic aid are controlled by the same speech processor which derives its raw audio input from a single microphone.
  • the speech processor extracts certain parameters from the incoming acoustic waveform that are relevant to the perception of speech sounds. Some of the speech parameters extracted by the speech processor are translated into an electrical signal and delivered to a cochlear implant. These features can also be used as a basis for modification of the speech waveform following which it is then amplified and presented to the acoustic aid.
  • the cochlear implant produces stimulation at positions in the cochlear that correspond to higher frequencies (usually above 700Hz) .
  • frequencies usually above 700Hz
  • the frequency and temporal resolution of residual hearing can be better than that provided by the pulsatile electric sig ⁇ al of the cochlear implant aid portion of the bimodal aid.
  • the acoustic aid driver part of the device can also be used as a speech processing hearing aid independent of the cochlear implant aid output.
  • a speech processing hearing aid independent of the cochlear implant aid output.
  • Conventional hearing aids are limited in practice because the adjustments to the frequency/gain characteristics are restricted to a small number of options and there are many users who are not optimally aided. There is a need for a hearing aid with a more flexible frequency/gain characteristic and this can be achieved with this aid.
  • the feature extraction circuits which are the basis of the cochlear implant aid allow the hardware to measure important characteristics of the speech signal in quiet conditions and in conditions of moderate amounts of background noise.
  • acoustic signal processor (12) which outputs to an acoustic aid.
  • the synthesized waveform is used to overcome special problems. For example, high frequency sounds above the limit of a user's hearing can be presented as lower frequencies within the user's hearing range. Broad peaks in the speech spectrum can be made narrower if this provides better frequency resolution and helps to reduce masking of one peak by other adjacent peaks. There is no,other single, wearable device capable of implementing all these processes.
  • the sound/speech processor can take in a speech sgnal from a microphone, measure selected features of that signal (including the frequency and amplitude of formants for voiced speech) and control the outputs to both a cochlear implant aid and an acoustic hearing aid in the case of the first embodiment or only the acoustic hearing aid in the case of the second embodiment.
  • FIG. 2 shows a schematic diagram of the operation of the device.
  • the cochlear implant aid portion of the device is covered by existing patents or patent applications to the same applicant and the implant aid operates upon one ear of a patient using similar strategies to those already developed for implant users.
  • the users of the bimodal device will receive an auditory signal via an acoustic aid in the non-implanted ear.
  • the capabilities of the bimodal aid allow this signal to be specially tailored in order to convey information complementary to the implant and utilise the' residual hearing of the patient maximally.
  • Fig. 2 discloses the body worn portion of the bimodal device comprises a speech processor 11 intimately connected to an acoustic aid processor 12 together with a microphone 13, an acoustic hearing aid 14 and an implant aid 15.
  • the implant aid 15 comprises an electrode array 16 electrically connected by harness 17 to a receiver stimulator 18 which is in radio communication with speech processor 11 by way of internal coil 19 and external coil 20.
  • Fig. 2 shows auxiliary items being the diagnostic and programming unit 21 and the diagnostic programming interface 22.
  • the diagnostic and programming unit 21 is implemented as a program running on a personal computer whilst the diagnostic programming interface 22 is a communications card connected to the PC bus.
  • the diagnostic and programming unit 21 is utilised in a clinical situation to test for and control device parameters of operation for the speech processor 11 and/or acoustic aid processor 12 which optimise hearing performance for a patient according to defined criteria. These parameters are communicated via the diagnostic programming interface 22 to a map memory storage 23 in the speech processor 11. It is by reference to the parameters stored in the map memory storage 23 that the manner of processing of the audio signal received from microphone 13 is determined both for the speech processor 11 when driving the implant aid 15 and the acoustic aid processor 12 when driving the acoustic aid 14.
  • Fig. 2 The components illustrated in Fig. 2 other than the acoustic aid processor 12 and the acoustic aid 14 and the computer program controlling the function of the diagnostic and programming unit 21 have been described elsewhere in earlier filed patents and patent applications and remain the same in so far as operation of the Cochlear implant aid is concerned.
  • the speech processor 11 and the precise methodology for exciting electrically the implant aid has varied since the inception of these devices and can be expected to continue to vary.
  • excitation of the stimulating electrodes placed within the ear of a patient can be either digital or analogue in nature.
  • Cochlear Pty. Limited has pursued a strategy of digital electronic stimulation using what have been termed pulsatile electrical signals applied to a pulsatile electrical signal transducer.
  • the speech processor 11 has been commercially available in a number of forms since around 1982 from Cochlear Pty. Limited (one of the co-applicants for the present application) .
  • the early units and, indeed, even the most recent units are primarily aimed at improving speech perception in favour of all other sounds received from microphone 13. This is done by causing speech processor 11 to discern and process from the raw audio input received from microphone 13 acoustic features of speech which have been determined to best characterise speech information as perceived by the human auditory system.
  • This latest coding scheme provides all of the information available in the F0F1F2 scheme while providing additional information from three high frequency band pass filters. These filters cover the following frequency ranges: 2000 to 2800 Hz, 2800 to 4000 Hz and 4000 to 8000 Hz. The energy within these ranges controls the amplitude of electrical stimulation of three fixed electrode pairs in the basal end of the electrode array. Thus, additional information about high frequency sounds is presented at a tonotopically appropriate place within the cochlear.
  • the overall stimulation rate for voiced sounds remains as F0 (fundamental frequency or voice pitch) but in the new scheme four electrical stimulation pulses occur for each glottal pulse. This compares with the F0F1F2 strategy in which only two pulses occur per voice pitch period.
  • the two pulses representing the first and second formant are still provided and additional stimulation pulses occur representing energy in the 2000 to 2800 Hz and the 2800 to 4000 Hz ranges.
  • the latest noise suppression algorithm operates in a continuous manner, rather than as a voice activated switch as previously been used. This removes the perceptually annoying switching on and off of the earlier system.
  • the noise flow is continuously assessed in each frequency band over a period of ten seconds. The lowest level over this period is assumed to be background noise and is subtracted from the amplitude relevant to that frequency band. Thus any increase in signal amplitude above the noise level is presented to the patient while the ambient noise level itself is reduced to near threshold.
  • Fig. 3 illustrates the basic filter and processing structure of a bimodal aid incorporating means to implement any of the above described latest processing scheme.
  • International Patent Application PCT/AU90/00407 to the present applicant entitled Multi-peak Speech Processor describes in detail the operation of these components. The entire text and drawings of the specification of that application are incorporated herein by cross reference. The most pertinent portions of that specification are included immediately below.
  • the coding strategy extracts and codes the Fl and F2 spectral peaks from the microphone audio signal, using the extracted frequency estimates to select a more apical and a more basal pair of electrodes for stimulation. Each selected electrode is stimulated at a pulse rate equal to the fundamental frequency F0.
  • Fl and F2 three high frequency bands of spectral information are extracted.
  • the amplitude estimates from band three (2000-2800 Hz), band four (2800-4000 Hz), and band five (above 4000 Hz) are presented to fixed electrodes, for example the seventh, fourth and first electrodes, respectively, of the electrode array 16 (Fig. 2 and Fig. 4) .
  • the first, fourth and seventh electrodes are selected as the default electrodes for the high,-frequency bands because they are spaced far enough apart so that most patients will be able to discriminate between stimulation at these three locations. Note that these default assignments may be reprogrammed as required. If the three high frequency bands were assigned only to the three most basal electrodes in the MAP, many patients might not find the additional high frequency information as useful since patients often do not demonstrate good place-pitch discrimination between adjacent basal electrodes. Additionally, the overall pitch percept resulting from the electrical stimulation might be too high.
  • Table I indicates the frequency ranges of the various formants employed in the speech coding scheme for the present invention.
  • FIG. 4 illustrates the pattern of electrical stimulation for various steady state phonemes when using this coding strategy.
  • a primary function of the MAP is to translate the frequency of the dominant spectral peaks (Fl and F2) to electrode selection.
  • Electrode 1 is the most basal electrode and electrode 22 is the most apical in the electrode array. Stimulation of different electrodes normally results in pitch perceptions that reflect the tonotopic organization of the cochlea. Electrode 22 elicits the lowest place-pitch percept, or the "dullest” sound. Electrode 1 elicits the highest place-pitch percept, or "sharpest” sound. To allocate the frequency range for the Fl and F2 spectral peaks to the total number of electrodes, a default mapping algorithm splits up the total number of electrodes available to use into a ratio of approximately 1:2, as shown in FIG. 4.
  • a random access memory stores a set of number tables, referred to collectively as the MAP memory storage 23.
  • the MAP determines both stimulus parameters for Fl, F2 and bands 3-5, and the amplitude estimates.
  • the encoding of the stimulus parameters follows a sequence of distinct steps. The steps may be summarized as follows:
  • the first formant frequency (Fl) is converted to a number based on the dominant spectral peak in the region between 280-1000 Hz.
  • the Fl number is used, in conjunction with one of the MAP tables, to determine the electrode to be stimulated to represent the first formant.
  • the indifferent electrode is determined by the mode.
  • the second formant frequency (F2) is converted to a number based on the dominant spectral peak in region between 800-4000 Hz.
  • the F2 number is used, in conjunction with " one of the MAP tables to determine the electrode to be stimulated to represent the second ormant.
  • the indifferent electrode is determined by the mode. 5.
  • the amplitude estimates for bands 3, 4 and 5 are assigned to the three default electrodes 7, 4 and 1 for bands 3, 4 and 5, respectively, or such other electrodes that may be selected when the MAP is being jprepared. 6.
  • the amplitude of the acoustic signal in each of the frequency bands is converted to a number ranging from 0 - 150.
  • the level of stimulation that will be delivered is determined by referring to set MAP tables that relate acoustic amplitude (in range of 0-150) to stimulation level for the specific electrodes selected in steps 2, 4 and 5, above.
  • the data are further encoded in the speech processor and transmitted to the receiver/stimulator 18. It, in turn, decodes the data and sends the stimuli to the appropriate electrodes. Stimulus pulses are presented at a rate equal to F0 during voiced periods and at a random a periodic rate within the range of FO and Fl formants (typically 200 to 300 Hz) during unvoiced periods.
  • the speech processor 11 additionally includes a non- linear loudness growth algorithm that converts acoustic signal amplitude to electrical stimulation parameters.
  • the speech processor 11 converts the amplitude of the acoustic signal into a digital linear scale with values from 0 - 150 as shown in FIG. 5. That digital scale (in combination with the information stored in the patient's MAP) determines the actual charge delivered to the electrodes in the electrode array 16. Improvements on this assembly are disclosed in co- pending applications to Cochlear Pty. Limited. Specifically International Application PCT/AU90/00406 discloses an improved connection system between migrophone 13 and speech processor 11 and between the external coil assembly 20 and the speech processor 11. The text and drawings of the specification of that application are incorporated herein by cross reference.
  • FIG. 6 is a block diagram of the processing circuitry showing the functional interconnection of components for driving the acoustic hearing aid 14.
  • the main components comprise microphone 13, automatic gain control 24, speech parameter extractor 25, encoder 26, patient MAP memory storage 23, noise generator 27 and acoustic aid signal processor 12.
  • the heart of the bimodal aid as far as allowing the acoustic aid 14 to be driven from the speech processor 11 is the acoustic aid signal processor 12.
  • the acoustic aid signal processor is software configurative and contains three two-pole filters each of which can be used in either bandpass, lowpass or highpass configuration. The centre frequency, bandwidth and output amplitude of these filters are controlled by the processor.
  • the filters can be used in series or in parallel and the input waveform can be the speech waveform, pulses, a noise signal or external signal.
  • the external signal can be from another microphone, other acoustic output or another acoustic signal processor. This results in a particularly flexible aid that can operate either in a manner similar to a conventional acoustic hearing aid (though with more accurate gain fitting than most currently available aids can provide) or as an aid providing different types of processed speech information.
  • the acoustic aid signal processor including the three programmable filters has been implemented on a single silicon chip.
  • Each filter is usable as a high-pass, band ⁇ pass, or low-pass filter.
  • This chip has the flexibility to cover a wide range of frequencies, amplitudes and spectral shapes.
  • DAC digital-to-analog converter
  • the DAC can produce waveforms of arbitrary shape (such as sinusoidal or pulsatile) controlled directly by the processor, or can be switched to provide excitation by the speech waveform, or a white noise generator.
  • a schematic diagram of the three-filter circuit is shown in Fig 7.
  • a functional specification for a single chip implementation of the acoustic aid signal processor 12 is provided by FIGS. 8, 9 and 10. Details of the specification are as follows: ⁇ Topology
  • Fig. 8 shows the overall topology of the chip.
  • Three programmable filters in which centre frequency and band width can be independently controlled are provided. The outputs of these filters can be independently attenuated or amplified and then mixed. The output of one of the three filters can be inverted if necessary by setting an INV bit.
  • Fig. 9 shows details of one of the Biquad filters forming the three filter array together with the frequency latches and Q latches which determine the parameters of the Biquad filter.
  • the topology of the chip can be altered from serial to parallel or a mixed structure by three PARn bits.
  • the signal source for this structure can be selected by a four channel multiplexer (MUX) . This selects +5 volts, a buffered output of the audio signal, an internally generated noise source, or an external signal. This signal source is fed to a 7 bit digital to analog converter (DAC) as a reference voltage.
  • MUX four channel multiplexer
  • the multiplying DAC can convert the DC level into a pulse generator, or provide a fine gain control on the audio or external signal, or noise source.
  • the most significant bit (MSB) is used to invert the output. All filter outputs are summed and passed to a push pull earphone driver which can provide effectively 10 volts peak- to-peak across a 270 ohm (nominal) earphone.
  • the chip uses a single supply of 5 volts.
  • the earphone has a DC resistance of 88 ohm with the impedance rising gradually to 270 ohm at 1 kHz.
  • the output stage consists of a bridge of P and N transistor switches as shown in Fig. 10. The switches are pulse width modulated by a signal derived from a comparator driven from a triangle wave on one input and the audio signal on the other. The on resistance of the switches should be less than 5 ohms (lower if possible).
  • the chip is programmed by writing to the MAP of the speech processor. To distinguish between chip and MAP writes, bits A8 - A12 are decoded. Any write to the block 1800 - 18FF in MAP will ' also write to the chip. Two addresses, one odd and one even (Y13 and Y14) are decoded and ORed and the output of these (R/W new) can be used to write to the filter chip in a more selective manner. Odd writes are used to select the MUX and even writes set the auto sensitivity control (ASC) latch on that chip. The four lowest address bits are used to write to 14 registers which contain the programming information for the acoustic processor chip. Registers Y0 to Yll program frequency, Q, gain (attenuation or amplification) and configuration in turn for each of the 3 filters. Register
  • Register Y15 is used to write to the DAC.
  • the topology latch Y12 is as follows:
  • Dl - PAR1 sends Filter 1 output to summer D2 - PAR2 sends Filter 2 output to summer D3 - PAR3 sends Filter 3 output to summer D6 - DIR sends the DAC output direct to summer
  • the cascaded filter and attenuator/amplifier are sent to the summer but the filter output itself is sent to a bus so that it can be made available to other filters.
  • the filter-attenuator/amplifier combination is sent to the bus. In this way the filter gains may be scaled if the filters are cascaded.
  • the filter inputs selected by D6 and D5 vary as shown below
  • a clock pre-scaler is provided to extend the frequency ranges of the filters. This is done by dividing the clock by 2 or 4 before feeding it to the filter's own divider. Decoding is as follows:
  • the filters consist of two integrators in a loop with a variable gain feedback path.
  • the input may be a switched or an unswitched capacitor, it may be applied to either the first or second input and the output may be taken from either the first or second integrator. This produces various different transfer functions as given below.
  • the programmable filter clock divider comprises a 7 bit ripple counter which is compared with the contents of the frequency latch. The first match produces a counter reset.
  • the frequency latch must be fed with the complement of the count required. If for example a reset is required after a count of 2 then the latch is all high except for bit Dl which is low. The outputs of all the NOR gates connected to the latch will be low except for the one connected to Dl. Now if the counter counts up from zero, on the first occasion of Q2 going high, this NOR will also go low and the 7 input NOR at the output will reset the counter.
  • the filter Q is programmed by using a capacitor which has a 4 bit binary sequence, together with a 3 bit programmable resistive divider.
  • the resistors are programmed by a number n which is represented by bits D4-D6 in the Yl latch (Y5 and Y9 for filters 2 and 3).
  • the capacitors are programmed by m, represented by bits D0-D3.
  • the Q is given by:
  • the feedback resistor of an amplifier can be configured to be like a buffered attenuator - i.e. the output can drive another attenuator or inverter.
  • Two separate sections are used, one to give 8, 4 dB steps giving a total of +/- 28 dB and another to give 8, 0.5 dB steps with a maximum of +/- 3.5 dB. The total range is therefore +/- 31.5 dB.
  • Two 8 channel MUXes are used to select the required taps on the potential dividers.
  • the attenuator/amplifiers can be selected by addressing latches Y2, Y6 and Y10.
  • Bit D6 1 gives attenuation, 0 amplification.
  • Y2 gain in dB x 2
  • Y2 64 - (atten in dB x 2) If Y2 is set to 0 (all bits low) then the attenuator/amplifier is powered down and switched ⁇ f. a Note that setting HF adds 6 dB to the gain.
  • the attenuator is arranged so that its gain is not changed except when the signal passes through zero. This is to prevent clicks and pops during gain changes.
  • a zero crossing detector which produces a pulse on either a positive 10 or negative going zero crossing is used to strobe a latch which transfers the required gain to the attenuator MUXes.
  • the multiplying DAC is a 6 bit resistive ladder type and multiplies the input REF, selected by the SOURCE SELECT (Fig. 8) by the digital quantity in latch Y15. Bit D6 inverts the 15 output. The settling time is 50 microseconds.
  • the acoustic aid processor 12 by its flexible and programmable construction, allows many signal processing strategies to be tried and ultimately settled upon so as to best adapt the acoustic aid to provide information to the 20 wearer which compliments information received from an implant aid worn in the other ear of the wearer. This same flexibility and programability also can be used to tailor the bimodal processor for operation of an acoustic hearing aid only. 25 In both cases it is the combination of use of a single microphone together with the preprocessing capabilities of the speech processor 11 combined with the flexibility and programability of the acoustic aid processor 12 which provides features and advantages not found in hearing aid devices to date.
  • the bimodal aid will now be desc ibed when used to drive an acoustic hearing aid only.
  • the modes of operation to be described in relation thereto are equally usable to help obtain the complementary behaviour mentioned above in the first embodiment in a relation to the use of both an acoustic aid and an implant aid by a single wearer. The following description therefore and to that extent should be taken as applying equally to the first embodiment.
  • the inherent flexibility of the acoustic aid signal processor 12 incorporating the three software configurable filters provides for an almost unlimited degree of flexibility of processing signals in the frequency domain received from the speech processor 11 and destined for the acoustic aid 14.
  • Four particular modes of operation have been identified as desirable and achievable by the acoustic aid signal processor 12.
  • mode 1 the filter parameters are set by the audiologist during the iterative fitting procedure and remain fixed thereafter.
  • modes 2-4 the speech parameter extraction circuits provide instantaneous information about the speech signal that is used to change the filter parameters dynamically while the aid is in use.
  • the input signal is the speech waveform.
  • the output signal is manipulated in different ways by controlling the filters to emphasize the chosen parts of the waveform (such as the formants) and to attenuate other parts (such as the background noise) .
  • mode 4 the speech waveform is used only by the speech parameter extractor, and the output waveform is synthesized completely using the speech parameters. The differences between the original speech waveform and the output of the hearing aid become greater as one progresses from mode 1 to 4, and the control over the frequency spectrum and intensity of the output signal als ' o increases.
  • the acoustic output is tailored to match the patient's hearing loss.
  • the 6 poles of filtering enable this to be done accurately (usually within 2 dB of the ideal gain specified by the audiologist at all frequencies) and the Automatic Gain Control allows the limited dynamic range of the residual hearing to be used.
  • the acoustic signal processor of the second embodiment configured in mode 1 provides both operational and practical advantages over conventional hearing aids. These advantages can best be appreciated by considering the steps involved in setting up both types of hearing aid for operation:
  • the conventional aid The majority of commercially available hearing aids merely amplify, and sometimes compress, the incoming sound. To fit one of these aids the audiologist would normally measure the user's thresholds using an audiometer, calculate the appropriate ideal gain by hand using a prescribed fitting rule (e.g. the National Acoustics Laboratory (NAL) rule, Byrne and Dillon, 1986). The audiologist would then search through the specifications of the aids stocked in the clinic to find one with a gain that most closely resembled the ideal gain. On all aids some changes can be made by the audiologist although the amount of control depends on the type of aid. The features that can be varied may include any combination of the overall gain, the maximum output, the level at which compression begins.
  • NAL National Acoustics Laboratory
  • Frequency specific variation of gain is usually only in two frequency bands corresponding to 'high' frequencies and 'low' frequencies respectively.
  • Behind-the-ear aids and body-worn aids usually offer greater scope for change by the audiologist than in-the-ear aids. This is because, with these types of aid, the acoustic properties of the tube and earmould can be varied in addition to the controls on the aid itself.
  • In-the-ear aids also require an earmould to be made specifically for that aid by the manufacturer. This is a costly and time consuming business. This makes testing and comparing in-the-ear aids difficult and expensive and many clinics avoid using them.
  • In-the-ear aids are also usually more limited in their maximum output and are therefore not often suitable for more severe hearing losses.
  • the aid When the aid is configured it is then tested on the client. If it proves unacceptable the audiologist must choose and reconfigure another sort of aid. This is repeated until an aid is found that the client considers acceptable.
  • the speech-processing hearing aid of the second embodiment (mode 1): the audiologist measures the client's hearing thresholds and any other hearing levels that might be needed for the strategies to be tested, e.g. maximum comfortable level (MCL) .
  • MCL maximum comfortable level
  • the measurements are made using the hearing aid and diagnostic and programming unit with associated configuring software rather than a separate audiometer. These values are then stored in a data file automatically.
  • a strategy is chosen and the aid is configured accordingly taking a maximum time of about five minutes. Calculation and fitting of ideal gain is done automatically and can be quickly accessed in a graphical form at any time by the audiologist.
  • the configured aid is then presented to the subject for evaluation. Different fittings can be tried in quick succession until an appropriate one is found.
  • the audiologist can change the ideal gain function at will if he/she believes that the ideal gain based on the client's threshold measurements is not optimal for that client. With many conventional hearing aids this can only be done grossly by changing the gain in "high” or “low” frequency bands or by choosing a different aid with quoted specifications closer to the new requirement for the client.
  • the device of the se ond embodiment can be configured exactly as many conventional aids, often more accurately. Setting up and testing the device are quicker, more efficient, and less prone to sources of error.
  • Fig. 12 provides an example fitting in mode 1 which is achievable utilising the acoustic aid signal processor 12 of the bimodal speech processor.
  • mode 2 This is similar to mode 1 except that the level output at any specific frequency is mapped non-linearly on a frequency specific basis by dynamically changing the gain parameters of the three filters in response to amplitude and frequency variations measured by the processor.
  • This requires that the audiologist measure the maximum comfort levels to which maximum amplitude can be mapped in addition to client thresholds.
  • the advantage of this mode is that it makes a more accurate mapping of dynamic range possible.
  • the frequency parameters of the three filters are changed dynamically (unlike modes 1 and 2 where they are fixed).
  • salient speech features can be enhanced.
  • the centre frequencies of two bandpass filters can be used to track the Fl and F2 (first and second formant) peaks in the signal. This acts as both a form of noise cancellation and also a removal of parts of the signal that might mask the information in the peaks to be traced.
  • the resulting signal after filtering is amplified to the appropriate loudness for the user on a frequency- specific basis as in mode 2. This may be most useful for users with impaired frequency resolution as well as raised thresholds.
  • the device used in this mode can also be used to amplitude modulate the signal at the fundamental frequency (FO) which can be another way of enhancing this parameter.
  • the most similar commercially available devices are the
  • noise-cancelling hearing aids such as those containing the "Zeta-noise-blocker” chip. These devices calculate an average long-term spectrum that represents the background noise and this noise is then filtered out of the signal along with any speech that happens to be at the same frequencies.
  • This mode 3 scheme is based on enhancement of the speech signal at the measured formant frequencies rather than cancellation of noise. This means that speech information which is close in frequency to the noise will not be lost although the noise further from the formants will be reduced. The scheme will also enhance the selected speech features in quiet conditions as well as in noisy conditions.
  • Fig. 13 provides an example of mode 3 wherein selective peak sharpening is performed by the acoustic aid signal processor 12.
  • Mode 4 - Speech Reconstruction This mode differs from the other modes of operation of the second embodiment in that the user does not receive a modified version of the input signal, but a completely synthesized signal constructed using parameters extracted by the speech processor. The signal can be reconstructed in many different ways depending on the user's hearing loss. This reconstruction provides very tight control over the signals presented and hence allows very accurate mapping onto the user's residual hearing abilities. It may be most useful for users with very limited hearing for whom normal amplification would provide no open set speech recognition. A second example of the use of this mode is for frequency transposition.
  • Each mode of operation allows a ⁇ wide range of potential strategies.
  • the modes are not discrete and some strategies that combine elements from different modes can be implemented.
  • a reconstructed signal representing FO information (mode 4) can be added to a filtered speech signal (mode 3) . 3.0 USE OF THE BIMODAL AID
  • the bimodal aid is programmed by use of a diagnostic and programming unit 21 which communicates with the speech processor 11 and, in turn, with the acoustic aid processor 12 by way of a diagnostic programming interface 22.
  • the diagnostic and programming unit 21 is implemented as a program on a personal computer.
  • the interface 22 is a communications card connected on the PC bus.
  • Fig. 15 outlines the procedure in flow chart form with particular reference to obtaining an optimum setting for the acoustic aid 14 in mode 1.
  • Fig. 16 outlines the basic interaction between the control program in the diagnostic and programming unit and the fitting and mapping procedures performed by the Audiologist.

Landscapes

  • Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Signal Processing (AREA)
  • Radiology & Medical Imaging (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Biomedical Technology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Neurosurgery (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Prostheses (AREA)
  • Circuit For Audible Band Transducer (AREA)
EP19910918663 1990-11-01 1991-11-01 Bimodal speech processor Withdrawn EP0555278A4 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
AU3144/90 1990-11-01
AUPK314490 1990-11-01

Publications (2)

Publication Number Publication Date
EP0555278A1 true EP0555278A1 (de) 1993-08-18
EP0555278A4 EP0555278A4 (en) 1994-08-10

Family

ID=3775048

Family Applications (1)

Application Number Title Priority Date Filing Date
EP19910918663 Withdrawn EP0555278A4 (en) 1990-11-01 1991-11-01 Bimodal speech processor

Country Status (4)

Country Link
EP (1) EP0555278A4 (de)
JP (1) JPH06506322A (de)
CA (1) CA2095344A1 (de)
WO (1) WO1992008330A1 (de)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0663137B1 (de) 1993-07-01 2002-10-02 The University Of Melbourne Cochleare implantatgeräte
US5651071A (en) * 1993-09-17 1997-07-22 Audiologic, Inc. Noise reduction system for binaural hearing aid
US5511128A (en) * 1994-01-21 1996-04-23 Lindemann; Eric Dynamic intensity beamforming system for noise reduction in a binaural hearing aid
US5626629A (en) * 1995-05-31 1997-05-06 Advanced Bionics Corporation Programming of a speech processor for an implantable cochlear stimulator
US6005955A (en) * 1996-08-07 1999-12-21 St. Croix Medical, Inc. Middle ear transducer
US5836863A (en) 1996-08-07 1998-11-17 St. Croix Medical, Inc. Hearing aid transducer support
US5814095A (en) * 1996-09-18 1998-09-29 Implex Gmbh Spezialhorgerate Implantable microphone and implantable hearing aids utilizing same
US5935166A (en) 1996-11-25 1999-08-10 St. Croix Medical, Inc. Implantable hearing assistance device with remote electronics unit
CA2323983A1 (en) * 2000-10-19 2002-04-19 Universite De Sherbrooke Programmable neurostimulator
US6730015B2 (en) 2001-06-01 2004-05-04 Mike Schugt Flexible transducer supports
US7529587B2 (en) 2003-10-13 2009-05-05 Cochlear Limited External speech processor unit for an auditory prosthesis
WO2005097255A1 (en) 2004-04-02 2005-10-20 Advanced Bionics Corporation Electric and acoustic stimulation fitting systems and methods
US8244365B2 (en) * 2004-05-10 2012-08-14 Cochlear Limited Simultaneous delivery of electrical and acoustical stimulation in a hearing prosthesis
US8280087B1 (en) 2008-04-30 2012-10-02 Arizona Board Of Regents For And On Behalf Of Arizona State University Delivering fundamental frequency and amplitude envelope cues to enhance speech understanding
DE102008060056B4 (de) * 2008-12-02 2011-12-15 Siemens Medical Instruments Pte. Ltd. Verfahren und Hörgerätesystem zur Anpassung einer bimodalen Versorgung
WO2014123890A1 (en) * 2013-02-05 2014-08-14 Med-El Elektromedizinische Geraete Gmbh Fitting unilateral electric acoustic stimulation for binaural hearing
US10721574B2 (en) 2011-11-04 2020-07-21 Med-El Elektromedizinische Geraete Gmbh Fitting unilateral electric acoustic stimulation for binaural hearing
US9511225B2 (en) 2013-01-24 2016-12-06 Advanced Bionics Ag Hearing system comprising an auditory prosthesis device and a hearing aid
US20160151629A1 (en) * 2013-07-05 2016-06-02 Advanced Bionics Ag Cochlear implant system
US9717909B2 (en) 2014-02-28 2017-08-01 Advanced Bionics Ag Systems and methods for facilitating post-implant acoustic-only operation of an electro-acoustic stimulation (“EAS”) sound processor
WO2015170140A1 (en) * 2014-05-06 2015-11-12 Advanced Bionics Ag Systems and methods for cancelling tonal noise in a cochlear implant system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3629521A (en) * 1970-01-08 1971-12-21 Intelectron Corp Hearing systems
WO1988009105A1 (en) * 1987-05-11 1988-11-17 Arthur Jampolsky Paradoxical hearing aid

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3818149A (en) * 1973-04-12 1974-06-18 Shalako Int Prosthetic device for providing corrections of auditory deficiencies in aurally handicapped persons
US3894196A (en) * 1974-05-28 1975-07-08 Zenith Radio Corp Binaural hearing aid system
DE2716336B1 (de) * 1977-04-13 1978-07-06 Siemens Ag Verfahren und Hoergeraet zur Kompensation von Gehoerdefekten
AU541258B2 (en) * 1980-12-12 1985-01-03 Commonwealth Of Australia, The Speech processor
US4596902A (en) * 1985-07-16 1986-06-24 Samuel Gilman Processor controlled ear responsive hearing aid and method
US4852175A (en) * 1988-02-03 1989-07-25 Siemens Hearing Instr Inc Hearing aid signal-processing system
US5027410A (en) * 1988-11-10 1991-06-25 Wisconsin Alumni Research Foundation Adaptive, programmable signal processing and filtering for hearing aids
AU6339290A (en) * 1989-09-08 1991-04-08 Cochlear Pty. Limited Multi-peak speech processor

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3629521A (en) * 1970-01-08 1971-12-21 Intelectron Corp Hearing systems
WO1988009105A1 (en) * 1987-05-11 1988-11-17 Arthur Jampolsky Paradoxical hearing aid

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of WO9208330A1 *

Also Published As

Publication number Publication date
JPH06506322A (ja) 1994-07-14
WO1992008330A1 (en) 1992-05-14
CA2095344A1 (en) 1992-05-02
EP0555278A4 (en) 1994-08-10

Similar Documents

Publication Publication Date Title
EP0555278A1 (de) Zweifach-sprachprozessor
Loizou et al. The effect of parametric variations of cochlear implant processors on speech understanding
US4532930A (en) Cochlear implant system for an auditory prosthesis
US7171272B2 (en) Sound-processing strategy for cochlear implants
US5095904A (en) Multi-peak speech procession
US5271397A (en) Multi-peak speech processor
Fu et al. Effects of amplitude nonlinearity on phoneme recognition by cochlear implant users and normal-hearing listeners
US6728578B1 (en) Envelope-based amplitude mapping for cochlear implant stimulus
US9008786B2 (en) Determining stimulation signals for neural stimulation
US20210006914A1 (en) Bimodal hearing stimulation system and method of fitting the same
US8843205B2 (en) Stimulation channel selection for a stimulating medical device
Geurts et al. Enhancing the speech envelope of continuous interleaved sampling processors for cochlear implants
AU2016285966B2 (en) Selective stimulation with cochlear implants
US20110218592A1 (en) Cochlear Implant Fitting System
EP2542302A1 (de) Verfahren und vorrichtung zur erzeugung eines elektrodenstimulationssignals in einer neuronalen hörprothese
CN110520188B (zh) 双模式听觉刺激系统
US10357655B2 (en) Frequency-dependent focusing systems and methods for use in a cochlear implant system
McDermott et al. A portable programmable digital sound processor for cochlear implant research
US11478640B2 (en) Fitting device for bimodal hearing stimulation system
Arora Cochlear implant stimulation rates and speech perception
Kaiser et al. Using a personal computer to perform real-time signal processing in cochlear implant research
US20240292162A1 (en) Method for improving effective dynamic range of neural stimulation for artificial cochlear system, and reconfigurable current dac-based neural stimulation ic chip therefor
Yang Design, Fabrication & Evaluation of a Biomimetic Filter-bank Architecture for Low-power Noise-robust Cochlear Implant Processors
Webster et al. Mechanisms of Hearing

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 19930528

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): DE FR GB

A4 Supplementary search report drawn up and despatched

Effective date: 19940620

AK Designated contracting states

Kind code of ref document: A4

Designated state(s): DE FR GB

17Q First examination report despatched

Effective date: 19951215

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 19970809