WO2013020792A1 - Hörer und verfahren zum steuern eines hörers - Google Patents
Hörer und verfahren zum steuern eines hörers Download PDFInfo
- Publication number
- WO2013020792A1 WO2013020792A1 PCT/EP2012/064230 EP2012064230W WO2013020792A1 WO 2013020792 A1 WO2013020792 A1 WO 2013020792A1 EP 2012064230 W EP2012064230 W EP 2012064230W WO 2013020792 A1 WO2013020792 A1 WO 2013020792A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- handset
- microphone
- finger
- housing
- touch
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1041—Mechanical or electronic switches, or control elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1058—Manufacture or assembly
- H04R1/1075—Mountings of transducers in earphones or headphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1083—Reduction of ambient noise
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2410/00—Microphones
- H04R2410/05—Noise reduction with a separate noise microphone
Definitions
- the present invention relates to a handset and a method for controlling a handset.
- Earphones such as headphones, earphones, in-ear earphones and ear-canal phones are well known and typically include at least one electro-acoustic playback transducer for playing back a cable or wirelessly received audio signal.
- handset which have on their cable or on the housing of the listener, a control unit, for example, for adjusting the volume.
- receivers which have a control unit by means of which both the volume can be adjusted and an electronic device connected to the receiver can be controlled.
- Such control units have at least three control units (louder, quieter, forward / backward).
- US 2006/0045304 A1 shows a receiver with a housing and a capacitive touch sensor on the housing.
- a handset is provided with at least one electro-acoustic playback transducer for reproducing an audio signal and at least one microphone.
- the receiver further has an evaluation unit for evaluating an output signal of the microphone with regard to touches, in particular of the listener, and for generating control signals. command to control the handset depending on the output signal of the microphone.
- the evaluation unit is designed to detect a finger-earpiece touch and to issue a control command.
- the handset has a cable and the evaluation unit is configured to detect a finger-cable touch and to issue a control command.
- the finger-to-hand touch or finger-to-cable touch may be a tap, a knock, a finger strike on the handset or cable, or a squeeze or cord squeeze.
- the invention also relates to a method for controlling a receiver which has at least one electroacoustic reproduction converter for reproducing an audio signal and a microphone.
- the output signal of the microphone is evaluated with respect to touches and control signals for controlling the handset in response to the output signal of the microphone are generated.
- the invention relates to the idea of using a microphone in a housing of the listener (for example a microphone for active noise compensation) for detecting control commands.
- the control commands may be, for example, that a user tapes with a finger on the housing of the listener, which is then detected by the microphone. Subsequently, the detected signal may be subjected to signal processing to determine whether the signal is a control command or not. Only if it is detected that it is a control command, then the control command is implemented accordingly. Otherwise the audio signal will be ignored.
- control commands can be triggered by a finger-listener contact.
- a finger-listener contact For example, on / off switching, talk through functionality, bass boost functionality, or the like may be implemented through the finger-to-hand contact.
- the listener if a listener has an active noise compensation unit, then the listener has a microphone. This microphone can then be used to detect a finger-listener contact.
- the listener may have an evaluation unit to evaluate the audio signals recorded by the microphone and to determine whether there has been a finger-listener contact to be converted into a control command.
- FIG. 1 is a block diagram of a receiver according to the first embodiment
- FIG. 2 shows a schematic sectional view of a receiver according to a second exemplary embodiment
- FIG. 3 shows a schematic sectional view of a receiver according to a third embodiment
- FIG. 4 shows a time profile of a microphone signal according to a fourth exemplary embodiment
- FIG. 5 shows a time profile of a microphone signal according to a fifth embodiment
- FIG. 6 shows a schematic representation of a receiver according to a sixth embodiment
- FIG. 7 shows a schematic representation of a receiver according to a seventh exemplary embodiment.
- Fig. 1 shows a block diagram of a receiver according to the first embodiment.
- the handset 100 has an input IN, an electroacoustic reproduction transducer 110, at least one microphone 120, a control unit 130, an evaluation unit 170 and optionally an active noise compensation unit 131.
- an audio signal to be reproduced can be received, which is then reproduced by the electroacoustic reproduction converter 110.
- the microphone 120 indoor and / or outdoor microphone
- ambient noise be recorded. These ambient sounds can then be used for active noise compensation, for example in the active noise compensation unit 131.
- the active noise compensation unit 131 then generates a compensation signal which is reproduced together with the audio signal to be reproduced by the electroacoustic reproduction converter 110.
- the evaluation unit 170 is coupled to the microphone 120 and evaluates the microphone signal. In this case, the evaluation is carried out in particular with regard to a finger-listener contact. If a finger-listener contact is detected by the evaluation unit 170, then the evaluation unit 170 checks whether this can be a control command. If this is not a control command, then the finger-listener contact is ignored. If, however, it is a control command, then the evaluation unit 170 outputs a corresponding control command to the control unit 130.
- the finger-earpiece touch may, for example, be a knocking of the finger on the housing of the listener, a collapse of the listener, a deleveraging of the handset or the like.
- Fig. 2 shows a schematic sectional view of a receiver according to a second embodiment, which is placed in an ear canal.
- the handset according to the second embodiment may correspond to or be based on the handset according to the first embodiment.
- the handset 100 is configured as an in-ear handset or ear-canal handset and has an electroacoustic reproduction transducer 110, a microphone 120, a control unit 130, an evaluation unit 170 and a sound guide section 150 with an ear cushion 160.
- the sound guide section 150 with the pad 160 may be inserted into an ear canal 110 of a user.
- the user can touch with his finger 300, the housing of the listener and thus cause a finger-earpiece touch. This touch may be a knock, a press, a strike or the like.
- the microphone 120 is configured as an internal microphone.
- the earphone according to the third embodiment substantially corresponds to the earphone according to the first or second embodiment. While the earphone according to the second embodiment has an inside microphone 120, the earpiece according to the third embodiment has an outside one Microphone 121 on. Furthermore, the receiver 100 has an electroacoustic reproduction converter 110, an external microphone 121, a control unit 130 and an evaluation unit 170.
- the evaluation unit 170 serves to evaluate the microphone signal of the microphone 121 in order to detect a finger-to-handset touch and possibly to convert it into a control command.
- FIG. 4 shows a timing diagram of a microphone signal during a finger-receiver contact according to the fourth embodiment.
- the microphone signal shown here was recorded by a listener according to the second embodiment.
- Fig. 5 shows a timing chart of a microphone signal according to a fifth embodiment.
- a finger-earpiece touch is shown, which was recorded by a finger tap on the handset in a handset according to the third embodiment.
- the earphones according to the first and second embodiments may also have an active noise canceling unit 131 as described in the first embodiment.
- the microphones may be used to record sound events arising due to contact between a finger 300 and a listener.
- Some of these touches such as tapping or pressing the handset, have characteristic sound events or properties as shown in FIGS. 4 and 5.
- These characteristic sound properties can be evaluated and identified by the evaluation unit 170, which represents, for example, a DSP-based evaluation unit. Based on this evaluation / identification, the finger-listener touches can be converted into control commands.
- a slight press of the earpiece can cause a sharp, sudden pressure increase in the earphone, this can be detected for example by an internal microphone 120.
- the evaluation unit 170 it can be checked whether the signal of the internal microphone exceeds a certain value for a minimum time.
- the microphone signal of the indoor microphone can be filtered with a low-pass filter with a very low cut-off frequency. Afterwards, the power can be can be evaluated and if the power exceeds a lower threshold, this can be evaluated as a switching action or as a control command.
- the finger-to-hand contact may represent a tap or a finger tap on a housing.
- a pulse-like sound is shown for example in Fig. 5.
- the occurring pulse is compared, for example, with a reference signal (when a finger is tapped on a handset). If these two signals agree sufficiently well, then a control command can be issued.
- both the left and the right handset can have a microphone and a separate evaluation unit, so that the sound events on the left and right handset can also be compared with one another in order to recognize and output a control command.
- sound events may occur whose signal characteristics are similar to those of touching, tapping or pressing the headphones. External or click-like sounds may be similar to tapping on the cabinet.
- jaw, jaw and head movements cause pressure fluctuations that are difficult to differentiate from the pressure fluctuations caused by the listener's pressure. All of these ambient sound events can be mistakenly interpreted as a user command.
- the simultaneous viewing of the right and left side signals of the headphone can help solve this problem.
- Ambient sound events that are not caused by touching a headphone cap can be heard on both sides of the headset at the same time or with a slight time shift and are recorded by both microphones on both sides.
- the sound events caused by touching a headphone cap can only be measured by a microphone on this side.
- the evaluation unit simultaneously takes into account the signals of the two sides and compares their features.
- the evaluation unit In the microphone signal on one side of the headphone there are characteristics that are due to a touch of the headphones may indicate the simultaneously detected signal characteristics of the other side to derive a decision as to whether it is a touch sound event to trigger a command. If the same features in the two-sided signals are detected simultaneously or with a slight time shift, the evaluation unit does not issue a command. If only one headphone side has in its microphone signal characteristics of a potential headphone touch, the evaluation unit outputs the corresponding command.
- pressing a headphone side causes a strong, momentary pressure fluctuation.
- the evaluation unit can detect, for example, the time in which the absolute value of the microphone signal was continuously above a certain high threshold. If this time exceeds a certain minimum duration, then there is a characteristic for a pressure fluctuation.
- a similar pressure fluctuation can be caused by a chewing, jaw or head movement, but in this case on both sides of the headphone.
- the evaluation unit then issues only one command if it does not determine a characteristic for a pressure fluctuation in the signal of the other headphone side in a timely manner.
- the advantages of the invention are, in particular, that no further components are required than those already present in the receiver. Only one evaluation unit has to be implemented in the handset.
- no additional switches are needed on the housing of the listener or for example on the cable of the listener.
- the finger-to-hand touches may be typing, tapping, stroking the finger along the case, squeezing the handset, or the like.
- successive finger-earpiece touches are also possible, which can be recognized by the evaluation as a control command.
- a simple tap or tapping effect a turn on the handset or a single function.
- a double tap / tap may cause a switch of a function of the listener.
- a movement of the finger along the housing can be interpreted, for example, as a control command for increasing or decreasing the volume.
- the evaluation unit 170 may be designed to detect not only a finger-earpiece touch, but also a compression of the earpiece, for example by moving the ears and to generate a corresponding control command.
- the surface of the listener (left / right) may be configured so that a top-to-bottom sounding different (or a different characteristic audio signal is generated) than a bottom-to-top sounding. This can be achieved, for example, by adapting the structure of the surface of the listener accordingly.
- a microphone may be provided on the cable of the listener.
- This microphone can be used for example for recording speech signals when the handset is used as a headset.
- the microphone signals recorded by the microphone on the cable are evaluated by the evaluation unit and converted accordingly into control commands.
- the invention relates to the idea that a characteristic audio pattern, which results from a touch of a listener or cable, is converted into control commands.
- a knocking or a finger-cable contact can also be detected and evaluated by the evaluation unit.
- a knock on the handset or on the cable may be detected as activation of the evaluation unit.
- the handset according to a sixth embodiment is designed as a headphone 500 and has at least one headphone housing 510.
- a first surface unit 520 is provided on the headphone housing 510.
- the first surface unit 520 has a plurality of, for example, horizontal grooves whose distances increase from bottom to top. Alternatively, the distances from bottom to top can be smaller. Alternatively, the grooves may have a different orientation (eg, vertical, oblique).
- the receiver housing 510 may include a second surface unit 530.
- This second surface unit 530 may have grooves both vertically and horizontally, or in a different orientation so as to form a pattern of grooves. The distances between the vertical and / or horizontal grooves may vary along the length and width of the second surface unit 530.
- a microphone within the headphone housing is used to detect control commands. This is advantageous because touch sounds (sounds caused by touch of the headphone) develop a much greater level within the listener due to the occlusion than outside the headphone.
- characteristic signals can be generated in the interior of the headphone housing due to certain touches that can come from any source other than a touch of the headphone. Pressing the headphone housing causes a sharp jump in pressure, which can only be measured within the headphone housing and is characteristic for pressing the headphone. By placing a microphone inside the headphone case, for example, pressing the headphone can be very well distinguished from other sound events. Thus, a robust distinction can be made possible.
- the surface units 520 and 530 are configured such that, for example, wiping along the surface unit 520 from top to bottom generates a different sound event than sweeping from the bottom to the top. Thus, for example, with the first surface unit 520 two preferred directions (top to bottom and bottom to top) can be detected. In the second surface unit 530, at least four preferential directions (top-bottom, bottom-top, left-right, and right-to-left) may be recognized.
- identification of touch sounds may be made more robust by providing the first or second surface unit 520, 530 on a headphone housing 510.
- the surface units 520, 530 have a structure that makes it possible to trace along, for. B. a finger on the first or second surface unit 520, 530 to recognize and distinguish on the basis of an acoustic sound event. By stroking a finger across the grooves of the first or second surface unit, a characteristic sound or a characteristic sequence of noises. The spacing of the grooves from one another may be such that different directions of a grazing finger-housing contact can be detected.
- strikethrough of a finger can be detected via the first surface unit 520. If a sequence of patterns of noise is initially slow and then steadily rising, then top-down motion can be detected. If the sequence of the noise pattern is fast and then steadily slower, then movement from bottom to top can be detected. If a sequence of noise patterns is initially slow and then faster, then movement may be from right to left, and if the sequence of the noise pattern is initially fast and then slow, then movement from left to right may be detected.
- the earphone according to the seventh embodiment is configured as a headphone 500 having at least a headphone housing 510 and a third surface unit 540.
- the third surface unit 540 may be configured as shown, for example in the form of a sawtooth.
- the third surface unit has a structured coating or a structured surface.
- the detected characteristic noises can then be converted into a control command.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Telephone Function (AREA)
- Soundproofing, Sound Blocking, And Sound Damping (AREA)
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE112012003238.1T DE112012003238A5 (de) | 2011-08-05 | 2012-07-19 | Hörer und Verfahren zum Steuern eines Hörers |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102011080518.4 | 2011-08-05 | ||
DE201110080518 DE102011080518A1 (de) | 2011-08-05 | 2011-08-05 | Hörer und Verfahren zum Steuern eines Hörers |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013020792A1 true WO2013020792A1 (de) | 2013-02-14 |
Family
ID=46650504
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2012/064230 WO2013020792A1 (de) | 2011-08-05 | 2012-07-19 | Hörer und verfahren zum steuern eines hörers |
Country Status (2)
Country | Link |
---|---|
DE (2) | DE102011080518A1 (de) |
WO (1) | WO2013020792A1 (de) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9743170B2 (en) | 2015-12-18 | 2017-08-22 | Bose Corporation | Acoustic noise reduction audio system having tap control |
US9930440B2 (en) | 2015-12-18 | 2018-03-27 | Bose Corporation | Acoustic noise reduction audio system having tap control |
US10091573B2 (en) | 2015-12-18 | 2018-10-02 | Bose Corporation | Method of controlling an acoustic noise reduction audio system by user taps |
US10110987B2 (en) | 2015-12-18 | 2018-10-23 | Bose Corporation | Method of controlling an acoustic noise reduction audio system by user taps |
US10354641B1 (en) | 2018-02-13 | 2019-07-16 | Bose Corporation | Acoustic noise reduction audio system having tap control |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0475297A2 (de) | 1990-09-07 | 1992-03-18 | Matsushita Electric Industrial Co., Ltd. | Fernbedienung für ein Gerät zur Audiowiedergabe |
US20060045304A1 (en) | 2004-09-02 | 2006-03-02 | Maxtor Corporation | Smart earphone systems devices and methods |
WO2006036262A2 (en) | 2004-09-23 | 2006-04-06 | Thomson Licensing | Method and apparatus for controlling a headphone |
JP2007019787A (ja) * | 2005-07-07 | 2007-01-25 | Yamaha Motor Co Ltd | 入力装置、入力機能付きヘルメットシステムおよびそれを備えた車両システム |
US20070079206A1 (en) | 2005-09-09 | 2007-04-05 | Manish Arora | Method and apparatus to control operation of multimedia device |
US20080019556A1 (en) * | 2006-07-21 | 2008-01-24 | Moon Eric E | Tactilely identifiable earphones |
EP1940195A2 (de) * | 2006-12-27 | 2008-07-02 | Sony Corporation | Tonausgabevorrichtung, Tonausgabeverfahren, Tonausgabeverarbeitungsprogramm und Tonausgabesystem |
US20080260176A1 (en) | 2005-10-28 | 2008-10-23 | Koninklijke Philips Electronics, N.V. | System and Method For Controlling a Device Using Position and Touch |
WO2009071919A1 (en) | 2007-12-07 | 2009-06-11 | The University Court Of The University Of Glasgow | Controller |
WO2011001229A1 (en) * | 2009-07-03 | 2011-01-06 | Sony Ericsson Mobile Communications Ab | Tactile input for accessories |
-
2011
- 2011-08-05 DE DE201110080518 patent/DE102011080518A1/de not_active Withdrawn
-
2012
- 2012-07-19 WO PCT/EP2012/064230 patent/WO2013020792A1/de active Application Filing
- 2012-07-19 DE DE112012003238.1T patent/DE112012003238A5/de not_active Withdrawn
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0475297A2 (de) | 1990-09-07 | 1992-03-18 | Matsushita Electric Industrial Co., Ltd. | Fernbedienung für ein Gerät zur Audiowiedergabe |
US20060045304A1 (en) | 2004-09-02 | 2006-03-02 | Maxtor Corporation | Smart earphone systems devices and methods |
WO2006036262A2 (en) | 2004-09-23 | 2006-04-06 | Thomson Licensing | Method and apparatus for controlling a headphone |
JP2007019787A (ja) * | 2005-07-07 | 2007-01-25 | Yamaha Motor Co Ltd | 入力装置、入力機能付きヘルメットシステムおよびそれを備えた車両システム |
US20070079206A1 (en) | 2005-09-09 | 2007-04-05 | Manish Arora | Method and apparatus to control operation of multimedia device |
US20080260176A1 (en) | 2005-10-28 | 2008-10-23 | Koninklijke Philips Electronics, N.V. | System and Method For Controlling a Device Using Position and Touch |
US20080019556A1 (en) * | 2006-07-21 | 2008-01-24 | Moon Eric E | Tactilely identifiable earphones |
EP1940195A2 (de) * | 2006-12-27 | 2008-07-02 | Sony Corporation | Tonausgabevorrichtung, Tonausgabeverfahren, Tonausgabeverarbeitungsprogramm und Tonausgabesystem |
WO2009071919A1 (en) | 2007-12-07 | 2009-06-11 | The University Court Of The University Of Glasgow | Controller |
WO2011001229A1 (en) * | 2009-07-03 | 2011-01-06 | Sony Ericsson Mobile Communications Ab | Tactile input for accessories |
Non-Patent Citations (1)
Title |
---|
RODERICK MURRAY-SMITH ET AL: "Stane: Synthesized Surfaces for Tactile Input", THE 26TH ANNUAL CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2008 : CONFERENCE PROCEEDINGS ; APRIL 5-10, 2008 IN FLORENCE, ITALY, ACM, NEW YORK, NY, USA, 4 May 2008 (2008-05-04), pages 1 - 4, XP007913166, ISBN: 978-1-60558-011-1 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9743170B2 (en) | 2015-12-18 | 2017-08-22 | Bose Corporation | Acoustic noise reduction audio system having tap control |
US9930440B2 (en) | 2015-12-18 | 2018-03-27 | Bose Corporation | Acoustic noise reduction audio system having tap control |
US10091573B2 (en) | 2015-12-18 | 2018-10-02 | Bose Corporation | Method of controlling an acoustic noise reduction audio system by user taps |
US10110987B2 (en) | 2015-12-18 | 2018-10-23 | Bose Corporation | Method of controlling an acoustic noise reduction audio system by user taps |
US10354641B1 (en) | 2018-02-13 | 2019-07-16 | Bose Corporation | Acoustic noise reduction audio system having tap control |
Also Published As
Publication number | Publication date |
---|---|
DE102011080518A1 (de) | 2013-02-07 |
DE112012003238A5 (de) | 2014-04-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1744589B1 (de) | Hörvorrichtung und entsprechendes Verfahren zur Eigenstimmendetektion | |
WO2013020792A1 (de) | Hörer und verfahren zum steuern eines hörers | |
DE10002321C2 (de) | Sprachgesteuerte Vorrichtung und System mit einer derartigen sprachgesteuerten Vorrichtung | |
DE112009005469B4 (de) | Lautsprecherschutzvorrichtung und Verfahren dafür | |
DE102014106428B4 (de) | Vorrichtung und Verfahren zum Steuern einer Verstärkung eines Verstärkers, und Digitalisierungsschaltkreis und Mikrofonbaugruppe | |
DE19904538C1 (de) | Verfahren zur Rückkopplungserkennung in einem Hörgerät und Hörgerät | |
WO2016078786A1 (de) | Verfahren und vorrichtung zum schnellen erkennen der eigenen stimme | |
DE102006047965A1 (de) | Hörhilfsgerät mit einer Okklusionsreduktionseinrichtung und Verfahren zur Okklusionsreduktion | |
DE60018783T2 (de) | Verfahren zur erkennung von kopfbewegungen oder kopfmuskelaktivität und zum erzeugen eines ausgangssignals | |
EP2224752B1 (de) | Vorrichtung und Verfahren zur Reduzierung von Trittschallwirkungen bei Hörvorrichtungen mit aktiver Okklusionsreduktion | |
EP3596822A1 (de) | Verstärkereinheit für einen schallwandler und schallerzeugungseinheit | |
DE10058786A1 (de) | Verfahren zum Steuerung eines eine akustische Ausgabeeinrichtung aufweisenden Geräts | |
DE102015216822B4 (de) | Verfahren zur Unterdrückung einer Rückkopplung in einem Hörgerät | |
EP3576431A1 (de) | Verfahren zum betreiben einer lautsprechereinheit sowie eine lautsprechereinheit | |
EP1401242B1 (de) | Feedbackkompensation für Hörgeräte mit Systemabstandsschätzung | |
EP1523220A2 (de) | Hörhilfevorrichtung zum automatischen Schalten in einen Telefonbetrieb und entsprechendes Verfahren | |
DE10208096A1 (de) | Mobiles Kommunikationsendgerät | |
EP2200341B1 (de) | Verfahren zum Betrieb eines Hörhilfegerätes sowie Hörhilfegerät mit einer Quellentrennungseinrichtung | |
DE202017106383U1 (de) | Fühlbare Basswiedergabe | |
DE102008022533B3 (de) | Verfahren zum Betrieb eines Hörgeräts und Mikrofonsystem für ein Hörgerät | |
WO2021239864A1 (de) | Verfahren, vorrichtung, kopfhörer und computerprogramm zur aktiven unterdrückung des okklusionseffektes bei der wiedergabe von audiosignalen | |
DE102022203969A1 (de) | Steuereinrichtung und Verfahren zur Steuerung eines am Kopf getragenen Gerätes, Ohrhörer und Steuerungssystem | |
EP3048808B1 (de) | Digitales drahtlos-audioübertragungssystem mit optimierter dynamik | |
EP1696700A2 (de) | Hörhilfegerät mit benutzergesteuerter Einmessautomatik | |
DE102008061844A1 (de) | Hörer und Headset |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12746045 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 112012003238 Country of ref document: DE Ref document number: 1120120032381 Country of ref document: DE |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: R225 Ref document number: 112012003238 Country of ref document: DE Effective date: 20140430 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 12746045 Country of ref document: EP Kind code of ref document: A1 |