US20160142834A1 - Electronic communication system that mimics natural range and orientation dependence - Google Patents

Electronic communication system that mimics natural range and orientation dependence Download PDF

Info

Publication number
US20160142834A1
US20160142834A1 US15/002,452 US201615002452A US2016142834A1 US 20160142834 A1 US20160142834 A1 US 20160142834A1 US 201615002452 A US201615002452 A US 201615002452A US 2016142834 A1 US2016142834 A1 US 2016142834A1
Authority
US
United States
Prior art keywords
microphone
signals
microphone signals
user
shared
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/002,452
Inventor
Timothy R. Beevers
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US15/002,452 priority Critical patent/US20160142834A1/en
Publication of US20160142834A1 publication Critical patent/US20160142834A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B17/00Monitoring; Testing
    • H04B17/20Monitoring; Testing of receivers
    • H04B17/27Monitoring; Testing of receivers for locating or positioning the transmitter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/55Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
    • H04R25/554Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired using a wireless connection, e.g. between microphone and amplifier or using Tcoils
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B5/00Near-field transmission systems, e.g. inductive or capacitive transmission systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B5/00Near-field transmission systems, e.g. inductive or capacitive transmission systems
    • H04B5/40Near-field transmission systems, e.g. inductive or capacitive transmission systems characterised by components specially adapted for near-field transmission
    • H04B5/43Antennas
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2201/00Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
    • H04R2201/10Details of earpieces, attachments therefor, earphones or monophonic headphones covered by H04R1/10 but not provided for in any of its subgroups
    • H04R2201/107Monophonic and stereophonic headphones with microphone for two-way hands free communication
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2420/00Details of connection covered by H04R, not provided for in its groups
    • H04R2420/07Applications of wireless loudspeakers or wireless microphones

Definitions

  • hearing aids tend to amplify background noise as well as sounds that the wearer wishes to hear
  • the ability to hear a speaker is typically not as good as it is if the speaker talks into a microphone from which the signals are sent to an apparatus that reconstructs the sounds and projects them directly into the listener's ear.
  • This type of system is known, but in form that appears to presume a simple two person environment, in which just one person suffers from hearing impairment. If a group of 4-10 hearing impaired persons congregated, the system might not adequately permit an easy course of communications.
  • a method for facilitating communications between a group of persons mutually positioned close to one another.
  • Each said person is provided with a microphone, positioned to receive voice sounds from said person, and an earphone.
  • a microphone signal from each of said microphones is received and analyzed to determine a portion that is shared between said microphone signals and attenuating said shared portion.
  • a signal is directed, based on said microphone signals, to at least one of said earphones.
  • an apparatus for facilitating communications between a group of persons mutually positioned close to one another.
  • a set of headsets each headset including at least one microphone and at least one earphone, is shaped to be worn by a user, so that said microphone receives sound from said user's mouth and said earphone emits sound near said user's ear.
  • a radio frequency transceiver responsive to said microphone and adapted to receive radio frequency signals from other headsets.
  • a sound filtering unit is also provided, that analyzes said microphone signals to determine a portion that is shared between said microphone signals and attenuates said shared portion, thereby creating an earphone signal.
  • FIG. 1 is a block diagram, illustrating the operation of an absolute system, according to the present invention.
  • FIG. 2 is an illustration of a relative system according to the present invention, in use.
  • FIG. 3 is an isometric view of a headset used in the system of FIG. 1 .
  • FIG. 4 is a block diagram, illustrating the operation of a relative system, according to the present invention.
  • One use of such a system would be for a group of 4 to 8 users seated together about a table in a noisy environment, such as a busy restaurant. If a first user is close to a second user and a third user, but wishes to speak principally to the second user he is likely to lean toward the second user and turn his head toward the second user. The system detects this change in orientation and range and increases the volume of the first user's voice in the second user's ears. Accordingly, the system users may communicate much as they would naturally, with no system present, except that the ability to hear each other is greatly augmented. In one preferred embodiment the range is compared to a threshold range and if it is greater than the threshold range, there is no effective transmission of sound from one user to another.
  • an index representing both sound and orientation is compared to a threshold value, so that a user in a restaurant does not hear the sounds from the table behind him, where people are similarly equipped with system enabled headphones. If the wait staff is equipped with head phones, then when a waiter approaches a particular table populated by system users, the waiter and the customers can communicate using their headsets, but when the waiter is further away, the people at that table are not bothered by the waiter's voice.
  • One preferred embodiment includes “shout override” wherein if someone shouts into his or her microphone, a microphone volume threshold is exceeded and the range threshold is suspended and even people far away can hear what he is saying.
  • a group of interested persons at a soccer match all wear system units. If a first user wishes to speak to a second user he turns toward the second user, much as he would do naturally. If the second user can see the first user he may at this point recognize that the first user is preparing to speak to him and mentally prepare himself, perhaps without realizing it, to listen to the first user.
  • an absolute system accelerometers are used to keep track of orientation and location.
  • Original orientation may be measured relative to a built-in compass that reports a direction electronically to a data processing unit.
  • a unit is activated at a known location and orientation or a GPS receiving system is used to find the location of the system users.
  • a plumb bob is included to correct orientation relative to straight down.
  • the accelerometers are periodically corrected by the compass and the plumb bob, for example every ten seconds, to prevent drift from becoming destructive.
  • Two sets of three orthogonal accelerometers are present, one close to each ear on a head set, and the position of these two sets of accelerometers are compared to find the orientation of the head set, and by implication, the user's head. This comparison can also be used to correct errors, as while the head set is being worn, there must be a fixed relationship between the relative positions of the two sets of accelerometers and the orientation of each set of accelerometers.
  • FIG. 1 is a block diagram of such a system unit 10 , which would physically take the familiar form of a headset with a microphone 30 and two earphones 40 and 42 .
  • a GPS 12 provides an initial position and updates when possible, and accelerometers 14 and a digital compass and miniature plumb bob 16 fill in when the GPS is not available, and provide greater detail of movement, including head movement.
  • a data processing unit 18 receives and integrates data from the three aforementioned units, and also receives a similar data set from each of the identical units 10 from which unit 10 receives an RF signal by way of antenna 20 and transceiver 22 .
  • data processing unit 18 By receiving the position and head orientation data from the other units and comparing this information to the unit 10 location and orientation information, data processing unit 18 is able to determine if any one of the other users has his head pointed in the direction of unit 10 . More specifically, the case of two headset wearers facing each other can be detected, and may be treated as a special case for placing the users into voice contact with one another.
  • Another portion of the RF signal is an encoded representation of the sound entering a microphone 30 and modulated to RF by the transceiver 22 of another headset, which will typically be the voice of another user.
  • the signal will be represented to the user of unit 10 by way of earphones 40 and 42 , provided that the range and orientation between units satisfies a predetermined condition. For example, the other user's head is facing the user of unit 10 , and/or the other unit is within a designated range.
  • the orientation of a listener's orientation relative to the place where the sound is originating is used to place a volume difference and time delay into the sound waves generated by the speakers in order to mimic the natural volume difference and time delay that the sound waves would have if received directly from the sound originating location. This permits the user to tell from which direction the sound is originating.
  • the system is configured to operate in an area in which an RF antenna is broadcasting a directional signal that is received by two sets of three orthogonal inductors, one near each ear of a user.
  • the relative power of reception of each inductor is used to determine the orientation of the inductor set relative to the known orientation of the electromagnetic field.
  • both a first headset unit 110 and a second, identical, headset unit 111 form part of the system.
  • first unit 110 both systems (and additional units not shown in FIG. 1 , but included within the scope of the invention) are described by the following description of first unit 110 .
  • First unit 110 is equipped with a directional speaker 112 that produces sound at a range above that of human hearing, for example in the range of 100,000 Hz (wavelength of about 3 mm) to 1,000,000 Hz (wavelength of about 0.3 mm). Moreover, unit 110 has a data processing unit 114 having a coded signal producer 126 adapted to drive speaker 112 to produce a coded sound signal. Units 110 and 111 also include a right-side microphone 120 and a left-side microphone 122 , so that second unit 111 receives sound from the speaker 112 of first unit 110 , in both microphone 120 and microphone 122 , which send a related electrical signal to logic and control unit 114 .
  • a coded signal generator 126 of a first unit 110 produces a first sound signal marked with a first code.
  • a second unit 111 upon receiving the first coded signal by way of microphones 120 , 122 and decoding it by logic and control unit 114 , sends back a second coded signal in a second code, that is dependent on the first code, so that the first unit, upon receiving the second coded sound signal is able to determine that it is the response of the second unit 111 to the recently transmitted sound signal from the first unit 110 .
  • each signal sent by speaker 112 is time stamped and after reception by another unit 110 , this time stamp can be compared with the present time, to arrive at a sound signal travel time and therefore a distance.
  • Unit 110 is also equipped with a mouth microphone 124 positioned to receive the voice of a user 126 .
  • the resultant electrical signal from microphone 124 is used, by modulation unit 128 , to modulate an RF signal that is broadcast by antenna 130 and received by the matching antenna 130 of the second unit 111 .
  • the second unit 111 receives this signal by way of antenna 130 , demodulates the signal, and with some further processing that is described below, sends the signal to a right-side headphone 140 and a left-side headphone 142 .
  • the data processing units 114 use the signals received from microphones 120 , 122 much as the human brain does with the signal from the ears, to determine the position and orientation of the sound signal originating speaker 112 , relative to the microphones 120 , 122 . This is done by comparing the time delay and volume of the signal received in the right-side microphone 120 versus left-side microphone 122 .
  • the signals driving headphones 140 and 142 are modulated to reflect the position of the speaker 112 , relative to microphones 120 and 122 , to permit the user 126 to detect sound directionality.
  • the data processing unit 114 does not determine position and orientation of any speaker 112 , but simply adds the time delay and volume differences found between the signals from microphones 120 and 122 , into the signals going to headphones 140 and 142 , respectively, thereby effectively providing sound direction information to the user.
  • the range information found as described above can be used to determine the volume of sound to be produced by a right-side headphone 140 and a left-side headphone 142 .
  • the range determined by the sound signal system of speaker 112 and microphones 120 is above a threshold (decision box 154 , FIG. 4 )
  • no sound responsive to the microphone 124 of the first unit is produced by the headphones 140 and 142 of the second unit 111 .
  • the volume of the received sound signal also is helpful in determining how loud the system should represent the sounds associate with the received RF signal from the other unit, as this volume is proportional to range and relative orientation, much as the actual voice sounds from the user are.
  • the range is divided by the magnitude of the weaker of the two sound signals received by microphones 120 and 122 , to form an estimate of the degree to which the first and second users are facing each other (block 152 , FIG. 4 ), and this estimate is compared to a threshold (decision box 154 , FIG. 4 ) to determine if the voice signal from microphone 124 of unit 110 should be presented to the second user 128 , by way of the headphones 140 and 142 of the second unit.
  • the same calculation is performed, but with the stronger of the two signals received by microphones 120 and 122 , to determine if first unit 110 is facing second unit 111 , regardless of how second unit 111 is facing.
  • This second option would be useful in a commander, commanded situation, for example the workers at a fast food restaurant, enabling a supervisory employee to look at a subordinate, speak and be heard, regardless of whether the subordinate was facing the supervisory employee or whether loud machinery was operating. Also the employees not being faced by the supervisor would not be interrupted or distracted from their tasks or both, by the supervisor's voice on the headphone system. Additionally, a system that does not send a signal to everyone permits a greater degree of privacy in communications.
  • the altitude at which the units will be used is entered during an initialization process performed by docking each unit to a computer using a USB connector cord.
  • a user interface appears on the computer screen and the user is guided to input information such as altitude. Users are advised that if they change altitude by more than 1,500 feet that it is advisable to reenter the altitude information so that the range calculations will not be corrupted by change in speed of sound with altitude.
  • the wireless communications apparatus of each unit is used to establish a communicative link with a computer and in similar manner the altitude is entered by a user.
  • a similar coding scheme is used to establish range between units using RF signals.
  • each unit is headphone having a speaker for both ears and a microphone supported near the user's mouth.
  • some or all of the units may be in the form of a device that clips onto a user's ear and that has a microphone extending toward the user's mouth.
  • auditory noise is discriminated against by comparing the signals received by all of the microphones of the various units.
  • the data processing unit 114 of each unit does this separately.
  • One method of performing such an operation is to perform fast Fourier transforms of the incoming sound, and filter out those frequencies that are being received in all the microphones, likely background noise, while retaining those frequencies found in a single microphone, which would most likely correspond to the voice of a speaker.
  • dual microphones about a head's width apart are used to receive sound in a speaker phone, and this sound is reconstructed in the earphones of listeners of a telephone connected with the speaker phone, to enable the listeners to tell speakers apart on the basis of their reconstructed locations.
  • Headphones may be used to produce sound on both ends of the telephone connection, with a pair of spaced microphones adapted to receive sound on both ends.
  • each unit also includes a cell phone, so that a user may contact any cell phone equipped person by dialing their telephone number or may also be placed into communication with any other system unit user who is within range, has the correct orientation and has chosen to permit this type of contact.
  • a cell phone so that a user may contact any cell phone equipped person by dialing their telephone number or may also be placed into communication with any other system unit user who is within range, has the correct orientation and has chosen to permit this type of contact.
  • possible systems have been described in terms of relative or absolute systems, it falls within the scope of this application to have a system that includes elements of both types of systems.
  • one or more headsets is configured for a hearing impaired person, with amplification of received sound signals in the frequencies most needed.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Electromagnetism (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Neurosurgery (AREA)
  • Otolaryngology (AREA)
  • Acoustics & Sound (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

A method of facilitating communications between a group of people close to one another. Each said person is provided with a microphone, positioned to receive voice sounds from said person, and an earphone. A microphone signal from each of said microphones is received and analyzed to determine a portion that is shared between said microphone signals and attenuating said shared portion. Finally, a signal is directed, based on said microphone signals, to at least one of said earphones.

Description

    RELATED APPLICATIONS
  • This application is a continuation of application Ser. No. 14/276,281, filed on May 13, 2014, which is in turn a continuation of application Ser. No. 13/035,225, filed on Feb. 25, 2011, now U.S. Pat. No. 8,761,674, issued on Jun. 24, 2014, which are incorporated herein by reference as if fully set forth herein.
  • BACKGROUND
  • Various RF systems for short range (<1 mile) communications currently exist, but most are geared to systems for permitting communications between people who are further than 100 feet apart. There are various situations, however, in which people who are relatively close together, even within 6 feet apart, could benefit from assistance in communications.
  • In a first situation, one or both of the people suffer from hearing loss. Because hearing aids tend to amplify background noise as well as sounds that the wearer wishes to hear, the ability to hear a speaker is typically not as good as it is if the speaker talks into a microphone from which the signals are sent to an apparatus that reconstructs the sounds and projects them directly into the listener's ear. This type of system is known, but in form that appears to presume a simple two person environment, in which just one person suffers from hearing impairment. If a group of 4-10 hearing impaired persons congregated, the system might not adequately permit an easy course of communications.
  • In a second situation, a group of people are gathered in a noisy environment, such as a crowded restaurant, a cocktail party or a nightclub. Communications may become strained in such an environment. Even very mild hearing loss can result in missed understanding. Further, even a person with no hearing impairment may experience difficulty understanding what is being said and may have to fill in the blanks with guesswork, potentially leading to embarrassing gaffs.
  • In a third situation, the people who wish to communicate are separated by less than 100 yards, but by far enough so that they cannot communicate comfortably without assistance. Although, as noted before, systems exist for aiding the communicators in this situation, the use of these systems tends to become difficult if more than two communicators wish to use the same system. It may be difficult to distinguish one or more of the communicators. The difficulty in distinguishing speakers also is present in telephone conversations where more than one person is sharing a speaker phone.
  • SUMMARY
  • The following embodiments and aspects thereof are described and illustrated in conjunction with systems, tools and methods which are meant to be exemplary and illustrative, not limiting in scope. In various embodiments, one or more of the above-described problems have been reduced or eliminated, while other embodiments are directed to other improvements.
  • In one aspect, a method is described for facilitating communications between a group of persons mutually positioned close to one another. Each said person is provided with a microphone, positioned to receive voice sounds from said person, and an earphone. A microphone signal from each of said microphones is received and analyzed to determine a portion that is shared between said microphone signals and attenuating said shared portion. Finally, a signal is directed, based on said microphone signals, to at least one of said earphones.
  • In another aspect, an apparatus is described for facilitating communications between a group of persons mutually positioned close to one another. A set of headsets, each headset including at least one microphone and at least one earphone, is shaped to be worn by a user, so that said microphone receives sound from said user's mouth and said earphone emits sound near said user's ear. Also included is a radio frequency transceiver, responsive to said microphone and adapted to receive radio frequency signals from other headsets. A sound filtering unit is also provided, that analyzes said microphone signals to determine a portion that is shared between said microphone signals and attenuates said shared portion, thereby creating an earphone signal.
  • In addition to the exemplary aspects and embodiments described above, further aspects and embodiments will become apparent by reference to the drawings and by study of the following detailed descriptions.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Exemplary embodiments are illustrated in referenced drawings. It is intended that the embodiments and figures disclosed herein are to be considered illustrative rather than restrictive.
  • FIG. 1 is a block diagram, illustrating the operation of an absolute system, according to the present invention.
  • FIG. 2 is an illustration of a relative system according to the present invention, in use.
  • FIG. 3 is an isometric view of a headset used in the system of FIG. 1.
  • FIG. 4 is a block diagram, illustrating the operation of a relative system, according to the present invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT(S)
  • It is an object of some preferred embodiments of the present invention to provide a short range communications system that mimics the effects of distance and orientation between speaker and listener in human communications. Accordingly in a preferred embodiment the volume at which sounds are projected into a listening user's ears are related to the distance and orientation between a speaking user and a listening user.
  • One use of such a system would be for a group of 4 to 8 users seated together about a table in a noisy environment, such as a busy restaurant. If a first user is close to a second user and a third user, but wishes to speak principally to the second user he is likely to lean toward the second user and turn his head toward the second user. The system detects this change in orientation and range and increases the volume of the first user's voice in the second user's ears. Accordingly, the system users may communicate much as they would naturally, with no system present, except that the ability to hear each other is greatly augmented. In one preferred embodiment the range is compared to a threshold range and if it is greater than the threshold range, there is no effective transmission of sound from one user to another. In another preferred embodiment an index representing both sound and orientation is compared to a threshold value, so that a user in a restaurant does not hear the sounds from the table behind him, where people are similarly equipped with system enabled headphones. If the wait staff is equipped with head phones, then when a waiter approaches a particular table populated by system users, the waiter and the customers can communicate using their headsets, but when the waiter is further away, the people at that table are not bothered by the waiter's voice. One preferred embodiment includes “shout override” wherein if someone shouts into his or her microphone, a microphone volume threshold is exceeded and the range threshold is suspended and even people far away can hear what he is saying.
  • In another use environment, a group of interested persons at a soccer match all wear system units. If a first user wishes to speak to a second user he turns toward the second user, much as he would do naturally. If the second user can see the first user he may at this point recognize that the first user is preparing to speak to him and mentally prepare himself, perhaps without realizing it, to listen to the first user.
  • To gain these benefits it is necessary to have a system that measures range and relative orientation between transceiving units. There are a number of ways of performing these tasks, but the various methods can be divided into two broad categories: 1) an “absolute system” that monitors position and orientation of each unit relative to a coordinate system, such as latitude and longitude, deriving relative range and orientation from this information and 2) a “relative system” that directly measures relative range and orientation.
  • In one preferred embodiment of an absolute system accelerometers are used to keep track of orientation and location. Original orientation may be measured relative to a built-in compass that reports a direction electronically to a data processing unit. A unit is activated at a known location and orientation or a GPS receiving system is used to find the location of the system users.
  • In one embodiment a plumb bob is included to correct orientation relative to straight down. In one preferred embodiment, the accelerometers are periodically corrected by the compass and the plumb bob, for example every ten seconds, to prevent drift from becoming destructive. Two sets of three orthogonal accelerometers are present, one close to each ear on a head set, and the position of these two sets of accelerometers are compared to find the orientation of the head set, and by implication, the user's head. This comparison can also be used to correct errors, as while the head set is being worn, there must be a fixed relationship between the relative positions of the two sets of accelerometers and the orientation of each set of accelerometers.
  • FIG. 1 is a block diagram of such a system unit 10, which would physically take the familiar form of a headset with a microphone 30 and two earphones 40 and 42. A GPS 12 provides an initial position and updates when possible, and accelerometers 14 and a digital compass and miniature plumb bob 16 fill in when the GPS is not available, and provide greater detail of movement, including head movement. A data processing unit 18 receives and integrates data from the three aforementioned units, and also receives a similar data set from each of the identical units 10 from which unit 10 receives an RF signal by way of antenna 20 and transceiver 22. By receiving the position and head orientation data from the other units and comparing this information to the unit 10 location and orientation information, data processing unit 18 is able to determine if any one of the other users has his head pointed in the direction of unit 10. More specifically, the case of two headset wearers facing each other can be detected, and may be treated as a special case for placing the users into voice contact with one another.
  • Another portion of the RF signal is an encoded representation of the sound entering a microphone 30 and modulated to RF by the transceiver 22 of another headset, which will typically be the voice of another user. The signal will be represented to the user of unit 10 by way of earphones 40 and 42, provided that the range and orientation between units satisfies a predetermined condition. For example, the other user's head is facing the user of unit 10, and/or the other unit is within a designated range.
  • In a preferred embodiment, when the headphone for both ears is used, the orientation of a listener's orientation relative to the place where the sound is originating is used to place a volume difference and time delay into the sound waves generated by the speakers in order to mimic the natural volume difference and time delay that the sound waves would have if received directly from the sound originating location. This permits the user to tell from which direction the sound is originating.
  • In another preferred embodiment of an absolute system, the system is configured to operate in an area in which an RF antenna is broadcasting a directional signal that is received by two sets of three orthogonal inductors, one near each ear of a user. The relative power of reception of each inductor is used to determine the orientation of the inductor set relative to the known orientation of the electromagnetic field.
  • In another preferred embodiment, shown in FIGS. 2-4, both a first headset unit 110 and a second, identical, headset unit 111, form part of the system. For ease of presentation, because the units are identical, both systems (and additional units not shown in FIG. 1, but included within the scope of the invention) are described by the following description of first unit 110.
  • First unit 110 is equipped with a directional speaker 112 that produces sound at a range above that of human hearing, for example in the range of 100,000 Hz (wavelength of about 3 mm) to 1,000,000 Hz (wavelength of about 0.3 mm). Moreover, unit 110 has a data processing unit 114 having a coded signal producer 126 adapted to drive speaker 112 to produce a coded sound signal. Units 110 and 111 also include a right-side microphone 120 and a left-side microphone 122, so that second unit 111 receives sound from the speaker 112 of first unit 110, in both microphone 120 and microphone 122, which send a related electrical signal to logic and control unit 114. Accordingly a coded signal generator 126 of a first unit 110 produces a first sound signal marked with a first code. A second unit 111 upon receiving the first coded signal by way of microphones 120, 122 and decoding it by logic and control unit 114, sends back a second coded signal in a second code, that is dependent on the first code, so that the first unit, upon receiving the second coded sound signal is able to determine that it is the response of the second unit 111 to the recently transmitted sound signal from the first unit 110. The processing time that the second unit 111 takes to transmit the second signal after receiving the first signal is known, and can accordingly be subtracted from the delay between the time the first signal is sent and the second signal is received by the first unit 110, yielding the round trip time, which divided by the round trip speed of sound yields the distance (block 50, FIG. 4). In an alternative embodiment, each signal sent by speaker 112 is time stamped and after reception by another unit 110, this time stamp can be compared with the present time, to arrive at a sound signal travel time and therefore a distance.
  • Unit 110 is also equipped with a mouth microphone 124 positioned to receive the voice of a user 126. The resultant electrical signal from microphone 124 is used, by modulation unit 128, to modulate an RF signal that is broadcast by antenna 130 and received by the matching antenna 130 of the second unit 111. The second unit 111 receives this signal by way of antenna 130, demodulates the signal, and with some further processing that is described below, sends the signal to a right-side headphone 140 and a left-side headphone 142.
  • In addition to determining distance or range between units 110, in one preferred embodiment, the data processing units 114 use the signals received from microphones 120, 122 much as the human brain does with the signal from the ears, to determine the position and orientation of the sound signal originating speaker 112, relative to the microphones 120, 122. This is done by comparing the time delay and volume of the signal received in the right-side microphone 120 versus left-side microphone 122. The signals driving headphones 140 and 142 are modulated to reflect the position of the speaker 112, relative to microphones 120 and 122, to permit the user 126 to detect sound directionality. In an alternative preferred embodiment, the data processing unit 114, does not determine position and orientation of any speaker 112, but simply adds the time delay and volume differences found between the signals from microphones 120 and 122, into the signals going to headphones 140 and 142, respectively, thereby effectively providing sound direction information to the user.
  • In addition, the range information found as described above can be used to determine the volume of sound to be produced by a right-side headphone 140 and a left-side headphone 142. In a preferred embodiment, if the range determined by the sound signal system of speaker 112 and microphones 120 is above a threshold (decision box 154, FIG. 4), no sound responsive to the microphone 124 of the first unit is produced by the headphones 140 and 142 of the second unit 111.
  • The volume of the received sound signal also is helpful in determining how loud the system should represent the sounds associate with the received RF signal from the other unit, as this volume is proportional to range and relative orientation, much as the actual voice sounds from the user are.
  • In a preferred embodiment, the range is divided by the magnitude of the weaker of the two sound signals received by microphones 120 and 122, to form an estimate of the degree to which the first and second users are facing each other (block 152, FIG. 4), and this estimate is compared to a threshold (decision box 154, FIG. 4) to determine if the voice signal from microphone 124 of unit 110 should be presented to the second user 128, by way of the headphones 140 and 142 of the second unit. In another preferred embodiment, the same calculation is performed, but with the stronger of the two signals received by microphones 120 and 122, to determine if first unit 110 is facing second unit 111, regardless of how second unit 111 is facing. This second option would be useful in a commander, commanded situation, for example the workers at a fast food restaurant, enabling a supervisory employee to look at a subordinate, speak and be heard, regardless of whether the subordinate was facing the supervisory employee or whether loud machinery was operating. Also the employees not being faced by the supervisor would not be interrupted or distracted from their tasks or both, by the supervisor's voice on the headphone system. Additionally, a system that does not send a signal to everyone permits a greater degree of privacy in communications.
  • Because the speed of sound is dependent on altitude, in a preferred embodiment the altitude at which the units will be used is entered during an initialization process performed by docking each unit to a computer using a USB connector cord. A user interface then appears on the computer screen and the user is guided to input information such as altitude. Users are advised that if they change altitude by more than 1,500 feet that it is advisable to reenter the altitude information so that the range calculations will not be corrupted by change in speed of sound with altitude. In another preferred embodiment, the wireless communications apparatus of each unit is used to establish a communicative link with a computer and in similar manner the altitude is entered by a user. In another preferred embodiment, a similar coding scheme is used to establish range between units using RF signals. In one preferred embodiment each unit is headphone having a speaker for both ears and a microphone supported near the user's mouth. In an alternative preferred embodiment, some or all of the units may be in the form of a device that clips onto a user's ear and that has a microphone extending toward the user's mouth.
  • In a preferred embodiment of either an absolute or a relative system, auditory noise is discriminated against by comparing the signals received by all of the microphones of the various units. In one embodiment, the data processing unit 114 of each unit does this separately. One method of performing such an operation is to perform fast Fourier transforms of the incoming sound, and filter out those frequencies that are being received in all the microphones, likely background noise, while retaining those frequencies found in a single microphone, which would most likely correspond to the voice of a speaker.
  • In a similar system, dual microphones about a head's width apart are used to receive sound in a speaker phone, and this sound is reconstructed in the earphones of listeners of a telephone connected with the speaker phone, to enable the listeners to tell speakers apart on the basis of their reconstructed locations. Headphones may be used to produce sound on both ends of the telephone connection, with a pair of spaced microphones adapted to receive sound on both ends.
  • In a preferred embodiment each unit also includes a cell phone, so that a user may contact any cell phone equipped person by dialing their telephone number or may also be placed into communication with any other system unit user who is within range, has the correct orientation and has chosen to permit this type of contact. Although possible systems have been described in terms of relative or absolute systems, it falls within the scope of this application to have a system that includes elements of both types of systems. In one preferred embodiment, one or more headsets is configured for a hearing impaired person, with amplification of received sound signals in the frequencies most needed.
  • While a number of exemplary aspects and embodiments have been discussed above, those possessed of skill in the art will recognize certain modifications, permutations, additions and sub-combinations thereof. It is therefore intended that the following appended claims and claims hereafter introduced are interpreted to include all such modifications, permutations, additions and sub-combinations as are within their true spirit and scope.

Claims (13)

1. A method of facilitating communications between a group of persons mutually positioned close to one another, said method comprising:
(a) providing each said person with a microphone, positioned to receive voice sounds from said person, and an earphone;
(b) receiving a microphone signal from each of said microphones;
(c) analyzing said microphone signals to determine a portion that is shared between said microphone signals and attenuating said shared portion; and
(d) directing a signal, based on said microphone signals to at least one of said earphones.
2. The method of claim 1, wherein analyzing said microphone signals includes digitizing said signals.
3. The method of claim 2, wherein analyzing said microphone signals includes iteratively performing digital Fourier transforms on the digitized signals.
4. The method of claim 3, wherein said digital Fourier transforms are fast Fourier transforms.
5. The method of claim 1, wherein analyzing said microphone signals includes iteratively performing Fourier transforms on said microphone signals.
6. The method of claim 5, wherein frequency components that are shared by all said microphone signals are attenuated.
7. The method of claim 6, wherein said frequency components that are shared by all said microphone signals are attenuated to zero.
8. An apparatus for facilitating communications between a group of persons mutually positioned close to one another, said apparatus comprising:
(a) a set of headsets, each headset including at least one microphone and at least one earphone, each said headset shaped to be worn by a user, so that said microphone receives sound from said user's mouth and said earphone emits sound near said user's ear and a radio frequency transceiver, responsive to said microphone and adapted to received radio frequency signals from other headsets; and
(b) a sound filtering unit that analyzes said microphone signals to determine a portion that is shared between said microphone signals and attenuates said shared portion, thereby creating an earphone signal.
9. The apparatus of claim 8, further including an analog-to-digital convertor, to digitize said microphone signals, prior to analysis.
10. The apparatus of claim 9, wherein said sound filtering unit iteratively performs digital Fourier transforms on the digitized signals from said analog-to-digital convertor.
11. The apparatus of claim 10, wherein said digital Fourier transforms are fast Fourier transforms.
12. The apparatus of claim 8, wherein signal filtering unit iteratively performs Fourier transforms on said microphone signals.
13. The apparatus of claim 12, wherein frequency components that are shared by all said microphone signals are attenuated.
US15/002,452 2011-02-25 2016-01-21 Electronic communication system that mimics natural range and orientation dependence Abandoned US20160142834A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/002,452 US20160142834A1 (en) 2011-02-25 2016-01-21 Electronic communication system that mimics natural range and orientation dependence

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US13/035,225 US8761674B2 (en) 2011-02-25 2011-02-25 Electronic communication system that mimics natural range and orientation dependence
US14/276,281 US20140248839A1 (en) 2011-02-25 2014-05-13 Electronic communication system that mimics natural range and orientation dependence
US15/002,452 US20160142834A1 (en) 2011-02-25 2016-01-21 Electronic communication system that mimics natural range and orientation dependence

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/276,281 Continuation US20140248839A1 (en) 2011-02-25 2014-05-13 Electronic communication system that mimics natural range and orientation dependence

Publications (1)

Publication Number Publication Date
US20160142834A1 true US20160142834A1 (en) 2016-05-19

Family

ID=46719323

Family Applications (3)

Application Number Title Priority Date Filing Date
US13/035,225 Expired - Fee Related US8761674B2 (en) 2011-02-25 2011-02-25 Electronic communication system that mimics natural range and orientation dependence
US14/276,281 Abandoned US20140248839A1 (en) 2011-02-25 2014-05-13 Electronic communication system that mimics natural range and orientation dependence
US15/002,452 Abandoned US20160142834A1 (en) 2011-02-25 2016-01-21 Electronic communication system that mimics natural range and orientation dependence

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US13/035,225 Expired - Fee Related US8761674B2 (en) 2011-02-25 2011-02-25 Electronic communication system that mimics natural range and orientation dependence
US14/276,281 Abandoned US20140248839A1 (en) 2011-02-25 2014-05-13 Electronic communication system that mimics natural range and orientation dependence

Country Status (2)

Country Link
US (3) US8761674B2 (en)
WO (1) WO2012115785A2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018199846A1 (en) * 2017-04-23 2018-11-01 Audio Zoom Pte Ltd Transducer apparatus for high speech intelligibility in noisy environments

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150139448A1 (en) * 2013-11-18 2015-05-21 International Business Machines Corporation Location and orientation based volume control
CN108353458B (en) * 2015-06-10 2021-12-03 3M创新有限公司 Communication device and network using TDMA radio communication protocol
WO2019160953A1 (en) * 2018-02-13 2019-08-22 SentiAR, Inc. Intercom system for multiple users
CN109218882B (en) * 2018-08-16 2021-02-26 歌尔科技有限公司 Earphone and ambient sound monitoring method thereof
US11105636B2 (en) * 2019-04-17 2021-08-31 Google Llc Radio enhanced augmented reality and virtual reality with truly wireless earbuds
JP7362320B2 (en) * 2019-07-04 2023-10-17 フォルシアクラリオン・エレクトロニクス株式会社 Audio signal processing device, audio signal processing method, and audio signal processing program

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040057586A1 (en) * 2000-07-27 2004-03-25 Zvi Licht Voice enhancement system

Family Cites Families (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4720857A (en) 1985-12-06 1988-01-19 Plantronics, Inc. Miniaturized headset for two-way voice communication
US5086464A (en) 1990-03-05 1992-02-04 Artic Elements, Inc. Telephone headset for the hearing impaired
US5844984A (en) 1992-03-19 1998-12-01 Pan Communications, Inc. Two-way communications earset with filter
US5615410A (en) 1996-01-16 1997-03-25 Demars; Robert A. Head wear communication system
US7072476B2 (en) 1997-02-18 2006-07-04 Matech, Inc. Audio headset
US6681022B1 (en) 1998-07-22 2004-01-20 Gn Resound North Amerca Corporation Two-way communication earpiece
US6522765B1 (en) 1999-04-02 2003-02-18 Hm Electronics, Inc. Headset communication system and method of using same
US6148175A (en) 1999-06-22 2000-11-14 Freedland; Marat Audio entertainment system
US6516213B1 (en) * 1999-09-03 2003-02-04 Robin Medical, Inc. Method and apparatus to estimate location and orientation of objects during magnetic resonance imaging
GB2382952B (en) 2001-11-28 2005-12-07 Sendo Int Ltd Wireless headset-based communication
US7120388B2 (en) 2002-12-16 2006-10-10 3M Innovative Properties Company Wireless intercom system and method of communicating using wireless intercom system
US20090141923A1 (en) 2005-10-11 2009-06-04 Smith Richard C Earpiece with attached speaker
US20080085030A1 (en) 2006-04-26 2008-04-10 Surefire, Llc Inconspicuous communications assembly
US7106874B2 (en) 2004-07-13 2006-09-12 Motorola, Inc. Method and system for selective coupling of a communication unit to a hearing enhancement device
US20060140435A1 (en) 2004-12-28 2006-06-29 Rosemary Sheehy Headset including boom-actuated microphone switch
US7356473B2 (en) * 2005-01-21 2008-04-08 Lawrence Kates Management and assistance system for the deaf
US7343177B2 (en) * 2005-05-03 2008-03-11 Broadcom Corporation Modular ear-piece/microphone (headset) operable to service voice activated commands
US7327982B2 (en) 2005-07-06 2008-02-05 The Chamberlain Group, Inc. Direct connect wireless headset system
US20070058831A1 (en) 2005-09-13 2007-03-15 Jin Rie Ear ring type two way wireless mini-ear piece
US7689248B2 (en) 2005-09-27 2010-03-30 Nokia Corporation Listening assistance function in phone terminals
US7620429B1 (en) 2005-10-07 2009-11-17 At&T Mobility Ii Llc Hearing assistive system with low power interface
US20070087780A1 (en) 2005-10-14 2007-04-19 Shary Nassimi An Adaptive Wireless Headset System
US20090052703A1 (en) * 2006-04-04 2009-02-26 Aalborg Universitet System and Method Tracking the Position of a Listener and Transmitting Binaural Audio Data to the Listener
US20080102906A1 (en) 2006-10-30 2008-05-01 Phonak Ag Communication system and method of operating the same
US7912449B2 (en) * 2007-06-14 2011-03-22 Broadcom Corporation Method and system for 60 GHz location determination and coordination of WLAN/WPAN/GPS multimode devices
US20090017868A1 (en) 2007-07-13 2009-01-15 Joji Ueda Point-to-Point Wireless Audio Transmission
US20090023417A1 (en) 2007-07-19 2009-01-22 Motorola, Inc. Multiple interactive modes for using multiple earpieces linked to a common mobile handset
KR100821501B1 (en) * 2007-10-17 2008-04-14 김영언 Bi-directional broadcasting system for a small space
JP5095473B2 (en) * 2007-11-15 2012-12-12 ソニー株式会社 Wireless communication apparatus, audio data reproduction method, and program
US20090214064A1 (en) 2008-02-25 2009-08-27 Zounds, Inc. RF power supply for hearing aids
US20100048134A1 (en) 2008-08-19 2010-02-25 Mccarthy Randall T Wireless communication system and communication method with wireless headset
US20100045928A1 (en) 2008-08-25 2010-02-25 Tri-Specs, Inc. Fashion eyewear frame that houses circuitry to effect wireless audio communication while providing extraneous background noise cancellation capability
US20100077020A1 (en) * 2008-09-23 2010-03-25 Nokia Corporation Method, apparatus and computer program product for providing intelligent updates of emission values
WO2010062402A1 (en) 2008-11-26 2010-06-03 Global Market Development, Inc. Integrated telecommunications handset

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040057586A1 (en) * 2000-07-27 2004-03-25 Zvi Licht Voice enhancement system

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018199846A1 (en) * 2017-04-23 2018-11-01 Audio Zoom Pte Ltd Transducer apparatus for high speech intelligibility in noisy environments
US11146884B2 (en) * 2017-04-23 2021-10-12 Audio Zoom Pte Ltd Transducer apparatus for high speech intelligibility in noisy environments

Also Published As

Publication number Publication date
WO2012115785A3 (en) 2012-11-08
WO2012115785A2 (en) 2012-08-30
US20120220237A1 (en) 2012-08-30
US8761674B2 (en) 2014-06-24
US20140248839A1 (en) 2014-09-04

Similar Documents

Publication Publication Date Title
US20160142834A1 (en) Electronic communication system that mimics natural range and orientation dependence
US11051105B2 (en) Locating wireless devices
EP3202160B1 (en) Method of providing hearing assistance between users in an ad hoc network and corresponding system
US9516430B2 (en) Binaural hearing assistance system comprising binaural noise reduction
US8189804B2 (en) Sound provider adapter to cancel out noise
US9344813B2 (en) Methods for operating a hearing device as well as hearing devices
AU2021225419B2 (en) Hearing protector systems
JP6193844B2 (en) Hearing device with selectable perceptual spatial sound source positioning
US11805364B2 (en) Hearing device providing virtual sound
CN112351364B (en) Voice playing method, earphone and storage medium
US8553922B2 (en) Earphone microphone
US20160104501A1 (en) Method and Apparatus for Facilitating Conversation in a Noisy Environment
EP3072314B1 (en) A method of operating a hearing system for conducting telephone calls and a corresponding hearing system
US20230051367A1 (en) Coordinated dichotic sound compression
CN116074665A (en) Distributed microphone in wireless audio system
WO2024153805A1 (en) Methods and systems for maintaining confidentiality of vocal audio

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION