US20140379343A1 - Method, device, and system for audio data processing - Google Patents

Method, device, and system for audio data processing Download PDF

Info

Publication number
US20140379343A1
US20140379343A1 US14/372,233 US201214372233A US2014379343A1 US 20140379343 A1 US20140379343 A1 US 20140379343A1 US 201214372233 A US201214372233 A US 201214372233A US 2014379343 A1 US2014379343 A1 US 2014379343A1
Authority
US
United States
Prior art keywords
user
voice
communication device
vocal
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/372,233
Inventor
Bizhan Karimi-Cherkandi
Farrokh Mohammadzadeh Kouchri
Schah Walli Ali
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
RingCentral Inc
Original Assignee
Unify GmbH and Co KG
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Unify GmbH and Co KG filed Critical Unify GmbH and Co KG
Publication of US20140379343A1 publication Critical patent/US20140379343A1/en
Assigned to UNIFY GMBH & CO. KG reassignment UNIFY GMBH & CO. KG CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS ENTERPRISE COMMUNICATIONS GMBH & CO. KG
Assigned to RINGCENTRAL, INC. reassignment RINGCENTRAL, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RINGCENTRAL IP HOLDINGS, INC.
Assigned to RINGCENTRAL IP HOLDINGS, INC. reassignment RINGCENTRAL IP HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: UNIFY PATENTE GMBH & CO. KG, UNIFY SOFTWARE AND SOLUTIONS GMBH & CO. KG
Assigned to UNIFY PATENTE GMBH & CO. KG reassignment UNIFY PATENTE GMBH & CO. KG CONFIDENTIAL PATENT AGREEMENT Assignors: UNIFY GMBH & CO. KG
Assigned to RINGCENTRAL, INC. reassignment RINGCENTRAL, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RINGCENTRAL IP HOLDINGS, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/247Telephone sets including user guidance or feature selection means facilitating their use
    • H04M1/2474Telephone terminals specially adapted for disabled people
    • H04M1/2475Telephone terminals specially adapted for disabled people for a hearing impaired user

Definitions

  • the present invention relates to communication systems, such as communication systems that utilize cellular phones, mobile computer devices, tablets, laptops, or other terminal devices. More particularly, the present invention relates to a system, method and device configured to edit received audio data to improve voice quality voice data that is received for recording or transmission.
  • Telecommunication endpoints are often equipped with a noise reduction device.
  • a noise reduction device is typically configured to filter voice data received by the device to deliver a better voice content when that data is output or transmitted as outputted audio data.
  • filters used in such noise reduction mechanisms are typically general purpose filters designed to process audio in a generic manner. As a result, the noise reduction may fail to provide a great improvement in voice quality or other attributes to outputted audio data.
  • audio that is outputted may be configured to utilize hearing and speaking parameters that are based on speaking and hearing profiles of the users' performing the speaking and performing the hearing.
  • An embodiment of the method may include the steps of a communication device prompting a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user, the communication device determining voice metrics for the voice of the user based on the sample of the voice of the user, the communication device storing the voice metrics of the voice of the user, the communication device receiving voice data from the user, the communication device modifying the received voice data based on the at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user, and the communication device transmitting the modified voice data to another device so that the modified voice data is outputtable to at least one recipient.
  • the method may include additional steps or modifications.
  • the method may also include the steps of the communication device generating tones for emitting to a user so the user hears the tones, the communication device receiving input that identify volume settings at which the tones are hearable to the user, and the communication device determining a hearing profile of the user based on the received input that identified the volume settings at which the tones are hearable to the user.
  • embodiments of the method may include the step of the communication device adjusting a volume of audio data that is received by the communication device to output the audio data at a louder volume based on the hearing profile of the user.
  • the method may include the steps of the communication device transmitting the hearing profile of the user to another communication device and the other communication device modifying audio data based on the hearing profile of the user prior to transmitting that audio data to the communication device.
  • the other device may modify audio data prior to transmitting the audio data to the user by having a portion of the audio data that is within an audio spectrum that the hearing profile of the user indicates is unhearable to the user omitted from the audio data so that the transmitted audio data does not contain audio within the unhearable audio spectrum of the user.
  • the audio data may be modified by the other communication device by increasing the volume of the audio data so that all the transmitted voice data within the transmitted audio data is within a hearable audio spectrum of the user.
  • the voice metrics of the voice of the user is comprised of parameters that are based on the time domain for the voice, the frequency domain for the voice, the vocal dominants for the voice, and a vocal range of the voice of the user.
  • the communication device may modify the voice data based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants.
  • the vocal range of the voice of the user can be comprised of the communication device removing a portion of the received voice data that is within a frequency range that is outside of the frequency domain of the voice of the user.
  • the communication device modifying the voice data based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user may also be comprised of the communication device applying a parameter that is based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user of the user's voice to noise cancellation processing of the received voice data.
  • the communication device may be any of a number of types of devices.
  • the communication device may be a server, a workstation, a desktop computer, a tablet, a mobile cellular phone, an internet appliance, a laptop computer, a smart phone, a personal digital assistant, a mobile computer device, or a communication terminal.
  • the communication device may also be configured to determine that received voice data is from the user.
  • Some embodiments of the communication system may include a communication system configured to utilize an embodiment of the above discussed methods, which are further discussed below.
  • a non-transitory computer readable medium has an application stored thereon that defines a method that is performed by a communication device when the application is executed by the communication device.
  • the method may include any of the above noted embodiments of a method, which are further discussed below.
  • an embodiment of the method may include the steps of the communication device prompting a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user, the communication device determining voice metrics for the voice of the user based on the sample of the voice of the user, the communication device storing the voice metrics of the voice of the user and associating those saved voice metrics with the user, the communication device receiving voice data from the user, the communication device modifying the received voice data based on the at least one of the of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user, and the communication device transmitting the modified voice data to another device.
  • Some embodiments of the method defined by the application of the computer readable medium may also include the steps of the communication device generating tones for emitting to a user so the user hears the tones, the communication device receiving input that identify volume settings at which the tones are hearable to the user, and the communication device determining a hearing profile of the user based on the received input that identified the volume settings at which the tones are hearable to the user.
  • the communication device may also adjust a volume of audio data that is received by the communication device to output the audio data at a louder volume based on the hearing profile of the user.
  • the communication device may transmit the hearing profile of the user to another communication device and the other communication device may modify the audio data based on the hearing profile of the user prior to transmitting that audio data to the communication device based on the hearing profile of the user.
  • the other device may modify audio data prior to transmitting the audio data to the user by having a portion of the audio data that is within an audio spectrum that the hearing profile of the user indicates is unhearable to the user omitted from the audio data so that the transmitted audio data does not contain audio within the unhearable audio spectrum of the user.
  • the audio data may be modified by the other communication device by increasing the volume of the audio data so that all the transmitted voice data within the transmitted audio data is within a hearable audio spectrum of the user.
  • the communication device modifying the voice data based on the at least one of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user can include a number of steps. For instance, such a modification may be comprised of the communication device removing a portion of the received voice data that is within a frequency range that is outside of the frequency domain of the voice of the user and the communication device applying a parameter that is based on at least one of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user of the user's voice to noise cancellation processing of the received voice data.
  • FIG. 1 is a block diagram of an exemplary embodiment of a communication device, which may be for example a server, a computer device, a workstation, a tablet, a laptop computer, a telephony device, an internet appliance, a personal digital assistant, a smart phone, a cellular phone, a telephone, or other type of communication terminal.
  • a communication device may be for example a server, a computer device, a workstation, a tablet, a laptop computer, a telephony device, an internet appliance, a personal digital assistant, a smart phone, a cellular phone, a telephone, or other type of communication terminal.
  • more than one microphone and more than one speaker may be included within embodiments of the communication device (e.g. some embodiments of a communication device may include a one ear speaker and a loud speaker and may include two microphones, one for surrounding audio detection and one for surrounding noise cancellation).
  • FIG. 2 is a flow chart of an exemplary method for transmitting voice data or audio data.
  • FIG. 3 is a flow chart of an exemplary method for outputting voice data or audio data.
  • FIG. 4 is a block diagram of an exemplary embodiment of a communication system.
  • a communication device may be a computer device such as a server or workstation or a terminal device such as a desktop computer, a tablet, a mobile cellular phone, an internet appliance, a laptop computer, a smart phone, a personal digital assistant, a mobile computer device, or a communication terminal.
  • a network 24 may include the first and second communication terminals 21 , 22 that are communicatively connected via a communication path that is established and maintained via a computer device 25 .
  • the network 24 may be a wide area network or local area network such as an enterprise network in some embodiments of the system.
  • the first and second communication terminals 21 and 22 of the system shown in FIG. 4 may each be an embodiment of the communication device shown in FIG. 1 .
  • the computer device 25 shown in FIG. 4 may function as a telephony type communication device or media communication device such as a switch device, media server, or teleconference server that may be an embodiment of the communication device shown in FIG. 1 .
  • the communication device may include a processor unit 1 that is communicatively connected to a transceiver unit 3 that includes a receiver unit 3 a and a transmitter unit 3 b , a display unit 2 , at least one input device such as a microphone 6 and camera sensor 8 , memory 5 , and a speaker 4 .
  • the transceiver unit 3 may include a receiver unit 3 a and a transmitter unit 3 b for transmission and reception of data via a cellular network, a wireless network connection, or other network connection or may be configured to form a wireless connection via a near field communication, a Bluetooth compliant communication, or other type of wireless communication mechanism or permit a wired connection such as a wired connection made via a universal serial bus connection, for example.
  • the communication device may include additional transceiver units as well.
  • the processor unit 1 may be a central processing unit such as a microprocessor, interconnected microprocessors, or other hardware processor element.
  • the memory 5 may be a non-transitory computer readable medium that has at least one application 7 stored thereon that defines a method that is performed by the communication device, such as for example the method shown in FIG. 2 or 3 .
  • the memory 5 may include a hard drive, a flash drive, an optical disk or a hard disk or other type of computer readable medium that can stores data for a non-transitory period of time. Datasets such as a voice database 12 and hearing database 13 as well as other data that may be utilized by the communication device when running the application 7 may also be stored in the memory 5 .
  • the speaker 4 may include multiple speakers that emit audio and the microphone 6 may be an input device that records audio emitted by a user for receiving that audio data.
  • the microphone 6 may include a first microphone that detects surrounding noise and a second microphone that is used for cancellation of surrounding noise and the speaker 4 may include a first speaker for emitting noise to one human ear and a second speaker that is a loud speaker.
  • the display unit 2 may be a liquid crystal display or other display device.
  • the display unit 2 may be configured as an input device and an output device such as touch screen display that functions to output data visually while also permitting a user to provide input by touching different portions of the display.
  • Other types of input devices such as a button, keyboard, Bluetooth headset or Bluetooth ear-set, or a mouse, may also be connectable to the processor unit 1 so that a user may utilize that input device to provide input to the device.
  • the communication device may run an application 7 via processor unit 1 .
  • the application 7 may define a method that is executed by the communication device when the application is run. Examples of such a method may be appreciated from FIGS. 2 and 3 .
  • the communication device may run the application to create a hearing profile for storing in a hearing database for use in transmitting or outputting received audio or may generate a voice profile for use in a voice database that may be used to enhance voice data received via microphone 6 prior to transmitting that voice data to modify the received voice data to improve the clarity, reduce the noise within the received voice data, or otherwise enhance the voice data so that the transmitted voice data has an enhanced quality.
  • the communication device may run the application and prompt a user to enter voice input via the microphone 6 so that voice data can be collected of the user's voice to define a user voice profile for saving in the voice database 12 .
  • the user's spoken voice may be recorded so that a voice profile that defines a specific filter for the user can be prepared and saved as the user voice profile in the voice database 12 .
  • the voice database 12 may save parameters for sending to a filter when the user is detected as the speaker so that any voice data recorded from that user is filtered by the filter in accordance with the voice parameters defined in the voice database 12 for that user.
  • the communication device may send output to the user via the display unit 2 to prompt the user to provide certain voice data at different volumes or at different speeds to obtain a sample of the user's voice.
  • the sampling of the user's voice may utilize a sampling of the different vocals and frequencies of the user's voice.
  • the prompting output to the user may ask a user to speak a certain phrase at different octaves (e.g. in a low voice or in a high voice) or at different speeds (e.g. slowly and quickly) to record that phrase when spoken by the user when meeting those conditions.
  • the user may also be directed via audio to repeat a specific sentence in his or her preferred language that has significant phonemes for voice analysis.
  • the recorded voice data is subsequently assessed to determine what frequencies a user's spoken voice may not contain and what frequencies are a strength of the user's voice (e.g. the frequency ranges at which the user's voice is most often present when the user speaks).
  • the communication device may determine that a user's voice does not contain specific frequencies.
  • the range of frequencies that are not present within the user's voice may therefore be cut out by a filter so that no sound within those frequencies is output or transmitted to another device when recording the user's voice for sending to another device such as what may be done when the communication device records a video via the camera sensor 8 and microphone 6 or when the user is engaged in a phone call using the communication device.
  • the removal of this frequency range that does not include the user's voice can eliminate noise that may be present during a recording of the user's voice and may reduce the processing needed for performing noise reduction or other voice data cleanup that may be performed on the recorded voice data prior to saving that voice data or transmitting that voice data.
  • the frequency ranges that is a strong range for the user's voice may be determined from the sampling of the user' voice.
  • the filtration of the use's voice in the frequency range that is determined to be a strength for the user's voice may be identified so that such spectrums may be filtered with a finer granularity.
  • a voice profile for a user may be defined and ranged by an assessment of a recorded voice sample that is recorded upon a user speaking words and phrases that he or she is prompted to say via the display unit 2 .
  • the time domain, frequency domain, and dominants of the user's vocals may then be assessed by the communication device analyzing the recorded voice sample.
  • the assessed parameters of the user's voice may then be saved in a voice database 12 for defining the filtering of the user's voice for purposes of enhancing any voice data received from the user that is for saving in association with a video recording or transmitting via a voice channel, media transmission, phone call, or other voice data transmission.
  • the enhancement of the voice data to be transmitted may occur in parallel with noise cancellation that may be performed by the communication device.
  • the noise cancellation and voice enhancement may occur prior to the packetizing of the voice data that is modified via the noise cancellation and voice enhancement.
  • the packetized modified voice data may then be transmitted to another device.
  • the communication device may also prompt a user to provide input to the communication device in response to emitted audio that is output via at least one speaker 4 so that a hearing profile for the user may be determined and stored in a hearing database 13 .
  • the hearing database 13 may then be used in processing received audio data and may also be utilized in processing voice data to be output or transmitted to another device. For instance, a particular user may have a frequency spectrum that the user is less sensible to or unable to hear any audio when that audio falls within that frequency spectrum. Such a spectrum may be identified in the hearing profile of the user and subsequently utilized to modify the voice data to be output to that user or to be transmitted to a device associated with that user such as a user's telephone or cellular phone.
  • the communication device may prompt the user to respond to a series of tones emitted at different levels of loudness via speaker 4 .
  • the input provided by the user may indicate at what volume or loudness level the user is able to hear different emitted tones and may prompt the user to detect the same series of tones emitted at different levels of loudness for each ear individually via a one ear speaker and also for both ears together via a loud speaker.
  • the input provided by the user may then be utilized by the communication device to determine the hearing thresholds for each ear of the user as well as for both ears of the user.
  • the hearing profile of the user may then be linked with the user and utilized in a number of ways. Additionally, the formed hearing profile may also be sent to other devices that may send audio data to the communication device or other device associated with the user so that the user's hearing profile may be utilized to modify the voice data to be sent to the user to enhance the clarity of that data to that user.
  • the hearing profile may be stored in the memory 5 of the communication device by being within a hearing database 13 .
  • the communication device may then adjust the loudness of received voice data so that voice data or other audio data that is not within a frequency to which the user may hear is modified so that the voice data that is output is hearable by the user.
  • the hearing profile may be applied to a real-time transport protocol (“RTP”) decoder or other audio equalizer through the audio path to provide such an adjustment, for example.
  • RTP real-time transport protocol
  • the hearing profile within the user's own communication device may therefore adjust the volume setting at which received data is to be emitted so that it improves the clarity of that data to the user.
  • the hearing profile may be sent to another communication device so that audio data that is transmitted to the user's communication device may be modified by the device transmitting that data to the user to account for the user's specific hearing capabilities via the user's hearing profile.
  • the transmitting communication device that receives the user's hearing profile may enhance its noise reduction to voice data to be transmitted to the communication device associated with the user based on the user's hearing profile.
  • the voice data to be transmitted may be modified so that the voice data is within the spectrum of the user's hearing profile to improve the clarity of the transmitted data to the user.
  • the transmitting device may also, or alternatively, modify the audio data to be transmitted to the user's communication device to avoid, omit, or otherwise nullify audio within the spectrum where the user has no sensibility at all or is near deaf within that spectrum so that bandwidth and processing power associated with the transmission and receipt of that portion of the audio data may be saved.
  • the communication device of the user may send the user's hearing profile or hearing database 13 of the communication device to a transmitting device so that it can be saved by that other device for future calls or other transmissions of audio data.
  • the other device may associate the user's hearing profile with an address of the user's communication device such as a phone number or internet protocol address of that device so that any communication utilizing that address may result in the use of the user's hearing profile to modify voice data or other audio data to be transmitted to that user's communication device.
  • the transmitting device may save that profile for use in future communication sessions as well based upon an association of a device address such as a phone number associated with the user and a call being established with a device having that address.
  • the hearing profile of the user may also be associated with other devices to which the user is associated such as a voice mailbox or email address so that audio data transmitted to these other devices or addresses are modified to improve the clarity of the audio data to the user to which the information is addressed.
  • the user may enter input during a call or other communication session to actuate a sending of a message to another device engaged in that communication session. That input that is entered may indicate to a call partner that the user does not want a hearing profile associated with that user to be applied during a particular session or during any future call or future communication session. Such input may actuate a message to deactivate use of a hearing profile by another device to be sent to this device during an ongoing call or other communication session or may be transmitted. Such a message may alternatively be sent after or before a particular communication session. Such a message may be sent upon a user creating a new hearing profile or voice profile for example. Such a message may be sent along with a replacement hearing profile or voice profile for saving by the other device for use in subsequent communication sessions.
  • voice data or other audio data may result in skewing the audio data to place that data in a part of a spectrum that is more sensible to that user.
  • a skewing may result in the speaker's voice characteristic being changed. But, this may also result in improving the clarity of the communication to the user so that more data is actually heard by the user and understood by the user during a communication session such as a telephone call or conference call.
  • a first communication terminal 21 and second communication terminal 22 may each have a voice database and hearing database for use during a telephone call between the two devices.
  • Each communication terminal may modify voice data of the user speaking into a microphone of that terminal so that the transmitted voice data is modified to improve its clarity to the user of the receiving terminal.
  • a switch device such as a computer device 25 that is within the transmission path of the two terminals during the communication session may have the hearing and voice databases and modify the audio data to be transmitted to a particular user based upon the hearing profile of the user to receive an output of the audio data and the voice profile of the speaker that is transmitting that audio data to the receiving party.
  • more than one user may utilize that device.
  • Different users may create different hearing profiles and different voice profiles.
  • each use may utilize the communication device to have a hearing profile and voice profile created, saved and transmitted as discussed above.
  • the formed profiles of the users may be saved in the memory of the same device and be linked with a user identification to distinguish the different users.
  • Each user may then identify themselves in any of a number of ways. For instance, each use may enter a unique log-on identification or code that identifies that user as the one that is currently using the device.
  • the log-on identification may include entering a personal identification number or pressing certain portions of a touch screen display of the communication device in a predefined sequence of touches to enter the user's reference number or identification code.
  • the communication device may then utilize the profiles within the voice and hearing databases for the identified user.
  • a user may select a specific profile for user by entering input via an input device to select the user specific profiles.
  • a communication device may by default associate a user's hearing and voice profiles with the user identified via an entered identification. A user may then provide input to deactivate such a setting or alter the settings.

Abstract

A method and apparatus that filters audio data received from a speaking person that includes a specific filter for that speaker. The audio characteristics of the speaker's voice may be collected and the specific filter may be formed to reduce noise while also enhancing voice quality. For instance, if a speaker's voice does not contain specific frequencies, then a filter may cancel the noise at such frequencies to ease noise cancellation and reduce processing sound spectrum for cleaning that is not needed. Additionally, the strength frequencies of a speaker's voice may be identified from the collected audio characteristics and those spectrums can be filtered with finer granularity to provide a speaker specific filter that enhances the voice quality of the speaker's voice data that is transmitted or output by a communication device. The audio data may also be output based upon a user's predefined hearing spectrum.

Description

    FIELD OF INVENTION
  • The present invention relates to communication systems, such as communication systems that utilize cellular phones, mobile computer devices, tablets, laptops, or other terminal devices. More particularly, the present invention relates to a system, method and device configured to edit received audio data to improve voice quality voice data that is received for recording or transmission.
  • BACKGROUND OF THE INVENTION
  • Telecommunication endpoints are often equipped with a noise reduction device. Such a device is typically configured to filter voice data received by the device to deliver a better voice content when that data is output or transmitted as outputted audio data. However, filters used in such noise reduction mechanisms are typically general purpose filters designed to process audio in a generic manner. As a result, the noise reduction may fail to provide a great improvement in voice quality or other attributes to outputted audio data.
  • We have determined that a new method and apparatus are needed to improve audio characteristics of audio output that is output based upon received audio data. In some embodiments of our method and apparatus, audio that is outputted may be configured to utilize hearing and speaking parameters that are based on speaking and hearing profiles of the users' performing the speaking and performing the hearing.
  • SUMMARY OF THE INVENTION
  • A method of transmitting voice data is provided. An embodiment of the method may include the steps of a communication device prompting a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user, the communication device determining voice metrics for the voice of the user based on the sample of the voice of the user, the communication device storing the voice metrics of the voice of the user, the communication device receiving voice data from the user, the communication device modifying the received voice data based on the at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user, and the communication device transmitting the modified voice data to another device so that the modified voice data is outputtable to at least one recipient.
  • Some embodiments of the method may include additional steps or modifications. For instance, the method may also include the steps of the communication device generating tones for emitting to a user so the user hears the tones, the communication device receiving input that identify volume settings at which the tones are hearable to the user, and the communication device determining a hearing profile of the user based on the received input that identified the volume settings at which the tones are hearable to the user. Additionally, embodiments of the method may include the step of the communication device adjusting a volume of audio data that is received by the communication device to output the audio data at a louder volume based on the hearing profile of the user.
  • In one embodiment, the method may include the steps of the communication device transmitting the hearing profile of the user to another communication device and the other communication device modifying audio data based on the hearing profile of the user prior to transmitting that audio data to the communication device. In one embodiment, the other device may modify audio data prior to transmitting the audio data to the user by having a portion of the audio data that is within an audio spectrum that the hearing profile of the user indicates is unhearable to the user omitted from the audio data so that the transmitted audio data does not contain audio within the unhearable audio spectrum of the user. In other embodiments, the audio data may be modified by the other communication device by increasing the volume of the audio data so that all the transmitted voice data within the transmitted audio data is within a hearable audio spectrum of the user.
  • For certain embodiments, the voice metrics of the voice of the user is comprised of parameters that are based on the time domain for the voice, the frequency domain for the voice, the vocal dominants for the voice, and a vocal range of the voice of the user. The communication device may modify the voice data based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants. The vocal range of the voice of the user can be comprised of the communication device removing a portion of the received voice data that is within a frequency range that is outside of the frequency domain of the voice of the user. The communication device modifying the voice data based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user may also be comprised of the communication device applying a parameter that is based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user of the user's voice to noise cancellation processing of the received voice data.
  • The communication device may be any of a number of types of devices. For instance, the communication device may be a server, a workstation, a desktop computer, a tablet, a mobile cellular phone, an internet appliance, a laptop computer, a smart phone, a personal digital assistant, a mobile computer device, or a communication terminal. In some embodiments, the communication device may also be configured to determine that received voice data is from the user.
  • Communication systems are also provided. Some embodiments of the communication system may include a communication system configured to utilize an embodiment of the above discussed methods, which are further discussed below.
  • A non-transitory computer readable medium is also provided. The non-transitory computer readable medium has an application stored thereon that defines a method that is performed by a communication device when the application is executed by the communication device. The method may include any of the above noted embodiments of a method, which are further discussed below. For instance, an embodiment of the method may include the steps of the communication device prompting a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user, the communication device determining voice metrics for the voice of the user based on the sample of the voice of the user, the communication device storing the voice metrics of the voice of the user and associating those saved voice metrics with the user, the communication device receiving voice data from the user, the communication device modifying the received voice data based on the at least one of the of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user, and the communication device transmitting the modified voice data to another device.
  • Some embodiments of the method defined by the application of the computer readable medium may also include the steps of the communication device generating tones for emitting to a user so the user hears the tones, the communication device receiving input that identify volume settings at which the tones are hearable to the user, and the communication device determining a hearing profile of the user based on the received input that identified the volume settings at which the tones are hearable to the user. The communication device may also adjust a volume of audio data that is received by the communication device to output the audio data at a louder volume based on the hearing profile of the user. The communication device may transmit the hearing profile of the user to another communication device and the other communication device may modify the audio data based on the hearing profile of the user prior to transmitting that audio data to the communication device based on the hearing profile of the user. In one embodiment, the other device may modify audio data prior to transmitting the audio data to the user by having a portion of the audio data that is within an audio spectrum that the hearing profile of the user indicates is unhearable to the user omitted from the audio data so that the transmitted audio data does not contain audio within the unhearable audio spectrum of the user. In other embodiments, the audio data may be modified by the other communication device by increasing the volume of the audio data so that all the transmitted voice data within the transmitted audio data is within a hearable audio spectrum of the user.
  • In certain embodiments of the method defined by the application, the communication device modifying the voice data based on the at least one of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user can include a number of steps. For instance, such a modification may be comprised of the communication device removing a portion of the received voice data that is within a frequency range that is outside of the frequency domain of the voice of the user and the communication device applying a parameter that is based on at least one of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user of the user's voice to noise cancellation processing of the received voice data.
  • Other details, objects, and advantages of the invention will become apparent as the following description of certain present preferred embodiments thereof and certain present preferred methods of practicing the same proceeds.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Present preferred devices, systems, and apparatuses for providing communications between multiple communication devices are shown in the accompanying drawings and certain present preferred methods of practicing the same are also illustrated therein. It should be understood that like reference numbers used in the drawings may identify like components.
  • FIG. 1 is a block diagram of an exemplary embodiment of a communication device, which may be for example a server, a computer device, a workstation, a tablet, a laptop computer, a telephony device, an internet appliance, a personal digital assistant, a smart phone, a cellular phone, a telephone, or other type of communication terminal. It should be understood that more than one microphone and more than one speaker may be included within embodiments of the communication device (e.g. some embodiments of a communication device may include a one ear speaker and a loud speaker and may include two microphones, one for surrounding audio detection and one for surrounding noise cancellation).
  • FIG. 2 is a flow chart of an exemplary method for transmitting voice data or audio data.
  • FIG. 3 is a flow chart of an exemplary method for outputting voice data or audio data.
  • FIG. 4 is a block diagram of an exemplary embodiment of a communication system.
  • DETAILED DESCRIPTION OF PRESENT PREFERRED EMBODIMENTS
  • Referring to FIGS. 1-4, a communication device may be a computer device such as a server or workstation or a terminal device such as a desktop computer, a tablet, a mobile cellular phone, an internet appliance, a laptop computer, a smart phone, a personal digital assistant, a mobile computer device, or a communication terminal. In one embodiment of a system, a network 24 may include the first and second communication terminals 21, 22 that are communicatively connected via a communication path that is established and maintained via a computer device 25. The network 24 may be a wide area network or local area network such as an enterprise network in some embodiments of the system.
  • The first and second communication terminals 21 and 22 of the system shown in FIG. 4 may each be an embodiment of the communication device shown in FIG. 1. Alternatively, the computer device 25 shown in FIG. 4 may function as a telephony type communication device or media communication device such as a switch device, media server, or teleconference server that may be an embodiment of the communication device shown in FIG. 1.
  • The communication device may include a processor unit 1 that is communicatively connected to a transceiver unit 3 that includes a receiver unit 3 a and a transmitter unit 3 b, a display unit 2, at least one input device such as a microphone 6 and camera sensor 8, memory 5, and a speaker 4. The transceiver unit 3 may include a receiver unit 3 a and a transmitter unit 3 b for transmission and reception of data via a cellular network, a wireless network connection, or other network connection or may be configured to form a wireless connection via a near field communication, a Bluetooth compliant communication, or other type of wireless communication mechanism or permit a wired connection such as a wired connection made via a universal serial bus connection, for example. In some embodiments, the communication device may include additional transceiver units as well.
  • The processor unit 1 may be a central processing unit such as a microprocessor, interconnected microprocessors, or other hardware processor element. The memory 5 may be a non-transitory computer readable medium that has at least one application 7 stored thereon that defines a method that is performed by the communication device, such as for example the method shown in FIG. 2 or 3. The memory 5 may include a hard drive, a flash drive, an optical disk or a hard disk or other type of computer readable medium that can stores data for a non-transitory period of time. Datasets such as a voice database 12 and hearing database 13 as well as other data that may be utilized by the communication device when running the application 7 may also be stored in the memory 5.
  • The speaker 4 may include multiple speakers that emit audio and the microphone 6 may be an input device that records audio emitted by a user for receiving that audio data. For instance, the microphone 6 may include a first microphone that detects surrounding noise and a second microphone that is used for cancellation of surrounding noise and the speaker 4 may include a first speaker for emitting noise to one human ear and a second speaker that is a loud speaker. The display unit 2 may be a liquid crystal display or other display device. The display unit 2 may be configured as an input device and an output device such as touch screen display that functions to output data visually while also permitting a user to provide input by touching different portions of the display. Other types of input devices such as a button, keyboard, Bluetooth headset or Bluetooth ear-set, or a mouse, may also be connectable to the processor unit 1 so that a user may utilize that input device to provide input to the device.
  • The communication device may run an application 7 via processor unit 1. The application 7 may define a method that is executed by the communication device when the application is run. Examples of such a method may be appreciated from FIGS. 2 and 3. For example, the communication device may run the application to create a hearing profile for storing in a hearing database for use in transmitting or outputting received audio or may generate a voice profile for use in a voice database that may be used to enhance voice data received via microphone 6 prior to transmitting that voice data to modify the received voice data to improve the clarity, reduce the noise within the received voice data, or otherwise enhance the voice data so that the transmitted voice data has an enhanced quality.
  • For instance, the communication device may run the application and prompt a user to enter voice input via the microphone 6 so that voice data can be collected of the user's voice to define a user voice profile for saving in the voice database 12. The user's spoken voice may be recorded so that a voice profile that defines a specific filter for the user can be prepared and saved as the user voice profile in the voice database 12. Alternatively, the voice database 12 may save parameters for sending to a filter when the user is detected as the speaker so that any voice data recorded from that user is filtered by the filter in accordance with the voice parameters defined in the voice database 12 for that user.
  • The communication device may send output to the user via the display unit 2 to prompt the user to provide certain voice data at different volumes or at different speeds to obtain a sample of the user's voice. The sampling of the user's voice may utilize a sampling of the different vocals and frequencies of the user's voice. For instance, the prompting output to the user may ask a user to speak a certain phrase at different octaves (e.g. in a low voice or in a high voice) or at different speeds (e.g. slowly and quickly) to record that phrase when spoken by the user when meeting those conditions. The user may also be directed via audio to repeat a specific sentence in his or her preferred language that has significant phonemes for voice analysis. The recorded voice data is subsequently assessed to determine what frequencies a user's spoken voice may not contain and what frequencies are a strength of the user's voice (e.g. the frequency ranges at which the user's voice is most often present when the user speaks).
  • After the user's voice is assessed, the communication device may determine that a user's voice does not contain specific frequencies. The range of frequencies that are not present within the user's voice may therefore be cut out by a filter so that no sound within those frequencies is output or transmitted to another device when recording the user's voice for sending to another device such as what may be done when the communication device records a video via the camera sensor 8 and microphone 6 or when the user is engaged in a phone call using the communication device. The removal of this frequency range that does not include the user's voice can eliminate noise that may be present during a recording of the user's voice and may reduce the processing needed for performing noise reduction or other voice data cleanup that may be performed on the recorded voice data prior to saving that voice data or transmitting that voice data.
  • Additionally, the frequency ranges that is a strong range for the user's voice (e.g. a range in which the user's voice is almost always present) may be determined from the sampling of the user' voice. The filtration of the use's voice in the frequency range that is determined to be a strength for the user's voice may be identified so that such spectrums may be filtered with a finer granularity.
  • In one embodiment, a voice profile for a user may be defined and ranged by an assessment of a recorded voice sample that is recorded upon a user speaking words and phrases that he or she is prompted to say via the display unit 2. The time domain, frequency domain, and dominants of the user's vocals may then be assessed by the communication device analyzing the recorded voice sample. The assessed parameters of the user's voice may then be saved in a voice database 12 for defining the filtering of the user's voice for purposes of enhancing any voice data received from the user that is for saving in association with a video recording or transmitting via a voice channel, media transmission, phone call, or other voice data transmission. It should be understood that the enhancement of the voice data to be transmitted may occur in parallel with noise cancellation that may be performed by the communication device. The noise cancellation and voice enhancement may occur prior to the packetizing of the voice data that is modified via the noise cancellation and voice enhancement. The packetized modified voice data may then be transmitted to another device.
  • The communication device may also prompt a user to provide input to the communication device in response to emitted audio that is output via at least one speaker 4 so that a hearing profile for the user may be determined and stored in a hearing database 13. The hearing database 13 may then be used in processing received audio data and may also be utilized in processing voice data to be output or transmitted to another device. For instance, a particular user may have a frequency spectrum that the user is less sensible to or unable to hear any audio when that audio falls within that frequency spectrum. Such a spectrum may be identified in the hearing profile of the user and subsequently utilized to modify the voice data to be output to that user or to be transmitted to a device associated with that user such as a user's telephone or cellular phone.
  • For example, the communication device may prompt the user to respond to a series of tones emitted at different levels of loudness via speaker 4. The input provided by the user may indicate at what volume or loudness level the user is able to hear different emitted tones and may prompt the user to detect the same series of tones emitted at different levels of loudness for each ear individually via a one ear speaker and also for both ears together via a loud speaker. The input provided by the user may then be utilized by the communication device to determine the hearing thresholds for each ear of the user as well as for both ears of the user. The hearing profile of the user may then be linked with the user and utilized in a number of ways. Additionally, the formed hearing profile may also be sent to other devices that may send audio data to the communication device or other device associated with the user so that the user's hearing profile may be utilized to modify the voice data to be sent to the user to enhance the clarity of that data to that user.
  • For example, the hearing profile may be stored in the memory 5 of the communication device by being within a hearing database 13. The communication device may then adjust the loudness of received voice data so that voice data or other audio data that is not within a frequency to which the user may hear is modified so that the voice data that is output is hearable by the user. The hearing profile may be applied to a real-time transport protocol (“RTP”) decoder or other audio equalizer through the audio path to provide such an adjustment, for example. The hearing profile within the user's own communication device may therefore adjust the volume setting at which received data is to be emitted so that it improves the clarity of that data to the user.
  • As another example, the hearing profile may be sent to another communication device so that audio data that is transmitted to the user's communication device may be modified by the device transmitting that data to the user to account for the user's specific hearing capabilities via the user's hearing profile. For instance, the transmitting communication device that receives the user's hearing profile may enhance its noise reduction to voice data to be transmitted to the communication device associated with the user based on the user's hearing profile. The voice data to be transmitted may be modified so that the voice data is within the spectrum of the user's hearing profile to improve the clarity of the transmitted data to the user. The transmitting device may also, or alternatively, modify the audio data to be transmitted to the user's communication device to avoid, omit, or otherwise nullify audio within the spectrum where the user has no sensibility at all or is near deaf within that spectrum so that bandwidth and processing power associated with the transmission and receipt of that portion of the audio data may be saved.
  • The communication device of the user may send the user's hearing profile or hearing database 13 of the communication device to a transmitting device so that it can be saved by that other device for future calls or other transmissions of audio data. The other device may associate the user's hearing profile with an address of the user's communication device such as a phone number or internet protocol address of that device so that any communication utilizing that address may result in the use of the user's hearing profile to modify voice data or other audio data to be transmitted to that user's communication device. The transmitting device may save that profile for use in future communication sessions as well based upon an association of a device address such as a phone number associated with the user and a call being established with a device having that address. The hearing profile of the user may also be associated with other devices to which the user is associated such as a voice mailbox or email address so that audio data transmitted to these other devices or addresses are modified to improve the clarity of the audio data to the user to which the information is addressed.
  • In some embodiments of the communication device, the user may enter input during a call or other communication session to actuate a sending of a message to another device engaged in that communication session. That input that is entered may indicate to a call partner that the user does not want a hearing profile associated with that user to be applied during a particular session or during any future call or future communication session. Such input may actuate a message to deactivate use of a hearing profile by another device to be sent to this device during an ongoing call or other communication session or may be transmitted. Such a message may alternatively be sent after or before a particular communication session. Such a message may be sent upon a user creating a new hearing profile or voice profile for example. Such a message may be sent along with a replacement hearing profile or voice profile for saving by the other device for use in subsequent communication sessions.
  • It should be understood that the modification of voice data or other audio data to improve the clarity of that data to a particular user when that data is output to the user may result in skewing the audio data to place that data in a part of a spectrum that is more sensible to that user. Such a skewing may result in the speaker's voice characteristic being changed. But, this may also result in improving the clarity of the communication to the user so that more data is actually heard by the user and understood by the user during a communication session such as a telephone call or conference call.
  • It should be appreciated that a first communication terminal 21 and second communication terminal 22 may each have a voice database and hearing database for use during a telephone call between the two devices. Each communication terminal may modify voice data of the user speaking into a microphone of that terminal so that the transmitted voice data is modified to improve its clarity to the user of the receiving terminal. In one contemplated alternative embodiment, a switch device such as a computer device 25 that is within the transmission path of the two terminals during the communication session may have the hearing and voice databases and modify the audio data to be transmitted to a particular user based upon the hearing profile of the user to receive an output of the audio data and the voice profile of the speaker that is transmitting that audio data to the receiving party.
  • In some embodiments of the communication device, more than one user may utilize that device. Different users may create different hearing profiles and different voice profiles. For instance, each use may utilize the communication device to have a hearing profile and voice profile created, saved and transmitted as discussed above. The formed profiles of the users may be saved in the memory of the same device and be linked with a user identification to distinguish the different users. Each user may then identify themselves in any of a number of ways. For instance, each use may enter a unique log-on identification or code that identifies that user as the one that is currently using the device. The log-on identification may include entering a personal identification number or pressing certain portions of a touch screen display of the communication device in a predefined sequence of touches to enter the user's reference number or identification code. Upon detection of such an entered user identification, the communication device may then utilize the profiles within the voice and hearing databases for the identified user. As another option, a user may select a specific profile for user by entering input via an input device to select the user specific profiles. In some embodiments, a communication device may by default associate a user's hearing and voice profiles with the user identified via an entered identification. A user may then provide input to deactivate such a setting or alter the settings.
  • While certain present preferred embodiments of the communication device, communication system, and methods of making and using the same have been shown and described above, it is to be distinctly understood that the invention is not limited thereto but may be otherwise variously embodied and practiced within the scope of the following claims.

Claims (20)

1. A method of transmitting voice data comprising:
a communication device prompting a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user;
the communication device determining voice metrics for the voice of the user based on the sample of the voice of the user;
the communication device storing the voice metrics of the voice of the user;
the communication device receiving voice data from the user;
the communication device modifying the received voice data based on the at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user; and
the communication device transmitting the modified voice data to another device so that the modified voice data is outputtable to at least one recipient.
2. The method of claim 1 further comprising:
the communication device generating tones for emitting to a user so the user hears the tones;
the communication device receiving input that identify volume settings at which the tones are hearable to the user; and
the communication device determining a hearing profile of the user based on the received input that identified the volume settings at which the tones are hearable to the user.
3. The method of claim 2 further comprising:
the communication device adjusting a volume of audio data that is received by the communication device to output the audio data at a louder volume based on the hearing profile of the user.
4. The method of claim 2 further comprising:
the communication device transmitting the hearing profile of the user to another communication device;
the other communication device modifying audio data based on the hearing profile of the user prior to transmitting that audio data to the communication device.
5. The method of claim 4 wherein the voice metrics of the voice of the user is comprised of parameters that are based on the time domain for the voice, the frequency domain for the voice, the vocal dominants for the voice, and a vocal range of the voice of the user; and
wherein the other communication device modifying the audio data based on the hearing profile of the user prior to transmitting that audio data to the communication device comprises omitting a portion of the audio data that is within an audio spectrum identified as being unhearable to a user via the hearing profile of the user so that the transmitted audio data does not include that portion of the audio data.
6. The method of claim 1 wherein the communication device modifying the voice data based on the at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user is comprised of the communication device removing a portion of the received voice data that is within a frequency range that is outside of the frequency domain of the voice of the user.
7. The method of claim 6 wherein the communication device modifying the voice data based on the at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user is also comprised of the communication device applying a parameter that is based on at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user of the user's voice to noise cancellation processing of the received voice data.
8. The method of claim 1 wherein the communication device is one of a server, a workstation, a desktop computer, a tablet, a mobile cellular phone, an internet appliance, a laptop computer, a smart phone, a personal digital assistant, a mobile computer device, or a communication terminal and wherein the method further comprising:
the communication device determining that the received voice data is from the user.
9. A communication system comprising:
a communication device configured to prompt a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user;
the communication device configured to determine voice metrics for the voice of the user based on the sample of the voice of the user;
the communication device configured to store the voice metrics of the voice of the user in non-transitory memory;
the communication device configured to receive voice data from the user;
the communication device configured to modify the received voice data based on the at least one of the voice metrics, the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user; and
the communication device configured to transmit the modified voice data so that the modified voice data is outputtable to at least one recipient.
10. A non-transitory computer readable medium having an application stored thereon that defines a method that is performed by a communication device when the application is executed by the communication device, the method comprising:
the communication device prompting a user to speak to obtain a sample of a voice of the user to identify at least one of time domain, frequency domain, vocal dominants, and a vocal range of the voice of the user;
the communication device determining voice metrics for the voice of the user based on the sample of the voice of the user;
the communication device storing the voice metrics of the voice of the user and associating those saved voice metrics with the user;
the communication device receiving voice data from the user;
the communication device modifying the received voice data based on the at least one of the of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user; and
the communication device transmitting the modified voice data to another device.
11. The computer readable medium of claim 10 wherein the method further comprising:
the communication device generating tones for emitting to a user so the user hears the tones;
the communication device receiving input that identify volume settings at which the tones are hearable to the user; and
the communication device determining a hearing profile of the user based on the received input that identified the volume settings at which the tones are hearable to the user.
12. The computer readable medium of claim 11 wherein the method further comprising:
the communication device adjusting a volume of audio data that is received by the communication device to output the audio data at a louder volume based on the hearing profile of the user.
13. The computer readable medium of claim 11 wherein the communication device is a first communication device and method further comprising:
the communication device transmitting the hearing profile of the user to a second communication device; and
the second communication device modifying the audio data based on the hearing profile of the user prior to transmitting that audio data to the first communication device such that the modified audio data that is transmitted to the first communication device omits any portion of the audio data that is within a first audio spectrum, the first audio spectrum being an audio spectrum identified by the hearing profile of the user as being unhearable audio to the user.
14. The computer readable medium of claim 10 wherein the communication device modifying the voice data based on the at least one of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user is comprised of the communication device removing a portion of the received voice data that is within a frequency range that is outside of the frequency domain of the voice of the user.
15. The computer readable medium of claim 14 wherein the communication device modifying the voice data based on the at least one of the of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user is also comprised of the communication device applying a parameter that is based on at least one of the time domain, the frequency domain, the vocal dominants, and the vocal range of the voice of the user of the user's voice to noise cancellation processing of the received voice data.
16. The communication system of claim 9 wherein the communication device comprises hardware, the hardware comprising a processor and a transceiver unit, and wherein at least one of a display device, a microphone, and a camera sensor is connected to the communication device.
17. The communication system of claim 9 wherein the communication device comprises hardware, the hardware comprising a processor, a transceiver unit, a display device, a microphone, and a camera sensor.
18. The communication system of claim 9 comprising:
a first communication terminal and a second communication terminal, the communication device being communicatively connected to the first and second communication terminals.
19. The communication system of claim 18 wherein the user is a user of the first communication terminal and the recipient is a user of the second communication terminal.
20. The communication system of claim 19 wherein the communication device is configured to transmit the modified voice data to the second communication terminal so that the modified voice data is outputtable to the recipient.
US14/372,233 2012-11-20 2012-11-20 Method, device, and system for audio data processing Abandoned US20140379343A1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2012/065995 WO2014081408A1 (en) 2012-11-20 2012-11-20 Method, device, and system for audio data processing

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2012/065995 A-371-Of-International WO2014081408A1 (en) 2012-11-20 2012-11-20 Method, device, and system for audio data processing

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/665,537 Continuation US10325612B2 (en) 2012-11-20 2017-08-01 Method, device, and system for audio data processing

Publications (1)

Publication Number Publication Date
US20140379343A1 true US20140379343A1 (en) 2014-12-25

Family

ID=47326378

Family Applications (4)

Application Number Title Priority Date Filing Date
US14/372,233 Abandoned US20140379343A1 (en) 2012-11-20 2012-11-20 Method, device, and system for audio data processing
US15/665,537 Active US10325612B2 (en) 2012-11-20 2017-08-01 Method, device, and system for audio data processing
US16/400,237 Active US10803880B2 (en) 2012-11-20 2019-05-01 Method, device, and system for audio data processing
US17/018,647 Abandoned US20200411025A1 (en) 2012-11-20 2020-09-11 Method, device, and system for audio data processing

Family Applications After (3)

Application Number Title Priority Date Filing Date
US15/665,537 Active US10325612B2 (en) 2012-11-20 2017-08-01 Method, device, and system for audio data processing
US16/400,237 Active US10803880B2 (en) 2012-11-20 2019-05-01 Method, device, and system for audio data processing
US17/018,647 Abandoned US20200411025A1 (en) 2012-11-20 2020-09-11 Method, device, and system for audio data processing

Country Status (6)

Country Link
US (4) US20140379343A1 (en)
EP (1) EP2786376A1 (en)
KR (1) KR101626438B1 (en)
CN (1) CN104160443B (en)
TW (1) TWI527024B (en)
WO (1) WO2014081408A1 (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150011192A1 (en) * 2013-07-05 2015-01-08 Kuo-Ping Yang Method of processing telephone signals and electronic device thereof
US20150078575A1 (en) * 2013-02-11 2015-03-19 Symphonic Audio Technologies Corp. Audio apparatus and methods
US20150356980A1 (en) * 2013-01-15 2015-12-10 Sony Corporation Storage control device, playback control device, and recording medium
US9319019B2 (en) 2013-02-11 2016-04-19 Symphonic Audio Technologies Corp. Method for augmenting a listening experience
US9344815B2 (en) 2013-02-11 2016-05-17 Symphonic Audio Technologies Corp. Method for augmenting hearing
US20160239253A1 (en) * 2014-01-31 2016-08-18 Matteo Staffaroni Method for audio correction in electronic devices
US20170131961A1 (en) * 2015-11-10 2017-05-11 Optim Corporation System and method for sharing screen
US9973627B1 (en) * 2017-01-25 2018-05-15 Sorenson Ip Holdings, Llc Selecting audio profiles
US20180166103A1 (en) * 2016-12-09 2018-06-14 Baidu Online Network Technology (Beijing) Co., Ltd. Method and device for processing speech based on artificial intelligence
US20180271710A1 (en) * 2017-03-22 2018-09-27 Bragi GmbH Wireless earpiece for tinnitus therapy
US10121488B1 (en) * 2015-02-23 2018-11-06 Sprint Communications Company L.P. Optimizing call quality using vocal frequency fingerprints to filter voice calls
US20180341643A1 (en) * 2017-05-26 2018-11-29 Bose Corporation Dynamic text-to-speech response from a smart speaker
US20190278445A1 (en) * 2014-01-28 2019-09-12 International Business Machines Corporation Impairment-adaptive electronic data interaction system
US10564928B2 (en) * 2017-06-02 2020-02-18 Rovi Guides, Inc. Systems and methods for generating a volume- based response for multiple voice-operated user devices
US20200273477A1 (en) * 2019-02-21 2020-08-27 International Business Machines Corporation Dynamic communication session filtering
US20210152880A1 (en) * 2019-11-14 2021-05-20 Dish Network L.L.C. Method and system for adaptive audio modification
US11205440B2 (en) * 2018-12-28 2021-12-21 Pixart Imaging Inc. Sound playback system and output sound adjusting method thereof
US11475907B2 (en) * 2017-11-27 2022-10-18 Goertek Technology Co., Ltd. Method and device of denoising voice signal
US20220351733A1 (en) * 2018-12-12 2022-11-03 Sonos, Inc. Guest Access for Voice Control of Playback Devices

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10051120B2 (en) * 2013-12-20 2018-08-14 Ultratec, Inc. Communication device and methods for use by hearing impaired
CN104954555B (en) 2015-05-18 2018-10-16 百度在线网络技术(北京)有限公司 A kind of volume adjusting method and system
US9407989B1 (en) 2015-06-30 2016-08-02 Arthur Woodrow Closed audio circuit
CN105788612B (en) * 2016-03-31 2019-11-05 广州酷狗计算机科技有限公司 A kind of method and apparatus detecting sound quality
CN108877822A (en) * 2017-05-09 2018-11-23 塞舌尔商元鼎音讯股份有限公司 Intelligent voice system, the method for voice output adjustment and computer-readable memory media
US10991379B2 (en) * 2018-06-22 2021-04-27 Babblelabs Llc Data driven audio enhancement
CN108986830B (en) * 2018-08-28 2021-02-09 安徽淘云科技有限公司 Audio corpus screening method and device
TW202102010A (en) * 2019-05-24 2021-01-01 瑞典商赫爾桑茲股份有限公司 Methods, devices and computer program products for lossless data compression and decompression
US11257480B2 (en) * 2020-03-03 2022-02-22 Tencent America LLC Unsupervised singing voice conversion with pitch adversarial network
DE102021103310B4 (en) 2021-02-12 2024-01-04 Dr. Ing. H.C. F. Porsche Aktiengesellschaft METHOD AND DEVICE FOR IMPROVING SPEECH UNDERSTANDABILITY IN A ROOM
TWI801085B (en) * 2022-01-07 2023-05-01 矽響先創科技股份有限公司 Method of noise reduction for intelligent network communication

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4296277A (en) * 1978-09-26 1981-10-20 Feller Ag Electronic voice detector
US4538295A (en) * 1982-08-16 1985-08-27 Nissan Motor Company, Limited Speech recognition system for an automotive vehicle
US20030028385A1 (en) * 2001-06-30 2003-02-06 Athena Christodoulou Audio reproduction and personal audio profile gathering apparatus and method
US6820053B1 (en) * 1999-10-06 2004-11-16 Dietmar Ruwisch Method and apparatus for suppressing audible noise in speech transmission
US6944474B2 (en) * 2001-09-20 2005-09-13 Sound Id Sound enhancement for mobile phones and other products producing personalized audio for users
US20060282264A1 (en) * 2005-06-09 2006-12-14 Bellsouth Intellectual Property Corporation Methods and systems for providing noise filtering using speech recognition
US20070225984A1 (en) * 2006-03-23 2007-09-27 Microsoft Corporation Digital voice profiles
US20090086933A1 (en) * 2007-10-01 2009-04-02 Labhesh Patel Call routing using voice signature and hearing characteristics
US20090185704A1 (en) * 2008-01-21 2009-07-23 Bernafon Ag Hearing aid adapted to a specific type of voice in an acoustical environment, a method and use
US20100094619A1 (en) * 2008-10-15 2010-04-15 Verizon Business Network Services Inc. Audio frequency remapping
US20110178799A1 (en) * 2008-07-25 2011-07-21 The Board Of Trustees Of The University Of Illinois Methods and systems for identifying speech sounds using multi-dimensional analysis
US8010366B1 (en) * 2007-03-20 2011-08-30 Neurotone, Inc. Personal hearing suite
US20110282669A1 (en) * 2010-05-17 2011-11-17 Avaya Inc. Estimating a Listener's Ability To Understand a Speaker, Based on Comparisons of Their Styles of Speech
US20120215532A1 (en) * 2011-02-22 2012-08-23 Apple Inc. Hearing assistance system for providing consistent human speech
US8369549B2 (en) * 2010-03-23 2013-02-05 Audiotoniq, Inc. Hearing aid system adapted to selectively amplify audio signals
US20130339025A1 (en) * 2011-05-03 2013-12-19 Suhami Associates Ltd. Social network with enhanced audio communications for the Hearing impaired
US20150269953A1 (en) * 2012-10-16 2015-09-24 Audiologicall, Ltd. Audio signal manipulation for speech enhancement before sound reproduction
US9468401B2 (en) * 2010-08-05 2016-10-18 Ace Communications Limited Method and system for self-managed sound enhancement

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5706352A (en) 1993-04-07 1998-01-06 K/S Himpp Adaptive gain and filtering circuit for a sound reproduction system
US7016507B1 (en) 1997-04-16 2006-03-21 Ami Semiconductor Inc. Method and apparatus for noise reduction particularly in hearing aids
US5844983A (en) * 1997-07-10 1998-12-01 Ericsson Inc. Method and apparatus for controlling a telephone ring signal
AU1024600A (en) * 1998-11-09 2000-05-29 Hans Kunov System and method for processing low signal-to-noise ratio signals
US6463128B1 (en) * 1999-09-29 2002-10-08 Denso Corporation Adjustable coding detection in a portable telephone
DE19956747C1 (en) * 1999-11-25 2001-01-11 Siemens Ag Speech recognition method for telecommunications system
DE60204902T2 (en) * 2001-10-05 2006-05-11 Oticon A/S Method for programming a communication device and programmable communication device
US20050119894A1 (en) * 2003-10-20 2005-06-02 Cutler Ann R. System and process for feedback speech instruction
CA2454296A1 (en) 2003-12-29 2005-06-29 Nokia Corporation Method and device for speech enhancement in the presence of background noise
CN100440899C (en) * 2004-11-15 2008-12-03 北京中星微电子有限公司 A method and apparatus for automatic regulation of mobile phone volume
CN1855704B (en) * 2005-04-26 2011-07-20 佛山市顺德区顺达电脑厂有限公司 Method for automatic volume regulation of electronic device
WO2006130909A1 (en) * 2005-06-10 2006-12-14 Neuromonics Pty Ltd Digital playback device and method and apparatus for spectrally modifying a digital audio signal
DE102006015497B4 (en) * 2006-04-03 2008-01-10 Nebel, Wolfgang, Dr. Audio system and method and computer program and data carrier containing the computer program for adapting the transfer function of an audio system by means of voice control
US7616936B2 (en) * 2006-12-14 2009-11-10 Cisco Technology, Inc. Push-to-talk system with enhanced noise reduction
US8259954B2 (en) * 2007-10-11 2012-09-04 Cisco Technology, Inc. Enhancing comprehension of phone conversation while in a noisy environment
US8027743B1 (en) 2007-10-23 2011-09-27 Adobe Systems Incorporated Adaptive noise reduction
US20090287489A1 (en) * 2008-05-15 2009-11-19 Palm, Inc. Speech processing for plurality of users
WO2009143898A1 (en) 2008-05-30 2009-12-03 Phonak Ag Method for adapting sound in a hearing aid device by frequency modification and such a device
DE102008052176B4 (en) 2008-10-17 2013-11-14 Siemens Medical Instruments Pte. Ltd. Method and hearing aid for parameter adaptation by determining a speech intelligibility threshold
US8582790B2 (en) 2010-02-12 2013-11-12 Audiotoniq, Inc. Hearing aid and computing device for providing audio labels
EP2375781B1 (en) 2010-04-07 2013-03-13 Oticon A/S Method for controlling a binaural hearing aid system and binaural hearing aid system
US8639516B2 (en) * 2010-06-04 2014-01-28 Apple Inc. User-specific noise suppression for voice quality improvements
US8700396B1 (en) * 2012-09-11 2014-04-15 Google Inc. Generating speech data collection prompts

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4296277A (en) * 1978-09-26 1981-10-20 Feller Ag Electronic voice detector
US4538295A (en) * 1982-08-16 1985-08-27 Nissan Motor Company, Limited Speech recognition system for an automotive vehicle
US6820053B1 (en) * 1999-10-06 2004-11-16 Dietmar Ruwisch Method and apparatus for suppressing audible noise in speech transmission
US20030028385A1 (en) * 2001-06-30 2003-02-06 Athena Christodoulou Audio reproduction and personal audio profile gathering apparatus and method
US6944474B2 (en) * 2001-09-20 2005-09-13 Sound Id Sound enhancement for mobile phones and other products producing personalized audio for users
US20060282264A1 (en) * 2005-06-09 2006-12-14 Bellsouth Intellectual Property Corporation Methods and systems for providing noise filtering using speech recognition
US20070225984A1 (en) * 2006-03-23 2007-09-27 Microsoft Corporation Digital voice profiles
US8010366B1 (en) * 2007-03-20 2011-08-30 Neurotone, Inc. Personal hearing suite
US20090086933A1 (en) * 2007-10-01 2009-04-02 Labhesh Patel Call routing using voice signature and hearing characteristics
US20090185704A1 (en) * 2008-01-21 2009-07-23 Bernafon Ag Hearing aid adapted to a specific type of voice in an acoustical environment, a method and use
US20110178799A1 (en) * 2008-07-25 2011-07-21 The Board Of Trustees Of The University Of Illinois Methods and systems for identifying speech sounds using multi-dimensional analysis
US20100094619A1 (en) * 2008-10-15 2010-04-15 Verizon Business Network Services Inc. Audio frequency remapping
US8369549B2 (en) * 2010-03-23 2013-02-05 Audiotoniq, Inc. Hearing aid system adapted to selectively amplify audio signals
US20110282669A1 (en) * 2010-05-17 2011-11-17 Avaya Inc. Estimating a Listener's Ability To Understand a Speaker, Based on Comparisons of Their Styles of Speech
US9468401B2 (en) * 2010-08-05 2016-10-18 Ace Communications Limited Method and system for self-managed sound enhancement
US20120215532A1 (en) * 2011-02-22 2012-08-23 Apple Inc. Hearing assistance system for providing consistent human speech
US20130339025A1 (en) * 2011-05-03 2013-12-19 Suhami Associates Ltd. Social network with enhanced audio communications for the Hearing impaired
US20150269953A1 (en) * 2012-10-16 2015-09-24 Audiologicall, Ltd. Audio signal manipulation for speech enhancement before sound reproduction

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150356980A1 (en) * 2013-01-15 2015-12-10 Sony Corporation Storage control device, playback control device, and recording medium
US10607625B2 (en) * 2013-01-15 2020-03-31 Sony Corporation Estimating a voice signal heard by a user
US20150078575A1 (en) * 2013-02-11 2015-03-19 Symphonic Audio Technologies Corp. Audio apparatus and methods
US9319019B2 (en) 2013-02-11 2016-04-19 Symphonic Audio Technologies Corp. Method for augmenting a listening experience
US9344815B2 (en) 2013-02-11 2016-05-17 Symphonic Audio Technologies Corp. Method for augmenting hearing
US9344793B2 (en) * 2013-02-11 2016-05-17 Symphonic Audio Technologies Corp. Audio apparatus and methods
US9787824B2 (en) * 2013-07-05 2017-10-10 Unlimited Mfa Co. Ltd. Method of processing telephone signals and electronic device thereof
US20150011192A1 (en) * 2013-07-05 2015-01-08 Kuo-Ping Yang Method of processing telephone signals and electronic device thereof
US20190278445A1 (en) * 2014-01-28 2019-09-12 International Business Machines Corporation Impairment-adaptive electronic data interaction system
US11429255B2 (en) * 2014-01-28 2022-08-30 International Business Machines Corporation Impairment-adaptive electronic data interaction system
US20160239253A1 (en) * 2014-01-31 2016-08-18 Matteo Staffaroni Method for audio correction in electronic devices
US10121488B1 (en) * 2015-02-23 2018-11-06 Sprint Communications Company L.P. Optimizing call quality using vocal frequency fingerprints to filter voice calls
US10825462B1 (en) 2015-02-23 2020-11-03 Sprint Communications Company L.P. Optimizing call quality using vocal frequency fingerprints to filter voice calls
US9959083B2 (en) * 2015-11-10 2018-05-01 Optim Corporation System and method for sharing screen
US20170131961A1 (en) * 2015-11-10 2017-05-11 Optim Corporation System and method for sharing screen
US10475484B2 (en) * 2016-12-09 2019-11-12 Baidu Online Network Technology (Beijing) Co., Ltd. Method and device for processing speech based on artificial intelligence
US20180166103A1 (en) * 2016-12-09 2018-06-14 Baidu Online Network Technology (Beijing) Co., Ltd. Method and device for processing speech based on artificial intelligence
US10582044B2 (en) 2017-01-25 2020-03-03 Sorenson Ip Holdings, Llc Selecting audio profiles
US10284714B2 (en) 2017-01-25 2019-05-07 Sorenson Ip Holdings, Llc Selecting audio profiles
US9973627B1 (en) * 2017-01-25 2018-05-15 Sorenson Ip Holdings, Llc Selecting audio profiles
US20180271710A1 (en) * 2017-03-22 2018-09-27 Bragi GmbH Wireless earpiece for tinnitus therapy
US10521512B2 (en) * 2017-05-26 2019-12-31 Bose Corporation Dynamic text-to-speech response from a smart speaker
US20180341643A1 (en) * 2017-05-26 2018-11-29 Bose Corporation Dynamic text-to-speech response from a smart speaker
US10564928B2 (en) * 2017-06-02 2020-02-18 Rovi Guides, Inc. Systems and methods for generating a volume- based response for multiple voice-operated user devices
US11481187B2 (en) 2017-06-02 2022-10-25 Rovi Guides, Inc. Systems and methods for generating a volume-based response for multiple voice-operated user devices
US11475907B2 (en) * 2017-11-27 2022-10-18 Goertek Technology Co., Ltd. Method and device of denoising voice signal
US11790920B2 (en) * 2018-12-12 2023-10-17 Sonos, Inc. Guest access for voice control of playback devices
US20220351733A1 (en) * 2018-12-12 2022-11-03 Sonos, Inc. Guest Access for Voice Control of Playback Devices
US11205440B2 (en) * 2018-12-28 2021-12-21 Pixart Imaging Inc. Sound playback system and output sound adjusting method thereof
US10971168B2 (en) * 2019-02-21 2021-04-06 International Business Machines Corporation Dynamic communication session filtering
US20200273477A1 (en) * 2019-02-21 2020-08-27 International Business Machines Corporation Dynamic communication session filtering
US20210152880A1 (en) * 2019-11-14 2021-05-20 Dish Network L.L.C. Method and system for adaptive audio modification
US11818426B2 (en) * 2019-11-14 2023-11-14 Dish Network L.L.C. Method and system for adaptive audio modification

Also Published As

Publication number Publication date
CN104160443A (en) 2014-11-19
US10325612B2 (en) 2019-06-18
TW201430834A (en) 2014-08-01
US10803880B2 (en) 2020-10-13
EP2786376A1 (en) 2014-10-08
KR20140121447A (en) 2014-10-15
WO2014081408A1 (en) 2014-05-30
CN104160443B (en) 2016-11-16
KR101626438B1 (en) 2016-06-01
TWI527024B (en) 2016-03-21
US20190259406A1 (en) 2019-08-22
US20170330578A1 (en) 2017-11-16
US20200411025A1 (en) 2020-12-31

Similar Documents

Publication Publication Date Title
US10803880B2 (en) Method, device, and system for audio data processing
US10553235B2 (en) Transparent near-end user control over far-end speech enhancement processing
JP6849797B2 (en) Listening test and modulation of acoustic signals
US8918197B2 (en) Audio communication networks
US9183845B1 (en) Adjusting audio signals based on a specific frequency range associated with environmental noise characteristics
US10142483B2 (en) Technologies for dynamic audio communication adjustment
US20150149169A1 (en) Method and apparatus for providing mobile multimodal speech hearing aid
US8965005B1 (en) Transmission of noise compensation information between devices
US9774743B2 (en) Silence signatures of audio signals
WO2002095726A1 (en) Speech quality indication
CN104580764A (en) Ultrasound pairing signal control in teleconferencing system
TWI624183B (en) Method of processing telephone voice and computer program thereof
US9031836B2 (en) Method and apparatus for automatic communications system intelligibility testing and optimization
US10841713B2 (en) Integration of audiogram data into a device
US9161136B2 (en) Telecommunications methods and systems providing user specific audio optimization
US10483933B2 (en) Amplification adjustment in communication devices
US9661417B2 (en) System, method, and computer program product for voice decibel monitoring on electronic computing devices
US20150201057A1 (en) Method of processing telephone voice output and earphone
Kozma-Spytek et al. Factors Affecting the Accessibility of Voice Telephony for People with Hearing Loss: Audio Encoding, Network Impairments, Video and Environmental Noise
WO2019100371A1 (en) Method and apparatus for adjusting call volume
JP2022050407A (en) Telecommunication device, telecommunication system, method for operating telecommunication device, and computer program
TWI566233B (en) Wireless communication method which provides clear voice content

Legal Events

Date Code Title Description
AS Assignment

Owner name: UNIFY GMBH & CO. KG, GERMANY

Free format text: CHANGE OF NAME;ASSIGNOR:SIEMENS ENTERPRISE COMMUNICATIONS GMBH & CO. KG;REEL/FRAME:043392/0256

Effective date: 20131024

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: RINGCENTRAL IP HOLDINGS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:UNIFY SOFTWARE AND SOLUTIONS GMBH & CO. KG;UNIFY PATENTE GMBH & CO. KG;REEL/FRAME:053366/0404

Effective date: 20191230

Owner name: RINGCENTRAL, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RINGCENTRAL IP HOLDINGS, INC.;REEL/FRAME:053375/0383

Effective date: 20200702

AS Assignment

Owner name: UNIFY PATENTE GMBH & CO. KG, GERMANY

Free format text: CONFIDENTIAL PATENT AGREEMENT;ASSIGNOR:UNIFY GMBH & CO. KG;REEL/FRAME:054822/0001

Effective date: 20140930

AS Assignment

Owner name: RINGCENTRAL, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RINGCENTRAL IP HOLDINGS, INC.;REEL/FRAME:058851/0969

Effective date: 20220106