US9711135B2 - Electronic devices and methods for compensating for environmental noise in text-to-speech applications - Google Patents
Electronic devices and methods for compensating for environmental noise in text-to-speech applications Download PDFInfo
- Publication number
- US9711135B2 US9711135B2 US14/374,170 US201314374170A US9711135B2 US 9711135 B2 US9711135 B2 US 9711135B2 US 201314374170 A US201314374170 A US 201314374170A US 9711135 B2 US9711135 B2 US 9711135B2
- Authority
- US
- United States
- Prior art keywords
- environmental noise
- speech output
- location
- sound characteristics
- output signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 230000007613 environmental effect Effects 0.000 title claims abstract description 259
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000001228 spectrum Methods 0.000 claims description 16
- 238000004091 panning Methods 0.000 claims description 4
- 230000006978 adaptation Effects 0.000 description 11
- 238000004891 communication Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 230000001413 cellular effect Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000002411 adverse Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000004424 eye movement Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 238000009423 ventilation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/08—Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/033—Voice editing, e.g. manipulating the voice of the synthesiser
- G10L13/0335—Pitch control
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/04—Details of speech synthesis systems, e.g. synthesiser structure or memory management
- G10L13/047—Architecture of speech synthesisers
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/003—Changing voice quality, e.g. pitch or formants
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L21/0216—Noise filtering characterised by the method used for estimating noise
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
- G10L2015/0631—Creating reference templates; Clustering
Definitions
- TTS Text-to-speech capabilities are being increasingly incorporated into a wide variety of electronic devices.
- TTS is used to read articles, emails, text messages and the like and to communicate with intelligent personal assistant applications on electronic devices such as computers (e.g., portable computers) and telephones.
- TTS is also used for navigation with such devices as well as dedicated global positioning system (GPS) devices.
- GPS global positioning system
- the environment in which TTS is used may be noisy due to any number of factors such background conversation, music, passing trains and so forth.
- One solution is to detect the ambient or environmental noise and increase the volume of the TTS output in response to increased environmental noise.
- volume control which may not be sufficient for certain environmental noise. Prolonged periods with high volume can also be annoying for the user.
- embodiments of the invention are directed to a method by an electronic device for compensating for environmental noise in text-to-speech (TTS) speech output.
- the method includes: measuring environmental noise using a microphone signal; determining sound characteristics of the measured environmental noise; dynamically predicting expected future sound characteristics of the environmental noise based on the determined sound characteristics of the measured environmental noise; receiving a text input at a TTS engine at the device, with the TTS engine configured to convert the text input into a speech output signal; determining text characteristics of the text input at the TTS engine; and, at the TTS engine, dynamically adapting the speech output signal based on the determined text characteristics of the text input and the predicted expected future sound characteristics of the environmental noise.
- the speech output signal is dynamically adapted by varying the pace of the speech output and/or varying the pitch of the speech output.
- Dynamically predicting expected future sound characteristics of the environmental noise may include characterizing a time-varying pattern of the expected future sound characteristics of the environmental noise based on a time-varying pattern observed in determined sound characteristics of previously occurring environmental noise.
- characterizing a time-varying pattern of the expected future sound characteristics of the environmental noise includes characterizing expected differences over time in the volume of the environmental noise based on differences over time in volume of previously occurring environmental noise.
- characterizing a time-varying pattern of the expected future sound characteristics of the environmental noise includes predicting timing of gaps in the environmental noise where the environmental noise volume has less than a threshold amplitude based on differences over time in volume of previously occurring environmental noise.
- the speech output signal may be dynamically adapted by increasing the pace of the speech output such that the speech output is fully carried out within one of the predicted gaps.
- the speech output signal may be dynamically adapted by interrupting the speech output at an interrupted location of the speech output signal at the end of one of the predicted gaps and resuming the speech output from the interrupted location of the speech output signal at a subsequent predicted gap.
- characterizing a time-varying pattern of the expected future sound characteristics of the environmental noise includes characterizing expected differences over time in the frequency spectrum of the environmental noise based on differences over time in the frequency spectrum of the previously occurring environmental noise.
- the speech output signal may be dynamically adapted by varying the pitch of the speech output to compensate for the expected differences over time in the frequency spectrum of the environmental noise.
- dynamically predicting expected future sound characteristics of the environmental noise based on the determined sound characteristics of the measured environmental noise includes generating values of a time-varying control curve representing an expected pattern of the environmental noise.
- the speech output signal may be dynamically adapted by dynamically adapting the speech output signal using values of the time-varying control curve.
- the method includes operating the device in a learning mode to determine sound characteristics of the measured environmental noise.
- the method may include determining a location of the device, and dynamically predicting expected future sound characteristics of the environmental noise may be based on previously known sound characteristics of environmental noise at the location.
- characterizing a time-varying pattern of the expected future sound characteristics of the environmental noise comprises characterizing expected differences over time in the direction of the environmental noise based on differences over time in direction of previously occurring environmental noise.
- Dynamically adapting the speech output signal may include panning the speech output to compensate for the expected differences over time in the direction of the environmental noise.
- embodiments of the invention are directed to a method by an electronic device for compensating for environmental noise in text-to-speech (TTS) speech output.
- the method includes: measuring environmental noise using a microphone signal; determining identifying characteristics of the measured environmental noise; and determining whether the identifying characteristics of the measured environmental noise match those corresponding to an identified environmental noise source that is stored in an environmental noise library at the device.
- the method includes: receiving a text input at a TTS engine at the device, with the TTS engine configured to convert the text input into a speech output signal; and dynamically adapting the speech output signal based on a time-varying pattern of sound characteristics of the identified environmental noise source, with the time-varying pattern being stored in the environmental noise library.
- the speech output signal is dynamically adapted by varying the pace of the speech output and/or varying the pitch of the speech output.
- the method includes operating the device in a learning mode including: learning identifying characteristics of the measured environmental noise; learning a time-varying pattern of sound characteristics of the environmental noise; and storing the identifying characteristics and the time-varying pattern of sound characteristics as an identified environmental noise source in the environmental noise library.
- the identified environmental noise source may be audio from a media file.
- the identified environmental noise source may be one or more persons in a conversation.
- the identified environmental noise source may be a conversation type.
- embodiments of the invention are directed to an electronic device configured to compensate for environmental noise in TTS speech output.
- the device includes: microphone configured to measure environmental noise; an environmental noise analysis circuit; and a TTS processor.
- the environmental noise analysis circuit is configured to: determine sound characteristics of environmental noise measured by the microphone; and predict future expected sound characteristics of the environmental noise based on the determined sound characteristics of the measured environmental noise.
- the TTS processor is configured to: receive a text input and covert the text input into a speech output signal; determine text characteristics of the text input; and adapt the speech output signal based on the determined text characteristics of the text input and the predicted expected sound characteristics of the environmental noise.
- the TTS processor is configured to adapt the speech output signal by varying the pace of speech output and/or varying the pitch of the speech output.
- the environmental noise analysis circuit is configured to characterize a time-varying pattern of the expected future sound characteristics of the environmental noise based on a time-varying pattern observed in determined sound characteristics of previously occurring environmental noise to predict future expected sound characteristics of the environmental noise. In some embodiments, the environmental noise analysis circuit is configured to predict timing of gaps in the environmental noise where the environmental noise volume has less than a threshold amplitude based on differences over time in volume of previously occurring environmental noise.
- the TTS processor may be configured to adapt the speech output signal by increasing the pace of the speech output such that the speech output is fully carried out within one of the predicted gaps.
- the TTS processor may be configured to adapt the speech output signal by interrupting the speech output at an interrupted location of the speech output signal at the end of one of the predicted gaps and resuming the speech output from the interrupted location of the speech output signal at a subsequent predicted gap.
- the environmental noise analysis circuit may be configured to characterize expected differences over time in the frequency spectrum of the environmental noise based on differences over time in the frequency spectrum of the previously occurring environmental noise and the TTS processor may be configured to vary the pitch of the speech output to compensate for the expected differences over time in the frequency spectrum of the environmental noise.
- embodiments of the invention are directed to a method by an electronic device for compensating for environmental noise in text-to-speech (TTS) speech output.
- the method includes: determining one or more identifying location characteristics associated with a location of the device; and determining whether the identifying location characteristics match those corresponding to an identified location that is stored in a location library at the device.
- the method includes: receiving a text input at a TTS engine at the device, the TTS engine configured to convert the text input into a speech output signal; and dynamically adapting the speech output signal based on a time-varying pattern of sound characteristics of environmental noise at the identified location, the time-varying pattern being stored in the location library.
- the method includes operating the device in a learning mode including: learning one or more identifying location characteristics associated with the location of the device; measuring environmental noise using a microphone signal; learning a time-varying pattern of sound characteristics of the environmental noise; and storing the one or more identifying location characteristics and time-varying pattern of sound characteristics as an identified location in the location library.
- the identifying location characteristics are associated with a GPS location or signal.
- the identifying location characteristics are associated with a WiFi network or an identification thereof.
- any one or more aspects or features described with respect to one embodiment may be incorporated in a different embodiment although not specifically described relative thereto. That is, all embodiments and/or features of any embodiment can be combined in any way and/or combination. Applicant reserves the right to change any originally filed claim or file any new claim accordingly, including the right to be able to amend any originally filed claim to depend from and/or incorporate any feature of any other claim although not originally claimed in that manner.
- FIG. 1 is a block diagram of an electronic device according to some embodiments of the present invention.
- FIG. 2 is a chart illustrating an exemplary time-varying pattern of predicted sound characteristics of environmental noise using the device of FIG. 1 .
- FIG. 3 is a block diagram of a communication terminal and exemplary components thereof according to some embodiments of the present invention.
- FIG. 4 is a flowchart of methods and operations that may be carried out using by the device of FIG. 1 for compensating for environmental noise in text-to-speech (TTS) speech output.
- TTS text-to-speech
- FIG. 5 is a flowchart of further methods and operations that may be carried out using by the device of FIG. 1 for compensating for environmental noise in text-to-speech (TTS) speech output.
- TTS text-to-speech
- FIG. 6 is a flowchart of further methods and operations that may be carried out using by the device of FIG. 1 for compensating for environmental noise in text-to-speech (TTS) speech output.
- TTS text-to-speech
- Some embodiments may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.). Consequently, as used herein, the term “signal” may take the form of a continuous waveform and/or discrete value(s), such as digital value(s) in a memory or register. Furthermore, various embodiments may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system.
- circuit and “controller” and “processor” may take the form of digital circuitry, such as computer-readable program code executed by an instruction processing device(s) (e.g., general purpose microprocessor and/or digital signal microprocessor), and/or analog circuitry.
- instruction processing device e.g., general purpose microprocessor and/or digital signal microprocessor
- analog circuitry e.g., analog circuitry.
- embodiments of the invention are directed to devices and methods for compensating for ambient, background or environmental noise in text-to-speech (TTS) applications.
- TTS text-to-speech
- TTS is a type of speech synthesis wherein textual input is converted into speech output.
- Different environmental noise sources can create different challenges with regard to the intelligibility of the TTS speech output.
- the electronic device 100 includes a text-to-speech (TTS) engine or processor 102 that is configured to receive an input 104 and convert the input 104 to a speech output signal 106 .
- the speech output signal 106 is received at one or more speakers or transducers 108 which are configured to output a speech output 110 (e.g., synthesized speech).
- the TTS engine 102 includes a text analysis module 112 and a TTS adaptation circuit or processor 114 ; these components will be described in greater detail below.
- the input 104 is a text input 104 (e.g., a text input signal) such that the speech output signal 106 and the speech output 110 is based on text in the text input 104 .
- the input 104 is generally referenced as a text input 104 below, it will be appreciated that the input 104 can take other forms (e.g., the input may be a TTS input 104 ).
- the speech output signal 106 and the speech output 110 may be used as a substitute for a user's voice.
- the speech output signal 106 and the speech output 110 may be translated speech and the input 104 may be pre-translated speech (e.g., speech input by the user in a different language).
- the input 104 may also be input initiated by, for example, a person using assistive input technology.
- the input 104 may be signals from electrodes or other sensors in a subvocal recognition system.
- the input 104 may be signals that are generated from eye movement, mouth/tongue movement and the like.
- the input 104 may be signals from a brain computer or a brain-computer interface (e.g., neural signals).
- a user of the device 100 listens to the speech output 110 in TTS applications.
- environmental noise e.g., ambient or background noise
- the device 100 may include at least one environmental noise sensor 116 such as a microphone 116 that measures or senses the environmental noise.
- the microphone 116 provides a microphone signal 118 to an environmental noise analysis circuit or processor 120 .
- the environmental noise analysis processor 120 may include an audio analysis module 122 that is configured to perform a real time audio analysis of environmental noise measured by the microphone 116 .
- the audio analysis module 122 determines sound characteristics including the amplitude (volume) and frequency levels associated with the captured environmental noise.
- the microphone 116 may be located in another device 101 .
- the microphone 116 could be in a car, a headset or another electronic device.
- the device 101 may communicate (e.g., wirelessly) with the device 100 and provide the microphone signal 118 to the device 100 .
- An environmental noise prediction module 124 is configured to predict expected future sound characteristics of the environmental noise based on the sound characteristics determined by the audio analysis module 122 .
- the environmental noise prediction module 124 characterizes a time-varying pattern of the expected future sound characteristics of the environmental noise based on a time-varying pattern observed in determined sound characteristics of previously occurring environmental noise.
- the environmental noise prediction module 124 generates values of a time-varying control curve representing an expected pattern of the environmental noise. The control curve may be constantly or dynamically reconfigured based on the measured environmental noise.
- FIG. 2 illustrates an exemplary control curve representing an expected pattern of environmental noise.
- the primary source of the environmental noise in FIG. 2 may be a passing train, for example.
- the curve shown in FIG. 2 may represent predicted sound characteristics by the environmental noise prediction module 124 after analyzing a relatively short sample of measured environmental noise (e.g., up to point “A” in time).
- the environmental noise prediction module 124 may recognize the gradually increasing amplitude or volume of the noise as well as the gradually increasing frequency as the train approaches.
- the environmental noise analysis processor 120 may also recognize identifying characteristics of the environmental noise source (e.g., a train) to assist in generating the predicted noise pattern.
- the environmental noise prediction module 124 may characterize expected differences over time in the amplitude (volume) of the environmental noise based on differences over time in volume of previously occurring environmental noise.
- the environmental noise prediction module 124 may predict timing of gaps in the environmental noise where the environmental noise has less than a threshold amplitude or volume based on differences over time in volume of previously occurring environmental noise.
- the environmental noise prediction module 124 may also characterize expected differences over time in the frequency spectrum of the environmental noise based on differences over time in the frequency spectrum of the previously occurring environmental noise.
- the text analysis module 112 is configured to analyze the text input 104 and determine characteristics thereof. For example, the text analysis module 112 may determine how many words are in the text input, the minimum amount of time needed to output the text input as an intelligible speech output, whether any words need emphasis and/or extra output time (e.g., an obscure word) and so forth.
- the TTS adaptation processor 114 is configured to dynamically adapt the speech output signal 106 based on the determined characteristics of the text input and the predicted expected future sound characteristics of the environmental noise.
- the TTS adaptation processor 114 has a pace adjustment module 126 configured to vary the pace of the speech output 110 .
- the TTS adaptation processor 114 has a pitch adjustment module 128 configured to vary the pitch of the speech output 110 .
- the pitch adjustment module 128 may include an electronic filter.
- the TTS adaptation processor 114 may also have a volume adjustment module 130 configured to vary the amplitude or volume of the speech output 110 (or portions thereof for emphasis).
- the volume adjustment module 130 may include an amplifier.
- the pace adjustment module 126 may increase the pace of the speech output signal 106 such that the speech output 110 is carried out within one or more of the predicted gaps where the environmental noise has less than a threshold amplitude.
- the pace adjustment module 126 may be configured to adapt the speech output signal 106 by increasing the pace of the speech output 110 such that the speech output 110 is fully carried out within one of the predicted gaps.
- the pace adjustment module 126 may be configured to adapt the speech output signal 106 by interrupting the speech output 110 at an interrupted location of the speech output signal 106 (or the text corresponding thereto) at the end of one of the predicted gaps and resuming the speech output 110 from the interrupted location of the speech output signal 106 (or the text corresponding thereto) at a subsequent predicted gap.
- the pace adjustment module 126 may increase the pace of the speech output signal 106 such that the speech output 110 corresponding to the text input 104 is fully carried out in the first gap, e.g., as the train is approaching but before it creates sufficient background noise to make the speech output 110 unintelligible.
- the pace adjustment module 126 may also interrupt the speech output 110 as the predicted sound amplitude reaches the threshold amplitude.
- the pace adjustment module 126 may then resume the speech output 110 at the beginning of the second gap, e.g., after the train has passed and is a sufficient distance from the device to allow intelligible speech output.
- the pitch adjustment module 128 may vary the pitch of the speech output 110 to compensate for the predicted differences over time in the frequency spectrum of the environmental noise.
- the environmental noise analysis processor 120 can provide the determined sound characteristics of the measured sound to the TTS adaptation processor 114 .
- the environmental noise analysis processor 120 provides the determined frequencies of the measured environmental noise to the TTS adaptation processor 114 or the pitch adjustment module 128 such that the pitch of the speech output 110 can be varied in real time to compensate for changes in frequencies of the environmental noise and to “cut through” the environmental noise.
- the at least one environmental noise sensor 116 may detect the direction or location of the environmental noise or sound.
- the at least one environmental noise sensor 116 may include a plurality of microphones and/or a gyroscope and/or other suitable device(s) to detect the location of the noise (e.g., the direction from which the noise is emanating).
- the at least one speaker 106 may comprise a multi-directional speaker or a plurality of speakers that are configured for panning the speech output 110 . Therefore, in some embodiments, the TTS adaptation processor 114 may be configured to pan the speech output 110 and/or control 3 D audio aspects of the speech output 110 in response to directional sound characteristics of the measured environmental noise.
- the TTS adaptation processor 114 may be configured to characterize (predict) a time-varying pattern of the expected future sound characteristics of the environmental noise including characterizing expected differences over time in the direction or location of the environmental noise based on differences over time in direction or location of previously occurring environmental noise.
- the TTS adaptation processor 114 may dynamically adapt the speech output signal 106 by controlling 3 D audio output and/or panning the audio output that makes up the speech output 110 (e.g., dynamically adapt the direction of the speech output 110 ).
- the TTS adaptation processor 114 may adapt the direction of the speech output 110 in addition to adapting the pace, pitch and/or volume of the speech output 110 .
- devices and methods according to the invention may further change the intonation, dialect, voice, voice gender, choice of words, lengths of sentences, and so forth to further improve the intelligibility of the speech output 110 .
- the environmental noise analysis processor 120 may recognize identifying sound characteristics of the environmental noise.
- the environmental noise analysis processor 120 may include an environmental noise recognition module 132 that interacts with an environmental noise library 134 .
- the environmental noise recognition module 132 may determine identifying characteristics of the measured environmental noise and query the library 134 to determine whether the identifying characteristics match those corresponding to a pre-identified environmental noise source or type that is stored in the library 134 .
- the environmental noise library 134 may include a time-varying pattern of sound characteristics for each environmental noise source stored therein.
- Exemplary environmental noise sources include an airplane cabin, a train station (or a passing train), a song, audio from a movie or television show, participant(s) in a conversation and types of conversations.
- a song may have gaps in the audio amplitude (e.g., where the amplitude is less than a threshold) and the TTS speech output can be timed accordingly.
- the environmental noise in an airplane cabin may have a relatively constant amplitude (e.g., from the engine) but with an oscillating frequency and the pitch of the TTS speech output can be varied accordingly.
- a participant in a conversation may speak with a known voice cadence and the pace and/or pitch of the TTS speech output can be varied based on the known voice cadence.
- the type of conversation may be determined based on identifying sound characteristics (e.g., the name of a politician or a sports team). It may be known that conversations involving politics or sports may be heated and/or fast-paced, and the TTS speech output may be paused until the conversation has ceased, for example.
- the device 100 may be operated in a learning mode.
- the device 100 operates in a learning mode for a certain amount of time to “learn” characteristics of the environmental noise such that a time-varying pattern of sound characteristics can be established for a particular environment.
- the device 100 may operate in a learning mode for a number of seconds or minutes to learn time-varying patterns in the environmental noise, such as regular or irregular intervals of relatively high amplitude sound and/or timing of gaps as well as differences over time in the frequency spectrum.
- the device 100 may then be operated in a TTS mode and the time-varying pattern of sound characteristics can be used to adapt the speech output by adapting the pace, pitch and/or volume of the speech output to provide intelligible speech output in the learned environment.
- the device 100 may enter the learning mode automatically (e.g., upon receiving a TTS text input but prior to converting the input to a speech output).
- a user may manually place the device 100 in the learning mode, e.g., using a user input interface.
- the device 100 may automatically enter the learning mode when identifying characteristics of the environmental noise are not recognized or are not present in the environmental noise library 134 .
- the environmental noise recognition module 132 may make a determination as to whether identifying characteristics of the measured environmental noise match those of an identified environmental noise source that is stored in the environmental noise library 134 . If the identifying characteristics of the measured environmental noise do not match those of an identified environmental noise source that is stored in the environmental noise library 134 , the device 100 may switch to operate in a learning mode to learn identifying characteristics of the environmental noise that is being measured and learn a time-varying pattern of sound characteristics of the environmental noise. The identifying characteristics and the time-varying pattern of sound characteristics may then be stored in the environmental noise library 134 for later use.
- the environmental noise analysis processor 120 includes a location recognition module 136 and a location library 138 .
- the location recognition module 136 may determine one or more identifying location characteristics associated with the location of the device 100 (e.g., GPS coordinates) and query the location library 138 to determine whether the identifying location characteristics match those corresponding to an identified location that is stored in the location library 138 .
- the location library 138 may include a time-varying pattern of sound characteristics associated with the environmental noise at the identified location. For example, the location may be in or near a factory or office building and the time-varying pattern may include sound characteristics associated with environmental noise from sources such as machines, ventilation systems and the like at that location.
- the one or more identifying location characteristics associated with the location of the device 100 are associated with GPS coordinates or location (or an alternative or competing system such as GLONASS, Galileo, WiFi Positioning and the like).
- the one or more identifying location characteristics associated with the location of the device 100 are associated with a WiFi network, such as an WiFi access point (e.g., router), a WiFi hot spot a WLAN network and the like.
- the one or more identifying location characteristics associated with the location of the device 100 may be associated with a Bluetooth beacon.
- the one or more identifying location characteristics associated with the location of the device 100 may be associated with a NFC signal (e.g., a known NFC signal that is associated with a location). It will be appreciated that other network or location technologies can be used to provide the identifying location characteristics.
- a text input may be received at the TTS engine 102 at the device 100 .
- the speech output signal may be adapted based on the time-varying pattern of sound characteristics of environmental noise at the identified location that is stored in the location library 138 .
- the device 100 may be operated in a learning mode if the one or more identifying location characteristics associated with the location of the device 100 do not match those corresponding to an identified location that is stored in the location library 138 at the device 100 .
- the device 100 may switch (e.g., automatically switch) to operate in a learning mode to learn one or more identifying location characteristics associated with the location of the device 100 , and may learn a time-varying pattern of sound characteristics at the location.
- the identifying location characteristics and the time-varying pattern of sound characteristics may then be stored in the location library 138 as an identified location for later use.
- FIG. 3 is a diagram of a terminal 200 that includes a terminal housing 200 h .
- the illustrated terminal 200 includes a display 232 and a user input interface 234 (e.g., a keypad or touchscreen).
- the illustrated terminal further includes a general application controller 202 , a wireless communication protocol controller 204 , a cellular transceiver 206 , a WLAN transceiver 208 (e.g., compliant with one or more of the IEEE 801.11a-g standards), and/or a Bluetooth transceiver 210 .
- the cellular transceiver 206 can be configured to communicate bi-directionally according to one or more cellular standards, such as Long Term Evolution (LTE), enhanced data rates for General Packet Radio Service (GPRS) evolution (EDGE), code division multiple access (CDMA), wideband-CDMA, CDMA2000, and/or Universal Mobile Telecommunications System (UMTS) frequency bands.
- LTE Long Term Evolution
- EDGE General Packet Radio Service
- CDMA Code division multiple access
- CDMA2000 Wideband-CDMA2000
- UMTS Universal Mobile Telecommunications System
- the terminal 200 may include the components described above in connection with the device 100 , including the microphone 116 , the environmental noise analysis circuit 120 , the TTS engine 102 and/or the loudspeaker 108 .
- One or more of the illustrated components may be omitted from the terminal 200 and/or rearranged within the terminal 200 , and it is contemplated that additional components (e.g., one or more additional controllers or processors) may be included.
- a method 400 by an electronic device for compensating for environmental noise in TTS speech output includes measuring environmental noise using a microphone signal (e.g., at the device) (Block 402 ). Sound characteristics of the measured environmental noise are determined (Block 404 ). Expected future sound characteristics of the environmental noise are predicted based on the determined sound characteristics of the measured environmental noise (Block 406 ).
- a text input is received at a TTS engine at the device, and the TTS engine determines text characteristics of the text input (Block 408 ).
- the TTS engine is configured to convert the text input into a speech output signal.
- the speech output signal is adapted at the TTS engine based on the determined text characteristics of the text input and the predicted expected future sound characteristics of the environmental noise (Block 410 ).
- the speech output signal may be adapted by varying the pace of the speech output and/or varying the pitch of the speech output.
- a method 500 by an electronic device for compensating for environmental noise in TTS speech output includes measuring environmental noise using a microphone signal (e.g., at the device) (Block 502 ). Identifying characteristics of the measured environmental noise are determined (Block 504 ). A determination of whether the identifying characteristics of the measured environmental noise match those corresponding to an identified environmental noise source that is stored in an environmental noise library at the device (Block 506 ).
- a text input is received at a TTS engine at the device (Block 508 ).
- the TTS engine is configured to convert the text input into a speech output signal.
- the speech output signal is adapted based on a time-varying pattern of sound characteristics of the identified environmental noise (Block 510 ).
- the time-varying pattern is stored in the environmental noise library.
- the speech output signal may be adapted by varying the pace of the speech output and/or varying the pitch of the speech output.
- the device When the identifying characteristics of the measured environmental noise do not match those corresponding to an identified environmental noise source that is stored in the environmental noise library, the device is operated in a learning mode including learning identifying characteristics of the measured environmental noise (Block 512 ) and learning a time-varying pattern of sound characteristics of the environmental noise (Block 514 ). The identifying characteristics and the time-varying pattern of sound characteristics are stored in the environmental noise library (Block 516 ).
- a method 600 by an electronic device for compensating for environmental noise in TTS speech output includes determining one or more identifying location characteristics associated with a location of the device (Block 602 ). A determination is made as to whether the identifying location characteristics match those corresponding to an identified location that is stored in a location library at the device (Block 604 ).
- a text input is received at a TTS engine at the device (Block 606 ).
- the TTS engine is configured to convert the text input into a speech output signal.
- the speech output signal is adapted based on a time-varying pattern of sound characteristics of environmental noise at the identified location (Block 608 ).
- the time-varying pattern is stored in the location library.
- the device is operated in a learning mode including learning one or more identifying location characteristics corresponding to the location of the device (Block 610 ) measuring the environmental noise (Block 612 ), and learning a time-varying pattern of sound characteristics of the environmental noise at the location of the device (Block 614 ).
- the one or more identifying location characteristics and the time-varying pattern of sound characteristics is stored as an identified location in the location library (Block 616 ).
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Telephone Function (AREA)
Abstract
Description
Claims (17)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2013/084395 WO2015092943A1 (en) | 2013-12-17 | 2013-12-17 | Electronic devices and methods for compensating for environmental noise in text-to-speech applications |
Publications (2)
Publication Number | Publication Date |
---|---|
US20160275936A1 US20160275936A1 (en) | 2016-09-22 |
US9711135B2 true US9711135B2 (en) | 2017-07-18 |
Family
ID=49918786
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/374,170 Active 2034-11-11 US9711135B2 (en) | 2013-12-17 | 2013-12-17 | Electronic devices and methods for compensating for environmental noise in text-to-speech applications |
Country Status (2)
Country | Link |
---|---|
US (1) | US9711135B2 (en) |
WO (1) | WO2015092943A1 (en) |
Families Citing this family (113)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10255566B2 (en) | 2011-06-03 | 2019-04-09 | Apple Inc. | Generating and processing task items that represent tasks to perform |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
DE112014000709B4 (en) | 2013-02-07 | 2021-12-30 | Apple Inc. | METHOD AND DEVICE FOR OPERATING A VOICE TRIGGER FOR A DIGITAL ASSISTANT |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
EP3937002A1 (en) | 2013-06-09 | 2022-01-12 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
DE112014003653B4 (en) | 2013-08-06 | 2024-04-18 | Apple Inc. | Automatically activate intelligent responses based on activities from remote devices |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
TWI566107B (en) | 2014-05-30 | 2017-01-11 | 蘋果公司 | Method for processing a multi-part voice command, non-transitory computer readable storage medium and electronic device |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10460227B2 (en) | 2015-05-15 | 2019-10-29 | Apple Inc. | Virtual assistant in a communication session |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10200824B2 (en) | 2015-05-27 | 2019-02-05 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US9640169B2 (en) * | 2015-06-25 | 2017-05-02 | Bose Corporation | Arraying speakers for a uniform driver field |
US20160378747A1 (en) | 2015-06-29 | 2016-12-29 | Apple Inc. | Virtual assistant for media playback |
US10740384B2 (en) | 2015-09-08 | 2020-08-11 | Apple Inc. | Intelligent automated assistant for media search and playback |
US10331312B2 (en) | 2015-09-08 | 2019-06-25 | Apple Inc. | Intelligent automated assistant in a media environment |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9847764B2 (en) * | 2015-09-11 | 2017-12-19 | Blackberry Limited | Generating adaptive notification |
RU2632424C2 (en) | 2015-09-29 | 2017-10-04 | Общество С Ограниченной Ответственностью "Яндекс" | Method and server for speech synthesis in text |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10074359B2 (en) * | 2016-11-01 | 2018-09-11 | Google Llc | Dynamic text-to-speech provisioning |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
JP2018159759A (en) * | 2017-03-22 | 2018-10-11 | 株式会社東芝 | Voice processor, voice processing method and program |
JP6646001B2 (en) * | 2017-03-22 | 2020-02-14 | 株式会社東芝 | Audio processing device, audio processing method and program |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK201770428A1 (en) | 2017-05-12 | 2019-02-18 | Apple Inc. | Low-latency intelligent automated assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK201770411A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Multi-modal interfaces |
US10395659B2 (en) * | 2017-05-16 | 2019-08-27 | Apple Inc. | Providing an auditory-based interface of a digital assistant |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US20180336892A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Detecting a trigger of a digital assistant |
US20180350344A1 (en) | 2017-05-30 | 2018-12-06 | Motorola Solutions, Inc | System, device, and method for an electronic digital assistant having a context driven natural language vocabulary |
CN107909993A (en) * | 2017-11-27 | 2018-04-13 | 安徽经邦软件技术有限公司 | A kind of intelligent sound report preparing system |
WO2019119593A1 (en) * | 2017-12-18 | 2019-06-27 | 华为技术有限公司 | Voice enhancement method and apparatus |
CN107895579B (en) * | 2018-01-02 | 2021-08-17 | 联想(北京)有限公司 | Voice recognition method and system |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US11076039B2 (en) | 2018-06-03 | 2021-07-27 | Apple Inc. | Accelerated task performance |
KR102544250B1 (en) * | 2018-07-03 | 2023-06-16 | 삼성전자주식회사 | Method and device for outputting sound |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
DK201970510A1 (en) | 2019-05-31 | 2021-02-11 | Apple Inc | Voice identification in digital assistant systems |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
DK180129B1 (en) | 2019-05-31 | 2020-06-02 | Apple Inc. | User activity shortcut suggestions |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11468890B2 (en) | 2019-06-01 | 2022-10-11 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11484271B2 (en) * | 2019-08-20 | 2022-11-01 | West Affum Holdings Dac | Alert presentation based on ancillary device conditions |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11425523B2 (en) * | 2020-04-10 | 2022-08-23 | Facebook Technologies, Llc | Systems and methods for audio adjustment |
US11810578B2 (en) | 2020-05-11 | 2023-11-07 | Apple Inc. | Device arbitration for digital assistant-based intercom systems |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11038934B1 (en) | 2020-05-11 | 2021-06-15 | Apple Inc. | Digital assistant hardware abstraction |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020128838A1 (en) | 2001-03-08 | 2002-09-12 | Peter Veprek | Run time synthesizer adaptation to improve intelligibility of synthesized speech |
US20030061049A1 (en) * | 2001-08-30 | 2003-03-27 | Clarity, Llc | Synthesized speech intelligibility enhancement through environment awareness |
US20120296654A1 (en) * | 2011-05-20 | 2012-11-22 | James Hendrickson | Systems and methods for dynamically improving user intelligibility of synthesized speech in a work environment |
US8571871B1 (en) * | 2012-10-02 | 2013-10-29 | Google Inc. | Methods and systems for adaptation of synthetic speech in an environment |
-
2013
- 2013-12-17 WO PCT/JP2013/084395 patent/WO2015092943A1/en active Application Filing
- 2013-12-17 US US14/374,170 patent/US9711135B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020128838A1 (en) | 2001-03-08 | 2002-09-12 | Peter Veprek | Run time synthesizer adaptation to improve intelligibility of synthesized speech |
US20030061049A1 (en) * | 2001-08-30 | 2003-03-27 | Clarity, Llc | Synthesized speech intelligibility enhancement through environment awareness |
US20120296654A1 (en) * | 2011-05-20 | 2012-11-22 | James Hendrickson | Systems and methods for dynamically improving user intelligibility of synthesized speech in a work environment |
US8571871B1 (en) * | 2012-10-02 | 2013-10-29 | Google Inc. | Methods and systems for adaptation of synthetic speech in an environment |
Non-Patent Citations (2)
Title |
---|
International Preliminary Report on Patentability for corresponding PCT application No. PCT/JP2013/084395 mailed Jun. 30, 2016, 7 pages. |
International Search Report and Written Opinion for PCT/JP2013/084395 mailed May 28, 2014. |
Also Published As
Publication number | Publication date |
---|---|
WO2015092943A1 (en) | 2015-06-25 |
US20160275936A1 (en) | 2016-09-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9711135B2 (en) | Electronic devices and methods for compensating for environmental noise in text-to-speech applications | |
US10433075B2 (en) | Low latency audio enhancement | |
JP6113302B2 (en) | Audio data transmission method and apparatus | |
EP2821992B1 (en) | Method for updating voiceprint feature model and terminal | |
US9864745B2 (en) | Universal language translator | |
US9293133B2 (en) | Improving voice communication over a network | |
EP2577658B1 (en) | User-specific noise suppression for voice quality improvements | |
US20090034750A1 (en) | System and method to evaluate an audio configuration | |
US20140278416A1 (en) | Method and Apparatus Including Parallell Processes for Voice Recognition | |
CN107799126A (en) | Sound end detecting method and device based on Supervised machine learning | |
US20090287489A1 (en) | Speech processing for plurality of users | |
CN110149805A (en) | Double-directional speech translation system, double-directional speech interpretation method and program | |
JP2015521404A (en) | Instant translation system | |
US20230206897A1 (en) | Electronic apparatus and method for controlling thereof | |
US9208798B2 (en) | Dynamic control of voice codec data rate | |
US11218666B1 (en) | Enhanced audio and video capture and presentation | |
US20160027435A1 (en) | Method for training an automatic speech recognition system | |
EP2196990A2 (en) | Voice processing apparatus and voice processing method | |
US9208781B2 (en) | Adapting speech recognition acoustic models with environmental and social cues | |
WO2019228329A1 (en) | Personal hearing device, external sound processing device, and related computer program product | |
CN103871416B (en) | Speech processing device and method of speech processing | |
US11948567B2 (en) | Electronic device and control method therefor | |
WO2014194273A2 (en) | Systems and methods for enhancing targeted audibility | |
JP6268916B2 (en) | Abnormal conversation detection apparatus, abnormal conversation detection method, and abnormal conversation detection computer program | |
JP5988077B2 (en) | Utterance section detection apparatus and computer program for detecting an utterance section |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:THORN, OLA;REEL/FRAME:033525/0742 Effective date: 20140807 |
|
AS | Assignment |
Owner name: SONY MOBILE COMMUNICATIONS INC., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY CORPORATION;REEL/FRAME:038542/0224 Effective date: 20160414 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY MOBILE COMMUNICATIONS, INC.;REEL/FRAME:048691/0134 Effective date: 20190325 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |