US9271089B2 - Voice control device and voice control method - Google Patents
Voice control device and voice control method Download PDFInfo
- Publication number
- US9271089B2 US9271089B2 US13/924,071 US201313924071A US9271089B2 US 9271089 B2 US9271089 B2 US 9271089B2 US 201313924071 A US201313924071 A US 201313924071A US 9271089 B2 US9271089 B2 US 9271089B2
- Authority
- US
- United States
- Prior art keywords
- sound
- received
- sending
- compensation
- hearing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000000034 method Methods 0.000 title claims description 12
- 238000001228 spectrum Methods 0.000 description 50
- 230000000694 effects Effects 0.000 description 10
- 238000012545 processing Methods 0.000 description 10
- 238000010586 diagram Methods 0.000 description 8
- 230000004044 response Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 5
- 230000032683 aging Effects 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000011835 investigation Methods 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 208000016354 hearing loss disease Diseases 0.000 description 1
- 238000012074 hearing test Methods 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/30—Monitoring or testing of hearing aids, e.g. functioning, settings, battery power
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/12—Audiometering
- A61B5/121—Audiometering evaluating hearing capacity
- A61B5/123—Audiometering evaluating hearing capacity subjective methods
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/7475—User input or interface means, e.g. keyboard, pointing device, joystick
- A61B5/749—Voice-controlled interfaces
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0316—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
- G10L21/0364—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/70—Adaptation of deaf aid to hearing loss, e.g. initial electronic fitting
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/04—Time compression or expansion
- G10L21/057—Time compression or expansion for improving intelligibility
- G10L2021/0575—Aids for the handicapped in speaking
Definitions
- the invention relates to a voice control device, a voice control method, a voice control program and a portable terminal device that control received sound.
- an equalizer for emphasizing a voice signal in a specific range, and characteristics of the equalizer are adjusted based on a volume operation by a user (Patent document 3).
- Patent document 1 there is a problem with Patent document 1 in that it is not easy to use because a user needs to execute a hearing test that forces a complicated procedure upon the user.
- Patent document 2 there is a problem with Patent document 2 in that there are cases where sound quality may be bad depending on a user because the received sound volume is determined only by the sending sound volume, hence a hearing characteristic of the user is not taken into account.
- a voice control device includes a hearing estimate section configured to estimate hearing of a user based on a sending/received sound ratio representing a ratio of the volume of a sending sound to the volume of a received sound; a compensation-quantity calculating section configured to calculate a compensation quantity for a received signal of the received sound responsive to the estimated hearing; and a compensation section configured to compensate the received signal based on the calculated compensation quantity.
- FIG. 1 is a schematic view illustrating a change of hearing level with aging
- FIG. 2 is a block diagram illustrating an example of functions of a voice control device according to an embodiment
- FIG. 3 is a block diagram illustrating an example of a hearing estimate section
- FIG. 4 is a schematic view illustrating an example of a relationship between sending/received sound ratio and age
- FIG. 5 is a schematic view illustrating an example of a relationship between age and minimum audible range
- FIG. 6 is a block diagram illustrating an example of a configuration of a hearing compensation section
- FIG. 7 is a schematic view illustrating an example of a relationship between noise quantity and compensation quantity
- FIG. 8 is a schematic view illustrating an example of minimum audible ranges before and after compensation
- FIG. 9 is a schematic view illustrating an example of a relationship between noise quantity and minimum audible ranges before and after compensation
- FIG. 10 is a schematic view illustrating spectrum compensation
- FIG. 11 is a flowchart illustrating an example of a voice control procedure according to the embodiment.
- FIG. 12 is a flowchart illustrating an example of a hearing compensation procedure.
- FIG. 13 is a block diagram illustrating an example of hardware of a portable terminal device according to an embodiment.
- FIG. 1 is a schematic view illustrating a change of hearing level with aging. According to an experiment illustrated in FIG. 1 , it can be understood that there is a difference in the average hearing level between age 40 to 60. The hearing level of 60's notably drops in a high range (2 kHz to 4 kHz). Also, as for hearing impairment induced by aging, an investigation result was reported in Http://triosaka.jp/group/infoele/life/sense/data/katagiri/070622.pdf.
- the Lombard effect is an effect that if it is noisy in the surroundings, or it is difficult to hear a voice of a counterpart because the counterpart is a quiet talker, one's speaking voice becomes louder.
- background noise is 50 dBSPL (simply denoted as dB hereafter)
- the speaking volume becomes 4 dB greater than in a quiet situation (37 dB).
- FIG. 1 in “Effects of noise on speech production: acoustic and perceptual analyses”, W. Van. Summers et. al., J. Acoust. Soc. Am., Vol. 84, No. 3, September 1988.
- the Lombard effect affects not only volumes of the surrounding noise and the voice of a counterpart, but also the hearing of a listener. If hearing is reduced, it becomes harder to hear a voice of a counterpart, hence the speaking voice tends to become louder. As illustrated in FIG. 1 , hearing reduction is related to age, hence there is a relationship between the Lombard effect and age.
- FIG. 2 is a block diagram illustrating an example of a configuration of the voice control device 1 according to the embodiment.
- the voice control device 1 includes frequency transform sections 101 and 102 , a hearing estimate section 103 , a noise estimate section 104 , a hearing compensation section 105 , a spectrum compensation-quantity calculating section 106 , a spectrum compensation section 107 , and an inverse frequency transform section 108 .
- the time-frequency transform section 101 applies a time-frequency transform to a received signal r(t) of a received sound to obtain a spectrum R(f) according to the following formula (1).
- the time-frequency transform is, for example, a fast Fourier transform (FFT).
- FFT fast Fourier transform
- the frequency transform section 101 outputs the obtained spectrum R(f) to the hearing estimate section 103 , the spectrum compensation-quantity calculating section 106 , and the spectrum compensation section 107 .
- the time-frequency transform section 102 applies a time-frequency transform to a sending signal s(t) of a sending sound to obtain a spectrum S(f) according to the following formula (2).
- the time-frequency transform is, for example, a fast Fourier transform (FFT).
- FFT fast Fourier transform
- the frequency transform section 102 outputs the obtained spectrum S(f) to the hearing estimate section 103 and the noise estimate section 104 .
- the hearing estimate section 103 estimates a user's hearing based on the received sound volume and the sending sound volume.
- FIG. 3 is a block diagram illustrating an example of the hearing estimate section 103 .
- the hearing estimate section 103 includes a sending/received-sound-ratio calculating section 131 , an age estimate section 132 , and a minimum audible range estimate section 133 .
- the sending/received-sound-ratio calculating section 131 calculates the average electric power of the spectrum R(f) of the received sound and the spectrum S(f) of the sending sound by the following formula.
- R_ave average electric power of the spectrum of the received sound
- the sending/received-sound-ratio calculating section 131 obtains a sending/received sound ratio sp_ratio, for example, from the average power R_ave of the received sound and the average power S_ave of the sending sound by the following formula.
- sp _ratio S _ave/ R _ave FORMULA (5)
- the sending/received-sound-ratio calculating section 131 sets the sending/received sound ratio to the ratio of the volume of the received sound to the volume of the sending sound.
- the sending/received-sound-ratio calculating section 131 outputs the obtained sending/received sound ratio to the age estimate section 132 .
- the age estimate section 132 estimates the age of a user with referring to information that indicates a relationship between the sending/received sound ratio and age, which is stored beforehand.
- FIG. 4 is a schematic view illustrating an example of a relationship between a sending/received sound ratio and age.
- a relationship between age and the sending/received sound ratio can be estimated with the following steps:
- the age estimate section 132 holds the information that indicates a relationship between age and sending/received sound ratio.
- the age estimate section 132 estimates an age from the relationship illustrated in FIG. 4 .
- the age estimate section 132 identified the information indicating the relationship illustrated in FIG. 4 that corresponds to the received sound volume whose sending/received sound ratio has been obtained.
- the age estimate section 132 estimates an age from the information indicating the identified relationship and the calculated sending/received sound ratio.
- the age estimate section 132 outputs the estimated age to the minimum audible range estimate section 133 .
- the minimum audible range estimate section 133 estimates a minimum audible range.
- the minimum audible range estimate section 133 holds a minimum audible range for each generation, based on the relationship illustrated in FIG. 1 .
- FIG. 5 is a schematic view illustrating an example of a relationship between age and a minimum audible range.
- a 1 illustrated in FIG. 5 represents a minimum audible range of 60's
- a 2 represents a minimum audible range of 40's
- a 3 represents a minimum audible range of 20's.
- the relationship between generation and a minimum audible range are provided with intervals of 20 years in the example illustrated in FIG. 5 , it may be provided with intervals of 10 years.
- the minimum audible range estimate section 133 holds data that indicates the relationship between generation and a minimum audible range as illustrated in FIG. 5 to obtain a minimum audible range corresponding to the age obtained from the estimate section 132 .
- the minimum audible range estimate section 133 outputs the obtained minimum audible range to the hearing compensation section 105 .
- a hearing reduction quantity for each generation may be used.
- a minimum audible range or a hearing reduction quantity based on gender may be used. For a difference of hearing characteristics by gender, see p. 72-73 of “Building Environment for Aged People”, edited by Architectural Institute of Japan, 1994 Jan. 10, Shokokusha Publishing Co., Ltd.
- the noise estimate section 104 estimates noise of the surroundings from the sending sound. For example, the noise estimate section 104 estimates noise from the average power S_ave of the sending sound of a current frame.
- noise_level(f) noise quantity
- an initial value of noise_level(f) is arbitrary.
- the initial value may be set to 0.
- ⁇ is a constant between 0 and 1.
- ⁇ is set to, for example, 0.1.
- the threshold value TH may be set from 40 to 50 dB.
- the threshold value TH is set smaller than a volume of a human voice because the volume of voices in people's conversation is 70 to 80 dB.
- the noise estimate section 104 outputs the estimated noise quantity to the hearing compensation section 105 .
- the hearing compensation section 105 compensates hearing (for example, a minimum audible range) based on the minimum audible range obtained from the hearing estimate section 103 and the noise quantity obtained from the noise estimate section 104 .
- FIG. 6 is a block diagram illustrating an example of a configuration of the hearing compensation section 105 .
- the hearing compensation section 105 includes a compensation-quantity calculating section 151 and a minimum audible range compensating section 152 .
- the compensation-quantity calculating section 151 calculates a compensation quantity in response to the noise quantity obtained from the noise estimate section 104 .
- the compensation-quantity calculating section 151 outputs the noise quantity to the minimum audible range compensating section 152 .
- the minimum audible range compensating section 152 compensates the minimum audible range based on the minimum audible range obtained from the hearing estimate section 103 and the compensation quantity obtained from the compensation-quantity calculating section 151 .
- the minimum audible range compensating section 152 adds the obtained compensation quantity to the obtained minimum audible range.
- the compensation-quantity calculating section 151 holds a compensation quantity suited to a noise quantity.
- FIG. 7 is a schematic view illustrating an example of a relationship between noise quantity and compensation quantity.
- noise is classified into three levels, loud, normal, and soft, and corresponding compensation quantities to the levels are illustrated.
- B 1 represents a compensation quantity of a spectrum whose noise quantity is loud
- B 2 represents a compensation quantity of a spectrum whose noise quantity is normal
- B 3 represents a compensation quantity of the spectrum whose noise quantity is soft.
- the compensation-quantity calculating section 151 determines which noise level corresponds to the obtained noise quantity with a determination using a threshold value or the like, to obtain a compensation quantity in response to the determination result from the relationship illustrated in FIG. 7 .
- the compensation-quantity calculating section 151 outputs the obtained compensation quantity to the minimum audible range compensating section 152 .
- FIG. 8 is a schematic view illustrating an example of minimum audible ranges before and after compensation.
- C 1 represents a minimum audible range after compensation
- C 2 represents a minimum audible range before compensation.
- the minimum audible range compensating section 152 obtains a minimum audible range after compensation (C 1 illustrated in FIG. 8 ) by adding to either of the compensation quantity B 1 to B 3 in FIG. 7 to the minimum audible range of C 2 illustrated in FIG. 8 .
- the minimum audible range compensating section 152 outputs the compensated minimum audible range H′(f) to the spectrum compensation-quantity calculating section 106 .
- the compensated minimum audible range H′(f) is, for example, C 1 illustrated in FIG. 8 .
- the compensation-quantity calculating section 151 calculates a compensation quantity by multiplying the noise quantity, or noise_level(f) which has been obtained from the noise estimate section 104 , by constant ⁇ .
- ⁇ is a constant set to, for example, 0.1.
- the compensation-quantity calculating section 151 outputs the calculated compensation quantity to the minimum audible range compensating section 152 .
- noise_level(f) noise quantity
- FIG. 9 is a schematic view illustrating an example of a relationship between noise quantity and minimum audible ranges before and after compensation.
- D 1 represents a minimum audible range after compensation
- D 2 represents a minimum audible range before compensation
- D 3 represents a noise quantity.
- the minimum audible range compensating section 152 obtains the compensated minimum audible range (D 1 illustrated in FIG. 9 ) by adding a compensation quantity multiplied by the constant ⁇ , to the minimum audible range of D 2 illustrated in FIG. 9 .
- the minimum audible range compensating section 152 outputs the compensated minimum audible range H′(f) to the spectrum compensation-quantity calculating section 106 .
- the spectrum compensation-quantity calculating section 106 outputs the obtained spectrum compensation quantity G(f) to the spectrum compensation section 107 .
- the spectrum compensation section 107 obtains compensated received sound spectrum R′(f) from, for example, the spectrum R(f) of the received sound and the spectrum compensation quantity G(f) by the following formula.
- R ′( f ) R ( f )+ G ( f ) FORMULA (8)
- FIG. 10 is a schematic view illustrating spectrum compensation.
- E 1 illustrated in FIG. 10 represents a compensated minimum audible range H′(f)
- E 2 represents a received sound spectrum R(f) before compensation
- E 3 represents a compensated received sound spectrum R′(f).
- the spectrum compensation section 107 compensates the received sound spectrum R(f) so that the compensated received sound spectrum R′(f) becomes greater than the minimum audible range H′(f).
- the spectrum compensation section 107 outputs the compensated spectrum R′(f) of the received sound to the inverse frequency transform section 108 .
- the spectrum compensation-quantity calculating section 106 may compensate the received sound spectrum only within a predetermined frequency band.
- the predetermined frequency band is, for example, a low frequency band and/or a high frequency band where hearing tends to be reduced. This is because bands where hearing tends to be reduced are known.
- the inverse frequency transform section 108 applies an inverse frequency transform (inverse FFT) to the compensated received sound spectrum R′(f) obtained from the spectrum compensation section 107 , to obtain compensated received signal r′(t).
- inverse FFT inverse frequency transform
- the voice control device 1 estimates hearing of a user based on a ratio of volume of sending sound and volume of received sound, and controls the voice in response to the hearing of a user to provide a voice easy to hear for the user automatically.
- the voice control device 1 compensates a minimum audible range estimated from the age of the user based on estimated noise to provide a voice even easier to hear for the user.
- the noise estimate section 104 and the hearing compensation section 105 are not necessarily required to be configured.
- the spectrum compensation-quantity calculating section 106 may calculate the spectrum compensation quantity using the hearing (minimum audible range) estimated by the hearing estimate section 103 .
- FIG. 11 is a flowchart illustrating an example of a voice control procedure according to the embodiment.
- the sending/received-sound-ratio calculating section 131 calculates a sending/received sound ratio from the volume of a received sound and the volume of a sending sound.
- the age estimate section 132 estimates an age from the information that indicates a relationship between the sending/received sound ratio and age.
- the minimum audible range estimate section 133 estimates a minimum audible range from the information that indicates a relationship between age (or generation) and minimum audible range.
- the hearing compensation section 105 compensates the estimated minimum audible range based on noise included in the sending sound. This compensation procedure will be described using FIG. 12 .
- the spectrum compensation-quantity calculating section 106 calculates the compensation quantity of the received sound spectrum so that the received sound becomes greater than the compensated minimum audible range.
- the spectrum compensation section 107 compensates the received signal by adding the calculated compensation quantity, or the like.
- FIG. 12 is a flowchart illustrating an example of a hearing compensation procedure.
- the noise estimate section 104 determines whether the average electric power of sending sound (also called, “sending power” hereafter) is less than the threshold value TH. If the sending power is less than the threshold value TH (Step S 201 —YES), Step S 202 is taken, or if the sending power is greater than or equal to the threshold value TH (Step S 202 —NO), Step S 203 is taken. If the sending power is less than TH, the noise estimate section 104 determines that the sending sound includes no sound.
- sending power also called, “sending power” hereafter
- the noise estimate section 104 updates the noise quantity using the sending sound spectrum of a current frame by the formula (6).
- the hearing compensation section 105 compensates the minimum audible range based on the estimated noise quantity (see FIGS. 8-9 ).
- voice control in response to a user's hearing without forcing a burden on the user according to the embodiment. Also, it is possible to execute voice control suited to a user because the user is not required to carry out a voice control operation, hence voice control can be done automatically during a call.
- the processing by the hearing estimate section 103 may be done at a predetermined timing (once a week, once a month, etc.) so that only hearing compensation by a noise quantity is executed usually. This is because it is not required to execute a hearing estimation every time if the user remains unchanged.
- the sending/received-sound-ratio calculating section 131 may calculate a sending/received sound ratio if a sending sound and a received sound include sound (voice).
- a determination of sound inclusion may be done with a known technology.
- voice or non-voice is determined for each frame of an input signal, based on a first voice characterizing quantity calculated using electric power, a zero crossing rate, a peak frequency of a power spectrum, pitch cycle, etc., and a second voice characterizing quantity calculated based on a difference of peak frequency of power spectrum only in high order components. This makes it possible to estimate the hearing of a user based on the volume of a sending sound and the volume of a received sound.
- FIG. 13 is a block diagram illustrating an example of hardware of a portable terminal device 200 according to a modified example.
- the portable terminal device 200 includes an antenna 201 , a radio section 202 , a baseband processing section 203 , a control section 204 , a microphone 205 , a speaker 206 , a main memory section 207 , an auxiliary storage section 208 , and a terminal interface section 209 .
- the antenna 201 sends a radio signal amplified by the sending amplifier, and receives a radio single sent from a base station.
- the radio section 202 applies a D/A conversion to the sending signal spread by the baseband processing section 203 , converts it to a high frequency signal by a quadrature modulation, and amplifies the signal by a power amplifier.
- the radio section 202 amplifies the received radio signal and applies an A/D conversion to the amplified signal to send it to the baseband processing section 203 .
- the baseband section 203 executes various baseband processing such as an addition of error correcting codes to sending data, data modulation, spread modulation, despreading of a received signal, determination of receiving environment, determination of a threshold value for each channel signal, error correction decoding, etc.
- the control section 204 executes radio control such as sending/receiving of a control signal. Also, the control section 204 executes the voice control program stored in the auxiliary storage section 208 to execute voice control according to the embodiment.
- the main memory section 207 is a ROM (Read-Only Memory), a RAM (Random Access Memory) or the like, which is a storage device to store or to temporarily store an OS, or the basic software executed by the control section 204 , programs such as application software or the like, and data.
- ROM Read-Only Memory
- RAM Random Access Memory
- the auxiliary storage section 208 is an HDD (Hard Disk Drive) or the like, which is a storage device to store data related to the application software and the like. For example, the information illustrated in FIGS. 4 , 5 and 7 is stored in the auxiliary storage section 208 .
- HDD Hard Disk Drive
- the terminal interface section 209 executes adapter processing for data, and interface processing with a handset and an external data terminal.
- the voice control device 1 makes it possible to provide a voice in response to hearing of a user automatically during a call on the portable terminal device 200 . Also, it is possible to implement the voice control device 1 according to the embodiments as one or multiple semiconductor integrated circuits in the portable terminal device 200 .
- the disclosed technology can be implemented not only on the portable terminal device 200 , but also on other devices.
- the voice control device is implemented on a portable terminal device
- the voice control device described above or the voice control procedures described above are applicable, for example, to a TV telephone conference device and an information processing device having a telephone function, a fixed telephone, and the like.
- a recording medium that records information optically, electrically, or magnetically such as a CD-ROM, a flexible disk, an optical magnetic disk and the like, and a semiconductor memory and the like that records information electrically such as a ROM, a flash memory, and the like.
Landscapes
- Health & Medical Sciences (AREA)
- Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Otolaryngology (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Medical Informatics (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- Animal Behavior & Ethology (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Surgery (AREA)
- Molecular Biology (AREA)
- Neurosurgery (AREA)
- Quality & Reliability (AREA)
- Computational Linguistics (AREA)
- Human Computer Interaction (AREA)
- Telephone Function (AREA)
- Circuit For Audible Band Transducer (AREA)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2011/050017 WO2012093470A1 (ja) | 2011-01-04 | 2011-01-04 | 音声制御装置、音声制御方法及び音声制御プログラム |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/050017 Continuation WO2012093470A1 (ja) | 2011-01-04 | 2011-01-04 | 音声制御装置、音声制御方法及び音声制御プログラム |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130279709A1 US20130279709A1 (en) | 2013-10-24 |
US9271089B2 true US9271089B2 (en) | 2016-02-23 |
Family
ID=46457338
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/924,071 Expired - Fee Related US9271089B2 (en) | 2011-01-04 | 2013-06-21 | Voice control device and voice control method |
Country Status (5)
Country | Link |
---|---|
US (1) | US9271089B2 (ja) |
EP (1) | EP2662855A4 (ja) |
JP (1) | JP5626366B2 (ja) |
CN (1) | CN103282960B (ja) |
WO (1) | WO2012093470A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10362394B2 (en) | 2015-06-30 | 2019-07-23 | Arthur Woodrow | Personalized audio experience management and architecture for use in group audio communication |
US10993048B2 (en) | 2017-05-09 | 2021-04-27 | Gn Hearing A/S | Speech intelligibility-based hearing devices and associated methods |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2014064093A (ja) * | 2012-09-20 | 2014-04-10 | Sony Corp | 信号処理装置およびプログラム |
CN110808723B (zh) | 2014-05-26 | 2024-09-17 | 杜比实验室特许公司 | 音频信号响度控制 |
KR20160000345A (ko) | 2014-06-24 | 2016-01-04 | 엘지전자 주식회사 | 이동 단말기 및 그 제어 방법 |
JP6596865B2 (ja) * | 2015-03-23 | 2019-10-30 | 日本電気株式会社 | 電話機、電話システム、電話機の音量設定方法、及びプログラム |
KR101941680B1 (ko) * | 2018-07-13 | 2019-01-23 | 신의상 | 사용자 맞춤형 오디오 주파수 이퀄라이저 조절 장치 및 방법 |
CN114900767B (zh) * | 2022-04-28 | 2023-06-13 | 歌尔股份有限公司 | 听力的保护方法、装置、终端设备及存储介质 |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06217398A (ja) | 1993-01-18 | 1994-08-05 | Hitachi Ltd | 聴覚特性補償用音声処理装置 |
JPH0766767A (ja) | 1993-08-23 | 1995-03-10 | Toshiba Corp | 無線通信装置 |
JPH08163121A (ja) | 1994-12-06 | 1996-06-21 | Canon Inc | 映像コミュニケーションシステム |
JPH08223256A (ja) | 1995-02-15 | 1996-08-30 | Nec Corp | 電話装置 |
US5777664A (en) | 1994-11-18 | 1998-07-07 | Canon Kabushiki Kaisha | Video communication system using a repeater to communicate to a plurality of terminals |
JP2000209698A (ja) | 1999-01-13 | 2000-07-28 | Nec Saitama Ltd | 音声補正装置及び音声補正機能付き移動機 |
JP2004165865A (ja) | 2002-11-12 | 2004-06-10 | Mitsubishi Electric Corp | 電話機及び受話音量調整方法 |
JP2004235708A (ja) | 2003-01-28 | 2004-08-19 | Mitsubishi Electric Corp | 電話装置 |
US20060088154A1 (en) | 2004-10-21 | 2006-04-27 | Motorola, Inc. | Telecommunication devices that adjust audio characteristics for elderly communicators |
US20070198263A1 (en) * | 2006-02-21 | 2007-08-23 | Sony Computer Entertainment Inc. | Voice recognition with speaker adaptation and registration with pitch |
JP2009171189A (ja) | 2008-01-16 | 2009-07-30 | Pioneer Electronic Corp | 音声補正装置及び音声補正機能を備えた通信端末装置 |
JP2010028515A (ja) | 2008-07-22 | 2010-02-04 | Nec Saitama Ltd | 音声強調装置、携帯端末、音声強調方法および音声強調プログラム |
WO2010035308A1 (ja) | 2008-09-24 | 2010-04-01 | 三菱電機株式会社 | エコー消去装置 |
JP2010081523A (ja) | 2008-09-29 | 2010-04-08 | Nec Corp | 携帯端末、携帯端末の制御方法、及びプログラム |
US8560308B2 (en) * | 2008-07-02 | 2013-10-15 | Fujitsu Limited | Speech sound enhancement device utilizing ratio of the ambient to background noise |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3849116B2 (ja) | 2001-02-28 | 2006-11-22 | 富士通株式会社 | 音声検出装置及び音声検出プログラム |
AU2003263380A1 (en) * | 2002-06-19 | 2004-01-06 | Koninklijke Philips Electronics N.V. | Audio signal processing apparatus and method |
CN101166017B (zh) * | 2006-10-20 | 2011-12-07 | 松下电器产业株式会社 | 用于声音产生设备的自动杂音补偿方法及装置 |
JP2010062663A (ja) * | 2008-09-01 | 2010-03-18 | Sony Ericsson Mobilecommunications Japan Inc | 音声信号処理装置、音声信号処理方法、及び、通信端末 |
JP2010239542A (ja) * | 2009-03-31 | 2010-10-21 | Yamaha Corp | 音声処理装置 |
CN103270740B (zh) * | 2010-12-27 | 2016-09-14 | 富士通株式会社 | 声音控制装置、声音控制方法以及移动终端装置 |
-
2011
- 2011-01-04 CN CN201180063528.8A patent/CN103282960B/zh not_active Expired - Fee Related
- 2011-01-04 EP EP11855034.2A patent/EP2662855A4/en not_active Withdrawn
- 2011-01-04 JP JP2012551766A patent/JP5626366B2/ja not_active Expired - Fee Related
- 2011-01-04 WO PCT/JP2011/050017 patent/WO2012093470A1/ja active Application Filing
-
2013
- 2013-06-21 US US13/924,071 patent/US9271089B2/en not_active Expired - Fee Related
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06217398A (ja) | 1993-01-18 | 1994-08-05 | Hitachi Ltd | 聴覚特性補償用音声処理装置 |
JPH0766767A (ja) | 1993-08-23 | 1995-03-10 | Toshiba Corp | 無線通信装置 |
US5777664A (en) | 1994-11-18 | 1998-07-07 | Canon Kabushiki Kaisha | Video communication system using a repeater to communicate to a plurality of terminals |
JPH08163121A (ja) | 1994-12-06 | 1996-06-21 | Canon Inc | 映像コミュニケーションシステム |
JPH08223256A (ja) | 1995-02-15 | 1996-08-30 | Nec Corp | 電話装置 |
JP2000209698A (ja) | 1999-01-13 | 2000-07-28 | Nec Saitama Ltd | 音声補正装置及び音声補正機能付き移動機 |
JP2004165865A (ja) | 2002-11-12 | 2004-06-10 | Mitsubishi Electric Corp | 電話機及び受話音量調整方法 |
JP2004235708A (ja) | 2003-01-28 | 2004-08-19 | Mitsubishi Electric Corp | 電話装置 |
US20060088154A1 (en) | 2004-10-21 | 2006-04-27 | Motorola, Inc. | Telecommunication devices that adjust audio characteristics for elderly communicators |
US20070198263A1 (en) * | 2006-02-21 | 2007-08-23 | Sony Computer Entertainment Inc. | Voice recognition with speaker adaptation and registration with pitch |
JP2009171189A (ja) | 2008-01-16 | 2009-07-30 | Pioneer Electronic Corp | 音声補正装置及び音声補正機能を備えた通信端末装置 |
US8560308B2 (en) * | 2008-07-02 | 2013-10-15 | Fujitsu Limited | Speech sound enhancement device utilizing ratio of the ambient to background noise |
JP2010028515A (ja) | 2008-07-22 | 2010-02-04 | Nec Saitama Ltd | 音声強調装置、携帯端末、音声強調方法および音声強調プログラム |
WO2010035308A1 (ja) | 2008-09-24 | 2010-04-01 | 三菱電機株式会社 | エコー消去装置 |
US20110135105A1 (en) | 2008-09-24 | 2011-06-09 | Atsuyoshi Yano | Echo canceller |
JP2010081523A (ja) | 2008-09-29 | 2010-04-08 | Nec Corp | 携帯端末、携帯端末の制御方法、及びプログラム |
Non-Patent Citations (2)
Title |
---|
Extended European Search Report dated Nov. 18, 2015 for corresponding European Patent Application No. 11855034.2, 8 pages. |
International Search Report, mailed in connection with PCT/JP2011/050017 and mailed Feb. 8, 2011. |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10362394B2 (en) | 2015-06-30 | 2019-07-23 | Arthur Woodrow | Personalized audio experience management and architecture for use in group audio communication |
US10993048B2 (en) | 2017-05-09 | 2021-04-27 | Gn Hearing A/S | Speech intelligibility-based hearing devices and associated methods |
Also Published As
Publication number | Publication date |
---|---|
EP2662855A1 (en) | 2013-11-13 |
EP2662855A4 (en) | 2015-12-16 |
JPWO2012093470A1 (ja) | 2014-06-09 |
WO2012093470A1 (ja) | 2012-07-12 |
CN103282960A (zh) | 2013-09-04 |
US20130279709A1 (en) | 2013-10-24 |
JP5626366B2 (ja) | 2014-11-19 |
CN103282960B (zh) | 2016-01-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9271089B2 (en) | Voice control device and voice control method | |
US9330678B2 (en) | Voice control device, voice control method, and portable terminal device | |
US9420370B2 (en) | Audio processing device and audio processing method | |
JP5704470B2 (ja) | オーディオ明瞭度増大方法および装置とコンピュータ装置 | |
US8571231B2 (en) | Suppressing noise in an audio signal | |
US9197181B2 (en) | Loudness enhancement system and method | |
US20100217606A1 (en) | Signal bandwidth expanding apparatus | |
TW201142831A (en) | Adaptive environmental noise compensation for audio playback | |
US9160404B2 (en) | Reverberation reduction device and reverberation reduction method | |
US8924199B2 (en) | Voice correction device, voice correction method, and recording medium storing voice correction program | |
JP2010532879A (ja) | アダプティブ・インテリジェント・ノイズ抑制システム及び方法 | |
US20100111290A1 (en) | Call Voice Processing Apparatus, Call Voice Processing Method and Program | |
EP2200340A1 (en) | Sound processing methods and apparatus | |
US20140185818A1 (en) | Sound processing device, sound processing method, and program | |
JP2008309955A (ja) | ノイズサプレス装置 | |
US8406430B2 (en) | Simulated background noise enabled echo canceller | |
US9177566B2 (en) | Noise suppression method and apparatus | |
JP4460256B2 (ja) | 雑音低減処理方法、この方法を実施する装置、プログラム、記録媒体 | |
US11195539B2 (en) | Forced gap insertion for pervasive listening | |
JP5172580B2 (ja) | 音補正装置及び音補正方法 | |
US8457955B2 (en) | Voice reproduction with playback time delay and speed based on background noise and speech characteristics |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FUJITSU LIMITED, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SUZUKI, MASANAO;OTANI, TAKESHI;TOGAWA, TARO;AND OTHERS;SIGNING DATES FROM 20130617 TO 20130618;REEL/FRAME:030835/0804 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20200223 |