JP2014090350A - Voice processing device, communication device, voice control method, and program - Google Patents

Voice processing device, communication device, voice control method, and program Download PDF

Info

Publication number
JP2014090350A
JP2014090350A JP2012239841A JP2012239841A JP2014090350A JP 2014090350 A JP2014090350 A JP 2014090350A JP 2012239841 A JP2012239841 A JP 2012239841A JP 2012239841 A JP2012239841 A JP 2012239841A JP 2014090350 A JP2014090350 A JP 2014090350A
Authority
JP
Japan
Prior art keywords
approach
voice
detection unit
sound
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP2012239841A
Other languages
Japanese (ja)
Inventor
Daisuke Sugii
大介 杉井
Toshiharu Aihara
敏治 相原
Yasuharu Onishi
康晴 大西
Atsushi Kuroda
淳 黒田
Original Assignee
Nec Casio Mobile Communications Ltd
Necカシオモバイルコミュニケーションズ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nec Casio Mobile Communications Ltd, Necカシオモバイルコミュニケーションズ株式会社 filed Critical Nec Casio Mobile Communications Ltd
Priority to JP2012239841A priority Critical patent/JP2014090350A/en
Publication of JP2014090350A publication Critical patent/JP2014090350A/en
Application status is Pending legal-status Critical

Links

Images

Abstract

To improve convenience by controlling a mute operation according to an actual action of a user without making the user aware of it.
An audio processing device includes an audio detection unit, a first approach detection unit that detects a first approach for closing a sound hole that guides audio to the audio detection unit, an audio output unit, and an audio output unit. Based on the second approach detection unit 23 for detecting the second approach for listening to the sound output from the sound output unit 22 and the detection result of the first approach by the first approach detection unit 21, the sound Controls whether or not the voice detected by the detection unit 20 is transmitted to the call partner, and the first approach detection result by the first approach detection unit 21 and the second approach by the second approach detection unit 23 And a control unit 24 for controlling whether to continue transmitting the sound detected by the sound detecting unit 20 or canceling the non-transmitting state and transmitting the sound based on the detection result.
[Selection] Figure 5

Description

  The present invention relates to a voice processing device, a communication device, a voice control method, and a program.

  In mobile phone terminals, in order to respond to actions such as being able to speak to a third party during a call, it is a common action to close the microphone sound hole with your finger or hand so that you do not convey the surrounding voice to the other party. . However, if the microphone sound hole is closed with a finger or hand in this way, the friction sound of the finger or hand is transmitted or the sound hole cannot be sufficiently closed, and the voice of the other party can be heard. In view of this, a technique has been proposed in which the microphone is turned off by a predetermined operation.

  In Patent Document 1, when the user releases the mobile phone terminal from the ear, the proximity sensor detects that the ear has been released, switches to mute to turn off the microphone, and the user switches the mobile phone terminal to resume the call. There has been proposed a technique for releasing mute when it is brought close to the ear again.

  In Patent Document 2, when a user touches a contact detection unit provided near the microphone during a call, the sound from the microphone is muted, and when the finger is released from the contact detection unit, the sound is muted. The technology to cancel is proposed.

JP 2012-010183 A JP 2007-081460 A

  However, in Patent Document 1, since it cannot be muted unless it is separated from the ear, the microphone sound hole is closed with a finger or a hand without being separated from the ear so as not to convey the surrounding speech to the other party during the call. Cannot be muted.

  Also, in Patent Document 2, in order to continue the muted state, the finger cannot be lifted from the contact detection unit. For example, when searching for a notepad or a document during a call, the cellular phone terminal is moved away from the ear and muted. In this state, no other work can be performed.

  As described above, Patent Documents 1 and 2 have a problem in that the mute operation cannot be controlled according to the actual behavior of the user during a call, and convenience is poor.

  Therefore, the present invention can control a mute operation according to a user's actual behavior without making the user aware of it, and can improve convenience, a voice processing device, a communication device, a voice control method, and a program The purpose is to provide.

  The speech processing apparatus of the present invention includes a speech detection unit, a first approach detection unit that detects a first approach for closing a sound hole that guides speech to the speech detection unit, a speech output unit, and the speech output. Detected by the voice detection unit based on a second approach detection unit for detecting a second approach for listening to the voice output from the unit and a detection result of the first approach by the first approach detection unit Based on the detection result of the first approach by the first approach detection unit and the detection result of the second approach by the second approach detection unit. And a control unit that controls whether the invalid state of the voice detected by the voice detection unit is continued or the invalid state is canceled and validated.

  The speech processing apparatus of the present invention includes a speech detection unit, a first approach detection unit that detects a first approach for closing a sound hole that guides speech to the speech detection unit, a speech output unit, and the speech output. A second proximity detection unit for detecting a second approach for listening to the voice output from the unit, and the voice detection unit based on a detection result of the first approach by the first proximity detection unit Control whether to transmit the transmitted voice to the call partner or not, and to detect a first approach detected by the first approach detector and a second approach detected by the second approach detector And a control unit that controls whether to continue the state in which the voice detected by the voice detection unit is not transmitted, or to cancel the state in which the voice is not transmitted and set the state to be transmitted. is there.

  The voice control method of the present invention includes a first approach detection step for detecting a first approach for closing a sound hole that guides voice to the voice detection unit, and a second for listening to the voice output from the voice output unit. Based on the second approach detection step for detecting the approach of the first approach and the detection result of the first approach, controlling whether to invalidate or validate the voice detected by the voice detector, A control step for controlling whether to continue the invalid state of the voice detected by the voice detection unit or to cancel the invalid state based on the detection result of the second approach and the detection result of the second approach Is a voice control method characterized by including:

  The program according to the present invention includes a first approach detection function for detecting a first approach for closing a sound hole that guides sound to the sound detection unit, and a second for listening to sound output from the sound output unit. Whether the voice detected by the voice detection unit is invalidated or validated based on the detection result of the first approach by the second approach detection function for detecting the approach of the first approach detection function; And is detected by the voice detection unit based on the detection result of the first approach by the first approach detection function and the detection result of the second approach by the second approach detection function. A program for executing a control function for controlling whether to keep a voice invalid state or to cancel and invalidate a voice.

  According to the present invention, the mute operation can be controlled in accordance with the actual action of the user without making the user aware of it, and the convenience can be improved.

It is a block diagram which shows the schematic structure (microphone mute function) of the mobile telephone terminal 1 by 1st Embodiment of this invention. 4 is a flowchart for explaining the operation of the mobile phone terminal 1 according to the first embodiment. 6 is a flowchart for explaining the operation of the mobile phone terminal 1 according to the second embodiment. 6 is a flowchart for explaining the operation of the mobile phone terminal 1 according to the second embodiment. FIG.

  Hereinafter, embodiments of the present invention will be described with reference to the drawings.

A. First Embodiment FIG. 1 is a block diagram showing a schematic configuration (microphone mute function) of a mobile phone terminal 1 according to a first embodiment of the present invention. In the figure, the mobile phone terminal 1 includes one or more microphones 2, a microphone amplifier 3, a receiver 4, a speaker 5, a speaker amplifier 6, a DSP (Digital Signal Processor) 7, and proximity sensors 11 and 12.

  The microphone 2 collects the user's voice in a voice call or the like. The microphone amplifier 3 amplifies the audio signal input from the microphone 2 and supplies it to the DSP 7. The receiver 4 is a voice output means for outputting the voice of the other party of the call when the mobile phone terminal 1 is pressed against the ear to make a call. The speaker 5 is a voice output means for outputting the voice of the other party of the call, such as hands-free, when the mobile phone terminal 1 is released without being pressed against the ear. The speaker amplifier 6 amplifies the voice signal of the other party from the codec unit 10 to be described later to an output level at which the voice of the other party can be heard even when the user releases the mobile phone terminal 1 and outputs it from the speaker 5. To do.

  The DSP 7 collects sound with the microphone 2 in accordance with digitization of the sound signal collected by the microphone 2 and amplified by the microphone amplifier 3, codec processing of the voice data of the other party, and detection results by the proximity sensors 11 and 12, which will be described later. It controls whether or not the sound is muted, that is, whether or not the user's voice collected by the microphone 2 is transmitted to the other party.

  The proximity sensor 11 is provided in the vicinity of the receiver 4, and detects the approach of the ear when the user presses the ear against the mobile phone terminal 1 (the receiver 4) and makes a call. The proximity sensor 12 is provided in the vicinity of the microphone 2 (in the vicinity of or around the sound hole provided at the position of the microphone 2 for guiding sound to the microphone 2). The proximity sensor 12 detects the approach of a hand, a finger, or the like when performing an operation of closing the sound hole with the user's hand, a finger, or the like so that the user cannot hear his / her voice or ambient sound.

  By the way, in the case of the cellular phone terminal 1, the user's face (mouth) approaches the microphone 2 during a normal call (not hands-free), so the proximity sensor 12 is connected to the user's face (mouth). May be erroneously detected as the approach of a hand, a finger, or the like when performing an operation of closing the sound hole. However, the hand, fingers, etc. when performing the operation of closing the sound hole are closer to the mobile phone terminal 1 (proximity sensor 12) than the user's face (mouth) during a call, or the case or proximity There is a difference that the sensor 12 is in direct contact. Therefore, in the present embodiment, paying attention to this difference, not the user's face (mouth), but the approach (including contact) of a hand or a finger when performing an operation to close the sound hole is detected. The detection sensitivity of the proximity sensor 12, the threshold value of the detection level, and the like are adjusted. In the following description, the proximity sensor 12 is described as detecting the approach of a hand or a finger when performing an operation of closing a sound hole that guides sound to the microphone 2, but the contact of the hand or the finger is detected. May be.

  The mute determination unit 9 is based on whether or not the proximity sensor 12 detects an approach when closing the sound hole, or whether or not the proximity sensor 11 detects an approach when the ear is pressed. Whether the audio data supplied from the microphone 2, the microphone amplifier 3, and the ADC (Analog Digital Converter) 8 is muted, that is, whether the user's voice collected by the microphone 2 is transmitted to the other party. Determine.

  The DSP 7 includes an ADC 8, a mute determination unit 9, and a codec unit 10. The ADC 8 digitizes the audio signal collected by the microphone 2 and amplified by the microphone amplifier 3 and supplies the audio signal to the mute determination unit 9 as audio data.

  The codec unit 10 decodes and decodes the encoded voice data of the communication partner supplied from the communication unit (not shown), and encodes the voice data collected by the microphone 2 and digitized by the ADC 8. Encode. Further, the codec unit 10 outputs the decoded voice data of the other party from the receiver 4 based on whether or not the proximity sensor 11 detects the approach of the ear to be heard by the receiver 4, or the speaker amplifier 6 to switch the output from the speaker 5 via 6.

  More specifically, when the proximity sensor 11 detects the approach of the ear for listening with the receiver 4, that is, when the user presses the mobile phone terminal 1 against the ear, the codec unit 10 The other party's voice data is controlled to be output from the receiver 4. On the other hand, when the proximity sensor 11 does not detect the approach of the ear for listening by the receiver 4, that is, when the user separates the mobile phone terminal 1 from the ear, the codec unit 10 transmits the voice data of the other party to the speaker. It controls to output from the speaker 5 through the amplifier 6.

Next, the operation of the above-described first embodiment will be described.
FIG. 2 is a flowchart for explaining the operation of the cellular phone terminal 1 according to the first embodiment. When a call is started in the mobile phone terminal 1, first, sound is collected by the microphone 2 (step S10), and the collected sound signal is input to the DSP 7 via the microphone amplifier 3 (step S12). In the DSP 7, the ADC 8 digitizes the sound (step S14).

  Next, the mute determining unit 9 determines whether or not the hands-free call state is set (step S16). The hands-free call state is a state in which a call is made with a certain distance away without bringing the mobile phone terminal 1 close to the face. If not in the hands-free call state (NO in step S16), the mute determination unit 9 determines whether or not the proximity sensor 12 detects the approach of a hand or a finger for closing the sound hole ( Step S18). In other words, the mute determination unit 9 performs an operation of closing the sound hole that guides the sound to the microphone 2 with its own hand or finger so that the proximity sensor 12 does not hear the user's own voice and surrounding sounds to the other party. It is determined whether or not an approach of a hand or a finger is detected.

  When the proximity sensor 12 does not detect the approach of the hand or finger for closing the sound hole, that is, when the operation for closing the sound hole is not performed (NO in step S18), the mute determination unit 9 Determines whether or not the proximity sensor 11 has detected an approach when the user presses the mobile phone terminal 1 against the ear to make a call (step S20).

  When the proximity sensor 11 detects the approach of the ear for listening by the receiver 4, that is, when the approach when pressing the ear and making a call is detected (YES in step S20), the mute determination is made. The unit 9 does not mute the input of the microphone 2 and maintains the call state (step S24).

  That is, when the user is talking by pressing the mobile phone terminal 1 against the ear without closing the sound hole that guides the voice to the microphone 2, the voice input from the microphone 2 is transmitted to the other party, A normal call operation is performed in which the encoded voice data of the other party is decoded and decoded by the codec unit 10 and output from the receiver 4.

  On the other hand, when the proximity sensor 12 detects the approach of a hand or a finger for closing the sound hole, that is, when the user closes the sound hole for guiding the sound to the microphone 2 during a call with the hand or the finger. (YES in step S18), the mute determination unit 9 performs control so that the input of the microphone 2 is muted and no sound is transmitted to the call partner (step S22).

  In other words, in order to respond to an action such as a user talking to a third party during a call, a normal operation of closing the sound hole of the microphone 2 with a finger or hand so as not to convey the surrounding speech to the other party is performed. Thus, the input of the microphone 2 is muted. Voice is never sent to the other party. In this case, since the input of the microphone 2 is muted, the frictional sound of the finger or hand may be transmitted, or the voice may be heard by the other party without sufficiently closing the sound hole. Absent.

  Further, during a call, the proximity sensor 12 does not detect the approach of a hand or a finger for closing the sound hole, and the proximity sensor 11 does not detect the approach of the ear for listening by the receiver 4 That is, even when the user does not block the sound hole for guiding the sound to the microphone 2, but the mobile phone terminal 1 is separated from the ear (NO in step S18, NO in step S20), the mute determination unit 9 The input of No. 2 is muted, and control is performed so that the voice is not transmitted to the other party (step S22).

  That is, even when the user removes the mobile phone terminal 1 from his / her ear, it is determined that a normal call (call with the mobile phone terminal 1 held against the ear) cannot be performed for any reason, and the voice is transmitted to the other party. Muted so that it will not be transmitted. For example, this corresponds to a case where the user forgets the operation of closing the sound hole for guiding the sound to the microphone 2 and moves the mobile phone terminal 1 away from the ear. Even in such a case, since the input of the microphone 2 is muted, the surrounding voice is not heard by the other party.

  Further, as can be seen from the above-described operation, the sound hole that guides the sound to the microphone 2 is closed with the user pressing the mobile phone terminal 1 against the ear (NO in step S18, step S22, mute at this point) ) Even if the cellular phone terminal 1 is separated from the ear as it is, the mute state is maintained while the sound hole is closed. Thereafter, even if the user releases the hand or finger that closed the sound hole that guides the sound to the microphone 2, the mobile phone terminal 1 is away from the ear (NO in step S20), so the mute state is still maintained. It will be.

  Therefore, the user does not need to keep the sound hole closed in order to mute, so that the mobile phone terminal 1 is kept away from the ear (such as on a desk) while searching for a notepad or a document during a call. You can do other work with it in place. During this time, the mute state is maintained, so that the surrounding voice is not heard by the other party.

  In the mute state described above, when the user brings the mobile phone terminal 1 to the ear (NO in step S18, YES in step S20), the mute determination unit 9 cancels the mute, and the microphone. The voice input from 2 is transmitted to the call partner, and the call partner's encoded voice data is decoded and decoded by the codec unit 10 and output from the receiver 4 (step S24). .

  If the hands-free call state is set (YES in step S16), the mute determination unit 9 cancels mute and transmits the voice input from the microphone 2 to the call partner (step S24). In the case of a hands-free call state, a call operation is performed in which the encoded voice data of the other party is decoded and decoded by the codec unit 10 and output from the speaker 5 through the speaker amplifier 6.

  Thereafter, in any of the above cases, it is determined whether or not there has been an end call operation (step S26). If there is no end call operation (NO in step S26), the process returns to step S10 and the above-described processing is repeated. . On the other hand, if there is an end call operation (YES in step S26), line disconnection processing is executed (step S28), and the processing ends.

  According to the first embodiment described above, the detection result by the proximity sensor 11 that detects the approach of the ear for listening to the sound output from the receiver 4, and the approach of a hand, a finger, or the like for closing the sound hole that guides the sound In accordance with the detection result of the proximity sensor 12 that detects whether or not the sound from the microphone 2 is transmitted to the other party (mute start, continuation, release), the user can be made unaware of the user. The mute operation (start, continuation, release) can be controlled in accordance with the operation or action, and convenience can be further improved.

  More specifically, according to the first embodiment described above, the microphone is detected when a normal operation of closing a sound hole for guiding sound is detected, or when it is detected that the mobile phone terminal 1 is separated from the ear. Since the mute that disables the voice input from 2 is started, it is possible to shift to the mute state without causing the user to perform an intentional operation such as pressing a key.

  Further, according to the first embodiment described above, when the user removes his / her hand or finger from the sound hole and brings the mobile phone terminal 1 close to the ear, the mute state is released. Without causing an intentional operation, the mute state can be canceled and the normal call operation can be resumed.

  Further, according to the first embodiment described above, when the mobile phone terminal 1 is separated from the ear in the mute state, the mute state is continued even if the sound hole for guiding the sound is not blocked. Therefore, it is no longer necessary to block the sound hole continuously to mute, and other operations can be performed in a hands-free manner such as searching for a memo pad or materials during a call.

B. Second Embodiment Next, a second embodiment of the present invention will be described.
The configuration of the mobile phone terminal 1 according to the second embodiment is the same as that of the first embodiment described above with reference to FIG. In the second embodiment, in order to further improve the convenience for the user, the continuation / cancellation of the mute operation is controlled when the user separates the mobile phone terminal 1 from the ear.

  3 and 4 are flowcharts for explaining the operation of the cellular phone terminal 1 according to the second embodiment. When a call is started in the cellular phone terminal 1, first, the microphone 2 collects sound (step S40), and the collected audio signal is input to the DSP 7 via the microphone amplifier 3 (step S42). In the DSP 7, the ADC 8 digitizes the sound (step S44).

  Next, the mute determination unit 9 determines whether or not it is in a hands-free call state (step S46). If not in the hands-free call state (NO in step S46), the mute determination unit 9 determines whether or not it is currently in a mute state in which the voice is not transmitted to the call partner (step S48). . If the mute state is not set (NO in step S48), the mute determination unit 9 determines whether or not the proximity sensor 12 detects the approach of a hand or a finger for closing the sound hole (step S48). S50). In other words, the mute determination unit 9 performs an operation of closing the sound hole that guides the sound to the microphone 2 with its own hand or finger so that the user cannot hear the user's own sound or ambient sound by the proximity sensor 12 with the proximity sensor 12. It is determined whether or not an approach of a hand or a finger is detected.

  When the proximity sensor 12 does not detect the approach of a hand or finger for closing the sound hole, that is, when the operation for closing the sound hole is not performed (NO in step S50), the mute determination unit 9 Maintains the call state without muting the input of the microphone 2 (step S54).

  In other words, regardless of whether or not the cellular phone terminal 1 is pressed against the ear, when the user is talking without closing the sound hole that guides the voice to the microphone 2, the voice inputted from the microphone 2 is called. In addition to transmitting to the other party, a normal call operation is performed in which the encoded voice data of the other party is decoded and decoded by the codec unit 10 and output from the receiver 4.

  On the other hand, when the proximity sensor 12 detects the approach of a hand or finger to close the sound hole, that is, when the user closes the sound hole that guides the sound to the microphone 2 during a call, or during a call If the user closes the sound hole after releasing the mobile phone terminal 1 from the ear (YES in step S50), the mute determining unit 9 mutes the input of the microphone 2 and does not transmit the voice to the other party. (Step S52).

  In other words, in order to respond to actions such as a user talking to a third party during a call, the cellular phone terminal 1 is pressed against the ear or away from the ear so as not to convey the surrounding voice to the other party. In this state, by performing a normal operation of closing the sound hole of the microphone 2 with a finger or hand, the input of the microphone 2 is muted. Voice is never sent to the other party. In this case, since the input of the microphone 2 is muted, the frictional sound of the finger or hand may be transmitted, or the voice may be heard by the other party without sufficiently closing the sound hole. Absent.

  On the other hand, if the operation is muted during the call due to the above-described operation (YES in step S48), the proximity sensor 11 listens to the receiver 4 when the user presses the mobile phone terminal 1 against the ear to make a call. It is determined whether or not an approaching ear is detected (step S60 in FIG. 4). Here, when the proximity sensor 11 detects the approach of the ear to be heard by the receiver 4, that is, when the user presses the mobile phone terminal 1 against the ear (YES in step S60), the proximity sensor 12 Thus, it is determined whether or not an approach of a hand or a finger for closing the sound hole is detected (step S62). That is, the mute determining unit 9 determines whether or not the sound hole that guides the sound to the microphone 2 is closed.

  When the proximity sensor 12 detects the approach of a hand or finger to close the sound hole, that is, when the sound hole that guides the sound to the microphone 2 is blocked (YES in step S62), While continuing the mute state, the process proceeds to step S56 in FIG. That is, when the sound hole that guides the sound to the microphone 2 remains closed, the state in which the input of the microphone 2 is muted is continued, and the sound is not transmitted to the other party.

  On the other hand, when the proximity sensor 12 does not detect the approach of a hand or a finger for closing the sound hole, that is, when the sound hole for guiding the sound to the microphone 2 is not blocked (NO in step S62), The mute discriminating unit 9 cancels the mute, transmits the voice input from the microphone 2 to the call partner, decodes and decodes the encoded voice data of the call partner by the codec unit 10, and outputs it from the receiver 4. A normal call operation is performed (step S64).

  Next, the codec unit 10 determines whether or not it is a speaker output mode in which the voice of the other party is currently output from the speaker 5 (step S66). If the speaker output mode is selected (YES in step S66), the user is currently talking by pressing the mobile phone terminal 1 against the ear. The mode is changed (step S68), and the process proceeds to step S56 in FIG. On the other hand, if it is not the speaker output mode (NO in step S66), the process proceeds to step S56 in FIG. 3 while remaining in the receiver output mode.

  That is, in the mute state described above, in a state where the user presses the mobile phone terminal 1 against the ear, the user releases his / her hand or finger from the sound hole that guides the sound to the microphone 2, or the mobile phone once released from the ear When the terminal 1 is brought close to the ear again, when the hand or finger is released from the sound hole, the mute is released and the normal call operation can be started.

  In the mute state, when the proximity sensor 11 does not detect the approach of the ear for listening by the receiver 4, that is, when the user does not press the mobile phone terminal 1 against the ear (NO in step S60), It is determined whether or not the first operation has been performed by the user (step S70). The first operation is an operation for instructing the change to the speaker output mode for outputting the voice of the other party, such as a specific hardware key operation, a specific software key operation, a touch panel, etc. If it is provided, it corresponds to a specific touch operation. At this time, it may be displayed on the display unit (not shown) that the speaker output mode can be changed by the first operation.

  Here, when the first operation is performed (YES in step S70), the codec unit 10 changes the voice of the other party to the speaker output mode for outputting the voice of the other party from the speaker 5 (step S72). ). In this case, the mute state is continued. Then, the process proceeds to step S58 in FIG.

  That is, when the first operation is performed with the mobile phone terminal 1 away from the ear, the mute determination unit 9 maintains the mute state, and the codec unit 10 outputs the voice of the other party from the speaker 5. Accordingly, the user can perform other operations in a hands-free manner such as searching for a memo pad or a document while listening to the voice of the other party.

  If the first operation is not performed (NO in step S70), it is determined whether or not the second operation is performed by the user (step S74). Unlike the first operation, the second operation is an operation for canceling the mute state and instructing the change to the speaker output mode for outputting the voice of the other party. If a hardware key operation, a specific software key operation, or a touch panel is provided, this corresponds to a specific touch operation.

  Here, when the second operation is performed (YES in step S74), the mute determination unit 9 cancels the mute state, transmits the voice input from the microphone 2 to the call partner, and the codec unit. 10 changes the other party's voice to the speaker output mode for outputting the other party's voice from the speaker 5 (step S76). That is, the operation is the same as in the hands-free mode. Then, the process proceeds to step S58 in FIG.

  That is, when the second operation is performed with the mobile phone terminal 1 away from the ear, the mute determination unit 9 releases the mute state, and the codec unit 10 outputs the voice of the other party from the speaker 5. The user can perform other operations in a hands-free manner such as searching for a memo pad or a document while continuing the call with the other party.

  Further, when the user does not perform the first operation or the second operation with the mobile phone terminal 1 away from the ear in the mute state (NO in step S70, NO in step S74). 3) Even if the sound hole for guiding the sound is not blocked, the process proceeds to step S56 in FIG. 3 while the mute state is continued.

  That is, when neither the first operation nor the second operation is performed with the mobile phone terminal 1 away from the ear, the mute state is continued. There is no need to keep blocking. For this reason, other operations can be performed in a hands-free manner such as searching for a notepad or a document during a call. During this time, the mute state is maintained, so that the surrounding voice is not heard by the other party.

  If it is in a hands-free call state (YES in step S46), the mute determination unit 9 cancels mute and transmits the voice input from the microphone 2 to the call partner (step S54). In the case of a hands-free call state, a call operation is performed in which the encoded voice data of the other party is decoded and decoded by the codec unit 10 and output from the speaker 5 through the speaker amplifier 6.

  In any case, it is determined whether or not there is an end operation (step S56 in FIG. 3). If there is no end operation (NO in step S56), the process returns to step S40 and the above-described processing is repeated. . On the other hand, if there is an end call operation (YES in step S56), line disconnection processing is executed (step S58), and the processing ends.

  According to the second embodiment described above, the detection result by the proximity sensor 11 that detects the approach of the ear for listening to the sound output from the receiver 4 and the approach of a hand, a finger, or the like for closing the sound hole that guides the sound. In accordance with the detection result of the proximity sensor 12 that detects whether or not the sound from the microphone 2 is transmitted to the other party (mute start, continuation, release), the user can be made unaware of the user. The mute operation (start, continuation, release) can be controlled in accordance with the operation or action, and convenience can be further improved.

  More specifically, according to the second embodiment described above, even when a normal operation by the user to close a sound hole that guides sound is detected while the mobile phone terminal 1 is pressed against the ear, Even when the normal operation by the user to close the sound hole for guiding the sound is detected with the mobile phone terminal 1 away from the ear, the mute that invalidates the sound input from the microphone 2 is started. As a result, it is possible to shift to the mute state without causing an intentional operation such as pressing a key.

  Further, according to the second embodiment described above, when the user removes his / her hand or finger from the sound hole, the mute is released, so that the user does not perform an intentional operation such as pressing a key. The mute state can be canceled and the normal call operation can be resumed.

  Further, according to the second embodiment described above, when the mobile phone terminal 1 is separated from the ear in the mute state, the mute state is continued even if the sound hole for guiding the sound is not blocked. Therefore, it is no longer necessary to block the sound hole continuously to mute, and other operations can be performed in a hands-free manner such as searching for a memo pad or materials during a call.

  Further, according to the second embodiment described above, when the first operation is performed in the mute state, the sound output for outputting the other party's voice from the speaker 5 is changed to the speaker output mode while the mute state is continued. So, while listening to the other party's voice, you can do other work in a hands-free manner, such as searching for a notepad or material.

  Further, according to the second embodiment described above, when the second operation is performed in the mute state, the mute state is canceled and the sound output for outputting the voice of the other party from the speaker 5 is changed to the speaker output mode. Because of this, you can do other tasks in a hands-free manner, such as searching for memos and materials while continuing to talk to the other party.

  In the second embodiment described above, when the first operation or the second operation is performed in steps S70 to S76 shown in FIG. 4, the sound output for outputting the voice of the other party from the speaker 5 is output to the speaker. However, the present invention is not limited to this. For example, the mode may be changed to the speaker output mode when the proximity sensor 11 detects that the user has released the mobile phone terminal 1 from the ear. Good.

  Alternatively, timing may be started by a timer when the proximity sensor 11 detects that the user has released the cellular phone terminal 1 from the ear, and the mode may be changed to the speaker output mode after a predetermined time has elapsed. However, inadvertently outputting the other party's voice from the speaker 5 may cause a problem. Therefore, as described in the second embodiment, it is effective to switch the first operation or the second operation as a trigger. Is.

  In the first and second embodiments described above, the voice processing device according to the present invention is applied to the mobile phone terminal 1, but the present invention is not limited to this, and voice calls are possible. As long as it is a communication device, it may be a basic phone, a future phone or a smartphone, and can also be widely applied to a landline phone.

  Further, the mute discrimination unit 9 shown in FIG. 1 detects the detection result of the proximity sensor 11 that detects the approach of the ear for listening to the sound output from the receiver 4 and the approach of the hand, finger, etc. for closing the sound hole. Based on the detection result of the proximity sensor 12, the mute process is simply executed and released. For this reason, the proximity sensors 11 and 12 are connected to a GPIO (General Purpose Input / Output) of a CPU (Central Processing Unit) (not shown), and the CPU executes and cancels mute processing based on the detection results of the proximity sensors 11 and 12. You may make it discriminate | determine execution. That is, the determination process for the detection results of the proximity sensors 11 and 12 can be substituted by the CPU. Since the DSP 7 can perform an acoustic operation with an amplifier unit (not shown) existing inside, the mute process or the release operation may be executed in accordance with a control signal from the CPU.

The features of the present invention will be described below.
A part or all of the above-described embodiment can be described as in the following supplementary notes, but is not limited thereto.
(Appendix 1)
FIG. 5 is a configuration diagram of Supplementary Note 1. Here, the correspondence between FIG. 5 and FIG. 1 will be described. 5 corresponds to the microphone 2 in FIG. 1, and the first approach detection unit 21 in FIG. 5 corresponds to the proximity sensor 12 in FIG. 5 corresponds to the receiver 4 in FIG. 1, the second approach detection unit 23 in FIG. 5 corresponds to the proximity sensor 11 in FIG. 1, and the control unit 24 in FIG. This corresponds to the mute determination unit 9 of FIG.

As shown in this figure, the invention described in Appendix 1 is
A voice detection unit 20;
A first approach detection unit 21 for detecting a first approach for closing a sound hole that guides the sound to the sound detection unit 20;
An audio output unit 22;
A second approach detection unit 23 for detecting a second approach for listening to the sound output from the sound output unit 22;
Based on the detection result of the first approach by the first approach detection unit 21, the voice detected by the voice detection unit 20 is controlled to be transmitted or not transmitted to the call partner, and the first approach detection is performed. Whether to continue the state in which the voice detected by the voice detection unit 20 is not transmitted based on the detection result of the first approach by the unit 21 and the detection result of the second approach by the second approach detection unit 23 And a control unit 24 for controlling whether to transmit after canceling the non-transmission state.

(Appendix 2)
The control unit controls the voice detected by the voice detection unit to be invalid while at least the first approach detection unit detects the first approach. The speech processing apparatus according to the above.

(Appendix 3)
While the control unit invalidates the voice detected by the voice detection unit, the second approach detection unit does not detect the second approach, and the first approach detection unit detects the second approach. The speech processing apparatus according to appendix 2, wherein when the approach of 1 is not detected, control is performed to continue the invalid state of the speech detected by the speech detection unit.

(Appendix 4)
While the control unit invalidates the voice detected by the voice detection unit, the second approach detection unit detects the second approach, and the first approach detection unit detects the first approach. 4. The speech processing apparatus according to appendix 2 or 3, wherein when the approach is not detected, control is performed so that the invalid state of the speech detected by the speech detection unit is canceled and validated.

(Appendix 5)
Listening to the sound output from the sound detection unit, the first approach detection unit for detecting the first approach for closing the sound hole that guides the sound to the sound detection unit, the sound output unit, and the sound output unit A second proximity detecting unit for detecting a second approach for transmitting the voice detected by the voice detecting unit based on a detection result of the first approach by the first proximity detecting unit. By the voice detection unit based on the detection result of the first approach by the first approach detection unit and the detection result of the second approach by the second approach detection unit. And a control unit that controls whether to continue the state in which the detected voice is not transmitted or to cancel the state in which the detected voice is not transmitted and to transmit the detected voice.

(Appendix 6)
The control unit performs control so that the voice detected by the voice detection unit is not transmitted to a call partner at least while the first approach detection unit detects the first approach. 5. The communication device according to 5.

(Appendix 7)
While the control unit is in a state in which the voice detected by the voice detection unit is not transmitted to the call partner, the second approach detection unit does not detect the second approach, and the first approach The communication apparatus according to appendix 6, wherein when the first approach is not detected by the detection unit, control is performed so as to continue the state in which the voice detected by the voice detection unit is not transmitted to the call partner. .

(Appendix 8)
The control unit detects the second approach by the second approach detector while the voice detected by the voice detector is not transmitted to the other party, and detects the first approach. 8. The communication apparatus according to appendix 6 or 7, wherein when the first approach is not detected by a unit, control is performed so as to cancel and transmit the state in which the voice detected by the voice detection unit is not transmitted. It is.

(Appendix 9)
A first approach detecting step for detecting a first approach for closing a sound hole for guiding sound to the sound detecting unit, and a second approach for detecting a second approach for listening to the sound output from the sound output unit. Based on the approach detection step and the detection result of the first approach, the voice detected by the voice detection unit is controlled to be invalidated or validated, and the detection result of the first approach and the first And a control step for controlling whether the invalid state of the voice detected by the voice detection unit is continued or is invalidated to be valid based on the detection result of the second approach. This is a voice control method.

(Appendix 10)
A first approach detecting function for detecting a first approach for closing a sound hole for guiding sound to the sound detecting unit in the computer, and a second approach for detecting a second approach for listening to the sound output from the sound output unit. Based on the detection result of the first approach by the second approach detection function and the first approach detection function, it is controlled whether to invalidate or validate the voice detected by the voice detection unit, Based on the detection result of the first approach by the first approach detection function and the detection result of the second approach by the second approach detection function, the invalid state of the voice detected by the voice detection unit is continued. A program for executing a control function for controlling whether to enable or cancel an invalid state.

(Appendix 11)
An output level at which the voice of the other party is output at an output level at which the voice can be heard without approaching, and an operation detection unit for detecting the first operation;
If the first operation is detected by the operation detection unit while the second approach is not detected by the second proximity detection unit, the control unit detects the sound detected by the sound detection unit. The communication device according to any one of appendices 5 to 8, wherein the communication device is controlled so as to continue the state in which the communication partner is not transmitted to the other party and the other partner's voice output unit is controlled to output the other party's voice. is there.

(Appendix 12)
The operation detection unit detects a second operation different from the first operation, and the control unit performs the operation while the second approach detection unit does not detect the second approach. When the second operation is detected by the detection unit, control is performed such that the state in which the voice detected by the voice detection unit is not transmitted is canceled and the voice is transmitted from the voice output unit. The communication device according to appendix 11, wherein the communication device is controlled so as to output.

DESCRIPTION OF SYMBOLS 1 Mobile phone terminal 2 Microphone 3 Microphone amplifier 4 Receiver 5 Speaker 6 Speaker amplifier 7 DSP
8 ADC
9 Mute discrimination part 10 Codec part 11, 12 Proximity sensor

Claims (10)

  1. A voice detector;
    A first approach detecting unit for detecting a first approach for closing a sound hole for guiding sound to the sound detecting unit;
    An audio output unit;
    A second approach detection unit for detecting a second approach for listening to the sound output from the sound output unit;
    Based on the detection result of the first approach by the first approach detection unit, the voice detected by the voice detection unit is controlled to be invalidated or validated, and the first approach detection unit performs Based on the detection result of the first approach and the detection result of the second approach by the second approach detection unit, the invalid state of the voice detected by the voice detection unit is continued or the invalid state is canceled and enabled A speech processing apparatus comprising: a control unit that controls whether or not.
  2. The controller is
    The voice according to claim 1, wherein the voice detected by the voice detector is controlled to be invalid while at least the first approach is detected by the first approach detector. Processing equipment.
  3. The controller is
    While the voice detected by the voice detector is invalid, the second approach is not detected by the second approach detector, and the first approach is detected by the first approach detector. The voice processing device according to claim 2, wherein, if not, control is performed to continue the invalid state of the voice detected by the voice detection unit.
  4. The controller is
    While the voice detected by the voice detection unit is invalidated, the second approach detection unit detects the second approach, and the first approach detection unit does not detect the first approach. In such a case, the voice processing device according to claim 2 or 3, wherein the voice processing device is controlled so as to cancel and invalidate the voice detected by the voice detector.
  5. A voice detector;
    A first approach detecting unit for detecting a first approach for closing a sound hole for guiding sound to the sound detecting unit;
    An audio output unit;
    A second proximity detector for detecting a second approach for listening to the voice output from the voice output unit;
    Based on the detection result of the first approach by the first approach detection unit, the voice detected by the voice detection unit is controlled to be transmitted or not transmitted to the call partner, and the first approach detection unit Based on the detection result of the first approach and the detection result of the second approach by the second approach detection unit, the state of not transmitting the voice detected by the voice detection unit is continued or the state of not transmitting And a control unit that controls whether to release and transmit the communication device.
  6. The controller is
    6. The control according to claim 5, wherein the voice detected by the voice detection unit is controlled not to be transmitted to a call partner while at least the first approach detection unit is detecting the first approach. Communication device.
  7. The controller is
    While the voice detected by the voice detector is not transmitted to the other party, the second approach detector does not detect the second approach, and the first approach detector detects the second approach. The communication apparatus according to claim 6, wherein when the approach of 1 is not detected, control is performed so as to continue a state in which the voice detected by the voice detection unit is not transmitted to the call partner.
  8. The controller is
    While the voice detected by the voice detection unit is not transmitted to the other party, the second approach detection unit detects the second approach, and the first approach detection unit detects the first approach. 8. The communication device according to claim 6, wherein when the approach of the voice is not detected, control is performed such that the state in which the voice detected by the voice detection unit is not transmitted is canceled and transmitted.
  9. A first approach detecting step for detecting a first approach for closing a sound hole for guiding sound to the sound detecting unit;
    A second approach detection step of detecting a second approach for listening to the sound output from the sound output unit;
    Based on the detection result of the first approach, the voice detected by the voice detection unit is controlled to be invalid or valid, and the detection result of the first approach and the detection of the second approach are controlled. And a control step of controlling whether to invalidate the voice detected by the voice detection unit based on the result or to cancel the invalid state and validate the voice.
  10. On the computer,
    A first approach detection function for detecting a first approach for closing a sound hole for guiding sound to the sound detection unit;
    A second approach detection function for detecting a second approach for listening to the voice output from the voice output unit;
    Based on the detection result of the first approach by the first approach detection function, it is controlled whether to invalidate or validate the voice detected by the voice detection unit, and by the first approach detection function Based on the detection result of the first approach and the detection result of the second approach by the second approach detection function, the invalid state of the voice detected by the voice detection unit is continued or the invalid state is canceled. A program for executing a control function for controlling whether or not to be effective.

JP2012239841A 2012-10-31 2012-10-31 Voice processing device, communication device, voice control method, and program Pending JP2014090350A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2012239841A JP2014090350A (en) 2012-10-31 2012-10-31 Voice processing device, communication device, voice control method, and program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2012239841A JP2014090350A (en) 2012-10-31 2012-10-31 Voice processing device, communication device, voice control method, and program

Publications (1)

Publication Number Publication Date
JP2014090350A true JP2014090350A (en) 2014-05-15

Family

ID=50791931

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2012239841A Pending JP2014090350A (en) 2012-10-31 2012-10-31 Voice processing device, communication device, voice control method, and program

Country Status (1)

Country Link
JP (1) JP2014090350A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016029466A (en) * 2014-07-16 2016-03-03 パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America Control method of voice recognition and text creation system and control method of portable terminal
JP2016201815A (en) * 2016-07-01 2016-12-01 Kddi株式会社 Communication terminal device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016029466A (en) * 2014-07-16 2016-03-03 パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America Control method of voice recognition and text creation system and control method of portable terminal
JP2016201815A (en) * 2016-07-01 2016-12-01 Kddi株式会社 Communication terminal device

Similar Documents

Publication Publication Date Title
CN1147187C (en) Communication apparatus influenced by accelerometer
US8872775B2 (en) Information processing apparatus
US9058801B2 (en) Robust process for managing filter coefficients in adaptive noise canceling systems
JP4919531B2 (en) Communication device with touch-sensitive screen
CN101325756B (en) Apparatus for identifying mobile phone voice and method for activating mobile phone voice identification
KR20100136649A (en) Method for embodying user interface using a proximity sensor in potable terminal and apparatus thereof
JP4530051B2 (en) Audio signal transmitter / receiver
US20110144984A1 (en) Voice coder with two microphone system and strategic microphone placement to deter obstruction for a digital communication device
JP5499633B2 (en) Reproduction device, headphone, and reproduction method
US6233462B1 (en) Portable terminal device for automatically controlling transmitting voice level and calling sound level
US8488411B2 (en) Microphone proximity detection
US20160019886A1 (en) Method and apparatus for recognizing whisper
US10425717B2 (en) Awareness intelligence headphone
US8666750B2 (en) Voice control system
JP5567914B2 (en) Mobile terminal device
US7010098B2 (en) Ultrasonic proximity detector for a telephone device
JP3699018B2 (en) Automatic volume control apparatus of the phone
JP5432185B2 (en) Portable device, method for operating portable device, and computer program
US20100040245A1 (en) Multi-function headset and function selection of same
GB2464642A (en) Automatic speakerphone/handset mode selection in a flip phone
EP1931119A1 (en) Portable terminal device and control method thereof
JP4247002B2 (en) Speaker distance detection apparatus and method using microphone array, and voice input / output apparatus using the apparatus
JP2007019898A (en) Portable telephone
JP2004517563A (en) Based on sound used in the mobile telephone device proximity detector
US7761106B2 (en) Voice coder with two microphone system and strategic microphone placement to deter obstruction for a digital communication device