WO2015102040A1 - 音声処理装置、音声処理システム、音声処理方法、音声処理用のプログラム製品 - Google Patents

音声処理装置、音声処理システム、音声処理方法、音声処理用のプログラム製品 Download PDF

Info

Publication number
WO2015102040A1
WO2015102040A1 PCT/JP2014/006172 JP2014006172W WO2015102040A1 WO 2015102040 A1 WO2015102040 A1 WO 2015102040A1 JP 2014006172 W JP2014006172 W JP 2014006172W WO 2015102040 A1 WO2015102040 A1 WO 2015102040A1
Authority
WO
WIPO (PCT)
Prior art keywords
voice
call
processing
application
voice processing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/JP2014/006172
Other languages
English (en)
French (fr)
Japanese (ja)
Inventor
伊藤 正也
義隆 尾崎
圭作 林
拡基 鵜飼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Denso Corp
Original Assignee
Denso Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Denso Corp filed Critical Denso Corp
Priority to US15/108,739 priority Critical patent/US20160329060A1/en
Publication of WO2015102040A1 publication Critical patent/WO2015102040A1/ja
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/60Substation equipment, e.g. for use by subscribers including speech amplifiers
    • H04M1/6033Substation equipment, e.g. for use by subscribers including speech amplifiers for providing handsfree use or a loudspeaker mode in telephone sets
    • H04M1/6041Portable telephones adapted for handsfree use
    • H04M1/6075Portable telephones adapted for handsfree use adapted for handsfree use in a vehicle
    • H04M1/6083Portable telephones adapted for handsfree use adapted for handsfree use in a vehicle by interfacing with the vehicle audio system
    • H04M1/6091Portable telephones adapted for handsfree use adapted for handsfree use in a vehicle by interfacing with the vehicle audio system including a wireless interface
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • G10L19/22Mode decision, i.e. based on audio signal content versus external parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0316Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0316Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
    • G10L21/0364Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/72412User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72442User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for playing music files
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72445User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting Internet browser applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/40Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2207/00Type of exchange or network, i.e. telephonic medium, in which the telephonic communication takes place
    • H04M2207/18Type of exchange or network, i.e. telephonic medium, in which the telephonic communication takes place wireless networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/02Details of telephonic subscriber devices including a Bluetooth interface
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/74Details of telephonic subscriber devices with voice recognition means

Definitions

  • the present disclosure relates to a speech processing apparatus (Speech Processing) Apparatus, a speech processing system, a speech processing method, and a program product for speech processing.
  • Speech Processing Speech Processing
  • the vehicle device transmits the acquired voice data to an external center server via the mobile terminal. Then, the center server performs voice recognition based on the acquired voice data, and returns a search result corresponding to the voice to the vehicle device.
  • the vehicle device transmits either voice data to the mobile terminal when performing a hands-free call, or transmits voice data to the mobile terminal when performing a search using voice recognition. Even in this case, the audio data is subjected to the same audio processing, specifically, the same noise cancellation processing, echo cancellation processing, gain control processing, and the like.
  • the optimum voice processing for a call and the optimum voice processing for voice recognition are different.
  • voice processing is performed to narrow down to a sound having a frequency that can be heard by human ears. If processing similar to this voice processing is performed in voice recognition, a voice waveform necessary for voice recognition is used. Will be distorted and the recognition rate will decrease.
  • An object of the present disclosure is to provide a voice processing device that can optimally perform voice processing for calls and voice processing for other than calls, a voice processing system that includes the voice processing device, and the voice processing device. And a program product for voice processing to be executed by being incorporated in the voice processing apparatus.
  • the acquired audio data is transmitted to an external mobile terminal, predetermined audio processing is performed on the audio data to be transmitted.
  • the voice processing it is possible to switch between the first voice processing for calling and the second voice processing for other than calling. Therefore, according to the application to be executed, the first voice processing for the call and the second voice processing for the other than the call can be appropriately switched and executed. Any of the second sound processing can be optimally performed.
  • FIG. 1 shows schematically the structural example of the speech processing system which concerns on one Embodiment.
  • the figure which shows the structural example of an audio processing apparatus roughly
  • the figure which shows the structural example of a portable terminal roughly
  • the flowchart which shows an example of the control content in the case of performing a telephone call application The figure which shows the state which an audio processing apparatus and a portable terminal mutually cooperate, and execute an application
  • the flowchart which shows an example of the control content in the case of performing a voice recognition search application Schematic configuration diagram of a voice processing system showing a modification of the present embodiment (part 1) Schematic configuration diagram of a voice processing system showing a modification of the present embodiment (No. 2) Schematic configuration diagram of a voice processing system showing a modification of the present embodiment (No. 3) Schematic configuration diagram of a voice processing system showing a modification of the present embodiment (No. 4)
  • the voice processing system 10 is constructed by a voice processing device 11 and a portable terminal 12.
  • the voice processing device 11 is composed of a navigation device mounted on a vehicle, for example.
  • the voice application 11 is equipped with a call application A.
  • the call application A is an application for realizing a so-called hands-free call function (also referred to as a hands-free telephone conversation function) that enables a user to call (phone conversation) without holding the mobile terminal 12 in his / her hand. is there.
  • the mobile terminal 12 is, for example, a mobile communication terminal owned by a passenger of the vehicle.
  • the audio processing device is compliant with the Bluetooth communication standard (Bluetooth: registered trademark) which is an example of the short-range wireless communication standard. 11 is communicably connected.
  • Bluetooth communication standard Bluetooth: registered trademark
  • the voice processing device 11 and the mobile terminal 12 are configured to acquire various applications distributed from the distribution center 14 by connecting to the external distribution center 14 via the communication network 100.
  • the distribution center 14 uses, for example, a voice recognition search application B that implements a search service using speech recognition (Speech Recognition) for recognizing speech uttered by a user, and Internet radio.
  • speech recognition Speech Recognition
  • Various applications such as an application to be realized and an application to realize a music distribution service are stored.
  • an application distribution request is received from an external terminal or device, the application is distributed to the request source via the communication network 100. .
  • the application distributed from the distribution center 14 includes various data necessary for executing the application.
  • the voice processing device 11 and the portable terminal 12 are configured to be connectable to a voice recognition search server 15 (hereinafter also referred to as a search server 15) via the communication network 100.
  • the voice recognition search server 15 stores well-known dictionary data necessary for voice recognition processing, search processing data necessary for search processing, and the like.
  • the search processing data includes data such as names and places of stores and facilities existing on the map.
  • the voice processing device 11 includes a control circuit 21, a communication connection unit 22, a storage unit 23, a voice input / output unit 24, a display output unit 25, an operation input unit 26, and the like.
  • the control circuit 21 is composed of a well-known microcomputer having a CPU, RAM, ROM, I / O bus and the like (not shown).
  • the control circuit 21 controls the overall operation of the sound processing apparatus 11 according to various computer programs stored in the ROM or the storage unit 23.
  • the control circuit 21 executes a sound processing program that is a computer program, so that the sound data acquisition processing unit 31, the sound data transmission processing unit 32, and the sound processing unit 33 are virtualized by software. Realize.
  • some or all of the functions can be provided as hardware components in each processing unit.
  • the communication connection unit 22 is configured by a wireless communication module, for example, and establishes a wireless communication line with the communication connection unit 42 included in the mobile terminal 12 and performs various communication with the mobile terminal 12 through the wireless communication line. I do.
  • the communication connection unit 22 includes various communication protocols such as a hands-free call profile (HFP: Hands Free Profile) and a data communication profile.
  • HFP hands-free call profile
  • the storage unit 23 is configured by a computer-readable non-transitional non-volatile storage medium such as a hard disk drive, for example, and executes applications in cooperation with various computer programs and application programs and external devices and terminals.
  • Various programs also referred to as program products including instructions
  • a cooperative application such as a cooperative application that realizes a cooperative function to be performed, and various data used in each program are stored.
  • the storage unit 23 stores various data necessary for speech recognition processing such as well-known dictionary data for speech recognition of the acquired speech data. Therefore, the voice processing apparatus 11 can perform voice recognition processing independently without depending on the voice recognition search server 15.
  • the voice input / output unit 24 is connected to a microphone and a speaker (not shown) and has a known voice input function and voice output function.
  • the voice input / output unit 24 sends voice data corresponding to the voice input from the microphone to the portable terminal 12. It is possible to output sound from the speaker based on the sound data transmitted and received from the mobile terminal 12. Thereby, the voice processing device 11 can realize a so-called hands-free call in cooperation with the portable terminal 12.
  • the display output unit 25 is configured by, for example, a liquid crystal display or an organic EL, and displays various information based on a display command signal from the control circuit 21.
  • the screen of the display output unit 25 is provided with a touch panel switch configured by a well-known pressure-sensitive method, electromagnetic induction method, capacitance method, or a combination thereof.
  • the display output unit 25 displays various screens such as an input interface such as an operation input screen for inputting an operation for the application and an output interface such as an output screen for outputting the execution contents and execution results of the application. .
  • the operation input unit 26 includes various switches such as a touch panel switch provided on the screen of the display output unit 25 and a mechanical switch provided around the display output unit 25.
  • the operation input unit 26 outputs operation detection signals to the control circuit 21 in accordance with various switch operations by the user.
  • the control circuit 21 analyzes the operation detection signal input from the operation input unit 26 to specify the operation content of the user, and executes various processes based on the specified operation content.
  • the voice processing device 11 includes a known position specifying unit for specifying the current position of the voice processing device 11 based on satellite radio waves received from a positioning satellite (not shown).
  • the voice data acquisition processing unit 31 is also referred to as a voice data acquisition unit / device / means.
  • voice data acquisition processing unit 31 When voice is input from the microphone of the voice input / output unit 24, the voice data acquisition processing unit 31 generates voice data corresponding to the acquired voice.
  • the voice data transmission processing unit 32 is also referred to as a voice data transmission unit / device / means, and the voice data acquired by the voice data acquisition processing unit 31 is transferred to an external portable terminal via a communication line established by the communication connection unit 22. 12 to send.
  • the voice data transmission processing unit 32 is configured to transmit both the voice data for a call and the voice data for a non-call using the same communication protocol.
  • a Bluetooth communication standard hands-free call profile HFP
  • the communication protocol that can be adopted is not limited to this.
  • the voice processing unit 33 is also referred to as a voice processing device / means, and performs predetermined voice processing on the voice data transmitted by the voice data transmission processing unit 32.
  • the voice processing unit 33 is also referred to as voice processing for calls (also referred to as first voice processing) and voice processing for other than calls (second voice processing) as voice processing. ), which is an example of the voice recognition search for voice recognition search.
  • the voice processing for a call is a process for narrowing down to only a sound having a frequency that can be heard by a human ear, and includes a noise cancellation process for a call, an echo cancellation process for a call, a gain control process for a call, and the like. .
  • the voice processing for voice recognition search is a process for narrowing down the sound to such an extent that the voice can be recognized including, for example, a sound of a frequency that can be heard by the human ear.
  • sounds other than the frequency that can be heard by the human ear remain to some extent without being canceled.
  • voice processing for a call can perform more reliable noise cancellation, echo cancellation, and gain control on voice data than voice processing for voice recognition search.
  • speech processing for speech recognition search since it is desired to pick up raw speech as close as possible to the speech uttered by the user, relatively loose noise cancellation, echo cancellation, and gain control are applied to the speech data. That is, in speech processing for speech recognition search, it is required to prevent the original speech information (speech waveform) from changing as much as possible.
  • the medium frequency that is audible to the human ear is reduced for each frequency band included in the audio data by reducing the gain in the high and low frequency bands that are difficult for the human ear to hear.
  • a process for amplifying the band is performed.
  • the original speech waveform is distorted and is not suitable for speech recognition.
  • the waveform (frequency) of the speech waveform differs for each vowel or consonant, if the original speech waveform is corrupted, speech recognition becomes extremely difficult.
  • gain control in speech processing for speech recognition for example, by changing the setting value (parameter) of the high frequency band and low frequency band where the gain is reduced, or by appropriately adjusting how the gain is reduced, It is preferable to perform a voice process in which the voice waveform remains in a state close to the original shape, rather than a process in which a voice waveform as close as possible to the waveform remains, that is, voice processing for a call.
  • the mobile terminal 12 includes a control circuit 41, a communication connection unit 42, a storage unit 43, a voice input / output unit 44, a display output unit 45, an operation input unit 46, a telephone communication unit 47, and the like.
  • the control circuit 41 is constituted by a well-known microcomputer having a CPU, RAM, ROM, I / O bus, etc. (not shown).
  • the control circuit 41 controls the overall operation of the mobile terminal 12 according to a computer program stored in the ROM or the storage unit 43.
  • some or all of the functions of the control circuit 41 can be executed by hardware components.
  • the communication connection unit 42 is configured by a wireless communication module, for example, and establishes a wireless communication line with the communication connection unit 22 included in the voice processing device 11, and performs various types of communication with the voice processing device 11 through the wireless communication line. To communicate.
  • the communication connection unit 42 includes various communication protocols such as a hands-free call profile (HFP) and a data communication profile.
  • the storage unit 43 is configured by a computer-readable non-transitional non-volatile storage medium such as a memory card, and executes applications in cooperation with various computer programs and application programs and external devices and terminals.
  • Various programs also referred to as program products including instructions
  • a cooperative application that realizes a cooperative function to be performed, and various data used in each program are stored.
  • the voice input / output unit 44 is connected to a microphone and a speaker (not shown) and has a known voice input function and voice output function.
  • the voice input / output unit 44 is connected to the portable terminal 12 when the voice application 11 is started in the voice processing apparatus 11 in a state where the voice processing apparatus 11 is communicably connected to the portable terminal 12.
  • the voice data corresponding to the voice input from the voice processor 11 is transmitted to the voice processor 11, and the voice data received from the voice processor 11 can be transmitted to the mobile terminal of the other party. Thereby, the portable terminal 12 can realize a so-called hands-free call in cooperation with the voice processing device 11.
  • the voice input / output unit 44 outputs the uttered voice input from the microphone to the control circuit 41 while the voice processing device 11 is not connected to the portable terminal 12 so as to be communicable.
  • the received voice is output from the speaker.
  • the portable terminal 12 can implement
  • the display output unit 45 is composed of, for example, a liquid crystal display or an organic EL display, and displays various information based on a display command signal from the control circuit 41.
  • the screen of the display output unit 45 is provided with a touch panel switch configured by a known pressure-sensitive method, electromagnetic induction method, electrostatic capacitance method, or a combination thereof.
  • the display output unit 45 displays various screens such as an input interface such as an operation input screen for inputting an operation for the application and an output interface such as an output screen for outputting the execution contents and execution results of the application. .
  • the operation input unit 46 includes various switches such as a touch panel switch provided on the screen of the display output unit 45 and a mechanical switch provided around the display output unit 45.
  • the operation input unit 46 outputs an operation detection signal to the control circuit 41 according to the operation of various switches by the user.
  • the control circuit 41 analyzes the operation detection signal input from the operation input unit 46, specifies the operation content of the user, and executes various processes based on the specified operation content.
  • the telephone communication unit 47 establishes a wireless telephone communication line with the communication network 100, and executes telephone communication via this telephone communication line.
  • the communication network 100 includes facilities for providing a mobile phone communication service using a public network such as a mobile phone base station and a base station controller (not shown).
  • the control circuit 41 is connected to the distribution center 14 or the voice recognition search server 15 connected to the communication network 100 through the telephone communication unit 47 so as to be communicable.
  • each section is expressed as, for example, A1, B1, C1, D1, E1.
  • each section can be divided into a plurality of subsections, while a plurality of sections can be combined into one section.
  • each section can be referred to as a device, module, or means.
  • each of the above sections or a combination thereof includes not only (i) a section of software combined with a hardware unit (eg, a computer), but also (ii) hardware (eg, an integrated circuit, As a section of (wiring logic circuit), it can be realized with or without the function of related devices.
  • the hardware section can be included inside the microcomputer.
  • the voice processing device 11 determines whether or not the call application A is activated in the voice processing device 11 (A1) and whether an incoming operation is input from the external portable terminal 12 (A2). ).
  • the voice processing device 11 monitors whether the user inputs a call operation via the call application A (A3).
  • the call operation is an example of a voluntary operation in the call application A and refers to making a call to an external mobile terminal.
  • the voice processing device 11 shifts from the normal mode to the hands-free call mode (A4).
  • the voice processing device 11 activates the call application A (A5). Then, the voice processing device 11 shifts from the normal mode to the hands-free call mode (A4).
  • the incoming call operation is an example of a different operation in the call application A, and means receiving an incoming call from an external portable terminal.
  • the mobile terminal 12 is set to input an incoming call operation to the voice processing device 11 when there is an incoming call from an external mobile terminal and the mobile terminal 12 is in the hands-free call mode.
  • the voice processing device 11 establishes a wireless communication line by HFP with the portable terminal 12, transmits voice data corresponding to the voice input from the microphone to the portable terminal 12, Based on the audio data received from the portable terminal 12, the audio can be output from the speaker.
  • the mobile terminal 12 when the mobile terminal 12 receives an incoming call from an external mobile terminal (not shown) (B1: YES), the mobile terminal 12 confirms whether or not a wireless communication line using HFP is established with the voice processing device 11 (B2). .
  • the portable terminal 12 performs a call by the portable terminal 12 alone in the normal call mode (B3). . That is, a normal call is performed between the mobile terminal 12 and the mobile terminal of the other party.
  • the mobile terminal 12 shifts from the normal call mode to the hands-free call mode (B4).
  • the portable terminal 12 uses the HFP wireless communication line established between the portable terminal 12 and the voice data corresponding to the voice input from the other party's portable terminal (not shown). Is transmitted to the voice processing device 11, and the voice data received from the voice processing device 11 can be transmitted to the mobile terminal of the other party.
  • the voice processing system 10 is in a state where a so-called hands-free call is possible when both the voice processing device 11 and the mobile terminal 12 are shifted to the hands-free call mode.
  • the voice data acquisition processing unit 31 acquires voice data (A6), and the acquired voice data is subjected to voice processing for calling by the voice processing unit 33 ( A7).
  • the voice processing device 11 detects a spontaneous operation or a spontaneous operation of the call application A, and thereby confirms that the application being executed is the call application A. Therefore, the voice processing device 11 switches the voice processing applied to the voice data to voice processing for a call. Then, the voice processing device 11 transmits the voice data subjected to voice processing for a call to the mobile terminal 12 (A8).
  • the process of step A6 is an example of an audio data acquisition step
  • the process of step A7 is an example of an audio process step
  • the process of step A8 is an example of an audio data transmission step.
  • the portable terminal 12 transmits the voice data received from the voice processing device 11 to the portable terminal of the call partner (B5).
  • the mobile terminal 12 receives voice data from the mobile terminal of the call partner (B6)
  • the mobile terminal 12 transmits the voice data to the voice processing device 11 (B7).
  • the audio processing device 11 receives audio data from the mobile terminal 12, the audio processing device 11 outputs audio from the speaker based on the audio data (A9).
  • the voice received from the mobile terminal of the call partner is output from the voice processing device 11.
  • the voice data of the uttered voice and the voice data of the received voice are appropriately transmitted and received between the voice processing device 11 and the mobile terminal of the other party of the call by relaying the mobile terminal 12, thereby realizing a so-called hands-free call Is done.
  • the voice data transmitted from the voice processing device 11 to the portable terminal 12 is included in the call data. Is processed. This hands-free call is continued until the call is terminated at the voice processing device 11 or the mobile terminal of the other party.
  • search application B the voice recognition search application B
  • search application B the voice recognition search application B
  • the portable terminal 12 when the portable terminal 12 is communicably connected to the voice processing device 11 and the cooperative application is activated in each of the voice processing device 11 and the portable terminal 12, the portable terminal 12
  • the execution process of the voice recognition search application B is executed by the mobile terminal 12, and the input interface and the output interface of the voice recognition search application B are provided by the voice processing device 11.
  • the voice processing device 11 starts the application that the portable terminal 12 has.
  • a button is displayed (C2). This activation button is an example of an input interface.
  • the voice processing device 11 transmits a start command signal for the voice recognition search application B to the mobile terminal 12 (C4). At this time, the voice processing device 11 also transmits to the portable terminal 12 current position information indicating the current position of the voice processing device 11 obtained by the position specifying unit.
  • the portable terminal 12 When the portable terminal 12 receives the activation command signal for the voice recognition search application B, the portable terminal 12 starts the voice recognition search application B (D2). And the portable terminal 12 transmits the starting completion signal which shows having started the speech recognition search application B to the speech recognition search server 15 (D3). At this time, the portable terminal 12 also transmits the current position information received from the voice processing device 11 to the voice recognition search server 15.
  • the voice recognition search server 15 When the voice recognition search server 15 receives the activation completion signal of the voice recognition search application B, the voice recognition search server 15 transmits voice data for collecting search conditions to the mobile terminal 12 (E1). In this case, for example, message data such as “Please say business” is set as voice data for collecting search conditions.
  • the portable terminal 12 transmits the search condition collection voice data received from the voice recognition search server 15 to the voice processing device 11 (D4).
  • the voice processing device 11 Upon receiving the search condition collection voice data, the voice processing device 11 outputs the search condition collection voice from the speaker based on the voice data (C5). In this case, for example, a guidance voice such as “Please say your business” is output.
  • the voice processing device 11 acquires the voice data by the voice data acquisition processing unit 31 (C6), and the acquired voice data.
  • the voice processing unit 33 performs voice processing for voice recognition search (C7). In this case, the voice processing device 11 has not detected a spontaneous operation or a spontaneous operation of the call application A, and thereby confirms that the application being executed is an application other than the call application A. ing.
  • the voice processing device 11 switches the voice processing applied to the voice data to voice processing for voice recognition search, which is an example of voice processing for other than calls. Then, the voice processing device 11 transmits the voice data subjected to voice processing for voice recognition search to the mobile terminal 12 (C8).
  • the process of step C6 is an example of an audio data acquisition step
  • the process of step C7 is an example of an audio processing step
  • the process of step C8 is an example of an audio data transmission step.
  • application specifying data for specifying an application being executed is transmitted from the mobile terminal 12 to the sound processing device 11, and the sound processing device 11 performs sound processing suitable for the application specified by the application specifying data. It may be configured to execute by switching.
  • the portable terminal 12 transmits the voice data received from the voice processing device 11 to the voice recognition search server 15 (D5).
  • the voice recognition search server 15 performs a known voice recognition process based on the voice data (E2).
  • the voice recognition search server 15 executes a well-known search process based on the recognized voice and the position information of the voice processing device 11 (E3), and transmits search result data indicating the search result to the portable terminal 12 (E4).
  • the voice recognition search server 15 also transmits voice data for search result output to the mobile terminal 12.
  • message data such as “Display a nearby Italian store” is set as the audio data for outputting the search result. That is, the voice recognition search server 15 reflects the search condition such as “Italian” in the voice data for outputting the search result.
  • the portable terminal 12 transmits the search result data received from the voice recognition search server 15 to the voice processing device 11 (D6). At this time, the portable terminal 12 also transmits the search result output voice data received from the voice recognition search server 15 to the voice processing device 11.
  • the audio processing device 11 when receiving the audio data for search result output, the audio processing device 11 outputs audio from the speaker based on the audio data (C9). In this case, for example, a guidance voice such as “Display a nearby Italian store” is output.
  • the speech processing apparatus 11 displays the search result based on the search result data (C10).
  • the search result output voice and the search result display screen are examples of an output interface.
  • a search service using voice recognition is realized by appropriately transmitting and receiving voice data and search result data between the voice processing device 11 and the voice recognition search server 15 via the portable terminal 12.
  • the voice processing device 11 does not detect the spontaneous operation or the other operation of the call application A. Therefore, the voice data transmitted from the voice processing device 11 to the portable terminal 12 is used for voice recognition. Is processed.
  • the audio processing device 11 when transmitting the acquired audio data to the external mobile terminal 12, the audio processing device 11 performs predetermined audio processing on the audio data to be transmitted.
  • voice processing it is possible to switch between voice processing for calling, which is an example of voice processing for calling, and voice processing for searching for voice recognition, which is an example of voice processing for other than calling. . Therefore, it is possible to appropriately switch and execute a voice process for a call and a voice process for a call other than the call according to a running application, and optimally perform both the voice process for a call and the voice process for a call. be able to.
  • the audio processing performed on the audio data may be configured such that noise cancellation processing, echo cancellation processing, auto gain control processing for gradually increasing the aperture of the noise cancellation processing, and the like are performed in a single shot.
  • the processes may be implemented by appropriately combining the processes.
  • the voice processing device 11 executes voice processing for a call when detecting a spontaneous operation or a spontaneous operation in the call application A. That is, based on whether an operation specific to the call application A, in other words, an operation that cannot be generated by an application other than the call application A is detected, the voice processing applied to the voice data is switched to voice processing for calling. Therefore, when the call application A is executed, the voice processing for the call can be surely executed. In addition, when an application other than the call application A is executed, it is possible to reliably execute voice processing for other than calls.
  • both voice data for call and voice data for voice recognition that is voice data for other than calls are configured to be transmitted and received by the same communication protocol.
  • audio data related to the application can be transmitted and received using the same protocol.
  • the call application may be configured to be executed on a mobile terminal.
  • the voice recognition search application may be configured to be executed by a voice processing device.
  • the voice processing device 11 more specifically, the voice processing unit 33 is configured not to execute voice processing when an application other than the call application is activated. Instead, the mobile terminal 12 or the voice recognition search server You may comprise so that 15 may perform an audio
  • the voice processing system 10 does not perform voice processing for voice recognition in the voice processing device 11, in other words, does not perform signal processing of voice data. You may comprise so that signal processing may be performed.
  • the speech processing system 10 does not perform speech recognition signal processing in the speech processing device 11 and the portable terminal 12, and the speech recognition search server 15 performs signal processing for speech recognition. May be configured to execute.
  • the voice processing system 10 includes a call app in both the voice processing device 11 and the mobile terminal 12, and the voice processing device 11 performs voice processing for calling on voice data for calling.
  • the portable terminal 12 may be configured not to perform voice processing for calls on voice data for calls, or to perform additional voice processing.
  • the voice processing system 10 is configured such that the voice processing device 11 does not perform voice processing for calls on voice data for calls or performs additional voice processing, and the mobile terminal 12 The voice processing for calling may be applied to the voice data for calling.
  • the voice processing system 10 includes a mobile terminal 12 that includes a voice recognition search application ⁇ corresponding to the voice recognition search server ⁇ and a voice recognition search application ⁇ corresponding to the voice recognition search server ⁇ . It is good.
  • the voice recognition search server does not perform voice processing for voice recognition on the voice data for voice recognition in the mobile terminal 12.
  • a configuration may be adopted in which speech processing for speech recognition is performed on speech data for speech recognition at ⁇ .
  • the portable terminal 12 When the search service of the voice recognition search server ⁇ is used by the voice recognition search application ⁇ , the portable terminal 12 performs voice processing for voice recognition on the voice data for voice recognition, and the voice recognition search server ⁇ Then, it is good also as a structure which does not perform the speech process for speech recognition with respect to the speech data for speech recognition. That is, the voice processing system 10 may be configured to appropriately change the execution subject that performs voice processing for voice recognition on voice data according to the type of voice recognition search application to be used.
  • Applications other than the call application may be any application that implements a service that requires voice recognition processing, and are not limited to the voice recognition search application.
  • the voice processing device 11 may be configured by a device in which an application program having a navigation function is installed, for example. Moreover, the voice processing device 11 may be configured by an in-vehicle device incorporated in a vehicle, or may be configured by a portable wireless device that can be attached to and detached from the vehicle.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Quality & Reliability (AREA)
  • Telephone Function (AREA)
  • Telephonic Communication Services (AREA)
PCT/JP2014/006172 2014-01-06 2014-12-11 音声処理装置、音声処理システム、音声処理方法、音声処理用のプログラム製品 Ceased WO2015102040A1 (ja)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/108,739 US20160329060A1 (en) 2014-01-06 2014-12-11 Speech processing apparatus, speech processing system, speech processing method, and program product for speech processing

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2014000285A JP6318621B2 (ja) 2014-01-06 2014-01-06 音声処理装置、音声処理システム、音声処理方法、音声処理プログラム
JP2014-000285 2014-01-06

Publications (1)

Publication Number Publication Date
WO2015102040A1 true WO2015102040A1 (ja) 2015-07-09

Family

ID=53493389

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/006172 Ceased WO2015102040A1 (ja) 2014-01-06 2014-12-11 音声処理装置、音声処理システム、音声処理方法、音声処理用のプログラム製品

Country Status (3)

Country Link
US (1) US20160329060A1 (enExample)
JP (1) JP6318621B2 (enExample)
WO (1) WO2015102040A1 (enExample)

Families Citing this family (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
EP4138075B1 (en) 2013-02-07 2025-06-11 Apple Inc. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
KR101959188B1 (ko) 2013-06-09 2019-07-02 애플 인크. 디지털 어시스턴트의 둘 이상의 인스턴스들에 걸친 대화 지속성을 가능하게 하기 위한 디바이스, 방법 및 그래픽 사용자 인터페이스
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
CN104301064B (zh) 2013-07-16 2018-05-04 华为技术有限公司 处理丢失帧的方法和解码器
WO2015020942A1 (en) 2013-08-06 2015-02-12 Apple Inc. Auto-activating smart responses based on activities from remote devices
WO2015184186A1 (en) 2014-05-30 2015-12-03 Apple Inc. Multi-command single utterance input method
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
CN105225666B (zh) 2014-06-25 2016-12-28 华为技术有限公司 处理丢失帧的方法和装置
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10200824B2 (en) 2015-05-27 2019-02-05 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10331312B2 (en) 2015-09-08 2019-06-25 Apple Inc. Intelligent automated assistant in a media environment
US10740384B2 (en) 2015-09-08 2020-08-11 Apple Inc. Intelligent automated assistant for media search and playback
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US12223282B2 (en) 2016-06-09 2025-02-11 Apple Inc. Intelligent automated assistant in a home environment
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US12197817B2 (en) 2016-06-11 2025-01-14 Apple Inc. Intelligent device arbitration and control
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
DK180048B1 (en) 2017-05-11 2020-02-04 Apple Inc. MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION
DK201770429A1 (en) 2017-05-12 2018-12-14 Apple Inc. LOW-LATENCY INTELLIGENT AUTOMATED ASSISTANT
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK201770411A1 (en) 2017-05-15 2018-12-20 Apple Inc. Multi-modal interfaces
US20180336892A1 (en) 2017-05-16 2018-11-22 Apple Inc. Detecting a trigger of a digital assistant
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. FAR-FIELD EXTENSION FOR DIGITAL ASSISTANT SERVICES
US20180336275A1 (en) 2017-05-16 2018-11-22 Apple Inc. Intelligent automated assistant for media exploration
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
DK179822B1 (da) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. VIRTUAL ASSISTANT OPERATION IN MULTI-DEVICE ENVIRONMENTS
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
DK201970511A1 (en) 2019-05-31 2021-02-15 Apple Inc Voice identification in digital assistant systems
US11227599B2 (en) 2019-06-01 2022-01-18 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
US11038934B1 (en) 2020-05-11 2021-06-15 Apple Inc. Digital assistant hardware abstraction
US11061543B1 (en) 2020-05-11 2021-07-13 Apple Inc. Providing relevant data items based on context
US12301635B2 (en) 2020-05-11 2025-05-13 Apple Inc. Digital assistant hardware abstraction
US11755276B2 (en) 2020-05-12 2023-09-12 Apple Inc. Reducing description length based on confidence
US11490204B2 (en) 2020-07-20 2022-11-01 Apple Inc. Multi-device audio adjustment coordination
US11438683B2 (en) 2020-07-21 2022-09-06 Apple Inc. User identification using headphones
JP2023131732A (ja) * 2022-03-09 2023-09-22 株式会社デンソーテン 通話処理装置および通話処理方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004153380A (ja) * 2002-10-29 2004-05-27 Nissan Motor Co Ltd 情報提供装置、情報取得装置および情報提供システム
JP2004343335A (ja) * 2003-05-14 2004-12-02 Denso Corp 音声入出力装置及び通話システム
JP2006222969A (ja) * 2005-02-09 2006-08-24 Bose Corp 車両通信
US20130332886A1 (en) * 2012-06-08 2013-12-12 Apple Inc. Identification of recently downloaded content
WO2014141574A1 (ja) * 2013-03-14 2014-09-18 日本電気株式会社 音声制御システム、音声制御方法、音声制御用プログラムおよび耐雑音音声出力用プログラム

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005024781A1 (en) * 2003-08-29 2005-03-17 Johnson Controls Technology Company System and method of operating a speech recognition system in a vehicle
US8254837B2 (en) * 2009-04-23 2012-08-28 Motorola Mobility Llc Establishing full-duplex audio over an asynchronous bluetooth link
WO2013172827A1 (en) * 2012-05-16 2013-11-21 Nuance Communications, Inc. Speech communication system for combined voice recognition, hands-free telephony and in-communication
US8831957B2 (en) * 2012-08-01 2014-09-09 Google Inc. Speech recognition models based on location indicia
US10593326B2 (en) * 2013-04-25 2020-03-17 Sensory, Incorporated System, method, and apparatus for location-based context driven speech recognition

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004153380A (ja) * 2002-10-29 2004-05-27 Nissan Motor Co Ltd 情報提供装置、情報取得装置および情報提供システム
JP2004343335A (ja) * 2003-05-14 2004-12-02 Denso Corp 音声入出力装置及び通話システム
JP2006222969A (ja) * 2005-02-09 2006-08-24 Bose Corp 車両通信
US20130332886A1 (en) * 2012-06-08 2013-12-12 Apple Inc. Identification of recently downloaded content
WO2014141574A1 (ja) * 2013-03-14 2014-09-18 日本電気株式会社 音声制御システム、音声制御方法、音声制御用プログラムおよび耐雑音音声出力用プログラム

Also Published As

Publication number Publication date
JP6318621B2 (ja) 2018-05-09
JP2015130554A (ja) 2015-07-16
US20160329060A1 (en) 2016-11-10

Similar Documents

Publication Publication Date Title
JP6318621B2 (ja) 音声処理装置、音声処理システム、音声処理方法、音声処理プログラム
US11676601B2 (en) Voice assistant tracking and activation
CN108924706B (zh) 蓝牙耳机切换控制方法、蓝牙耳机及计算机可读存储介质
CN108684029B (zh) 一种蓝牙配对连接方法及系统、蓝牙设备和终端
US12022019B2 (en) Method and electronic device for adjusting output level of speaker on basis of distance from external electronic device
JP5570641B2 (ja) 携帯端末装置、車載器、情報提示方法及び情報提示プログラム
CN106205593A (zh) 一种车内声音调整方法及相关设备
CN107993672B (zh) 频带扩展方法及装置
CN107633849B (zh) 蓝牙设备音量调整方法、设备及计算机可读存储介质
CN111971985A (zh) 耳机配对、耳机控制方法、蓝牙耳机收纳装置及介质
JP2017138536A (ja) 音声処理装置
CN107147767B (zh) 通话音量控制方法、装置、存储介质及终端
US11546693B2 (en) Method for generating audio signal using plurality of speakers and microphones and electronic device thereof
US8934886B2 (en) Mobile apparatus and method of voice communication
WO2017166751A1 (zh) 移动终端音频调节方法、装置和电子设备
CN105635482A (zh) 一种来电处理方法和装置
KR20150053276A (ko) 이동통신단말기와 차량 헤드유닛이 연계된 음성 처리 방법과 그 시스템
CN108668018B (zh) 移动终端、音量控制方法及相关产品
JP5350567B1 (ja) 携帯端末装置、車載器、情報提示方法及び情報提示プログラム
WO2014141574A1 (ja) 音声制御システム、音声制御方法、音声制御用プログラムおよび耐雑音音声出力用プログラム
CN111402876A (zh) 数据传输方法、装置及设备
CN118215025A (zh) 车载蓝牙的语音通话系统、方法、电子设备及存储介质
JP6319722B2 (ja) ハンズフリー機器
JP2015004703A (ja) 電子装置、音声信号の転送方法、音声信号の転送プログラム及び音声処理システム
CN115331672B (zh) 设备控制方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14877432

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 15108739

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14877432

Country of ref document: EP

Kind code of ref document: A1