WO2016183825A1 - 一种定位声音发出位置的方法和终端设备 - Google Patents

一种定位声音发出位置的方法和终端设备 Download PDF

Info

Publication number
WO2016183825A1
WO2016183825A1 PCT/CN2015/079391 CN2015079391W WO2016183825A1 WO 2016183825 A1 WO2016183825 A1 WO 2016183825A1 CN 2015079391 W CN2015079391 W CN 2015079391W WO 2016183825 A1 WO2016183825 A1 WO 2016183825A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
sound signals
sound signal
terminal device
voice commands
Prior art date
Application number
PCT/CN2015/079391
Other languages
English (en)
French (fr)
Inventor
龚树强
陶志东
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP15892204.7A priority Critical patent/EP3264266B1/en
Priority to CN201580076714.3A priority patent/CN107430524B/zh
Priority to KR1020177030167A priority patent/KR102098668B1/ko
Priority to PCT/CN2015/079391 priority patent/WO2016183825A1/zh
Priority to JP2017557075A priority patent/JP6615227B2/ja
Priority to US15/566,979 priority patent/US10410650B2/en
Publication of WO2016183825A1 publication Critical patent/WO2016183825A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0316Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
    • G10L21/0364Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R11/00Arrangements for holding or mounting articles, not otherwise provided for
    • B60R11/02Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof
    • B60R11/0217Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof for loud-speakers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0272Voice signal separating
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/20Arrangements for obtaining desired frequency or directional characteristics
    • H04R1/32Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
    • H04R1/40Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
    • H04R1/406Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • H04S7/303Tracking of listener position or orientation
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R16/00Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for
    • B60R16/02Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements
    • B60R16/037Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements for occupant comfort, e.g. for automatic adjustment of appliances according to personal settings, e.g. seats, mirrors, steering wheel
    • B60R16/0373Voice control
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L2021/02082Noise filtering the noise being echo, reverberation of the speech
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • G10L2021/02161Number of inputs available containing the signal or the noise to be suppressed
    • G10L2021/02166Microphone arrays; Beamforming
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2201/00Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
    • H04R2201/02Details casings, cabinets or mounting therein for transducers covered by H04R1/02 but not provided for in any of its subgroups
    • H04R2201/025Transducer mountings or cabinet supports enabling variable orientation of transducer of cabinet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2430/00Signal processing covered by H04R, not provided for in its groups
    • H04R2430/20Processing of the output signals of the acoustic transducers of an array for obtaining a desired directivity characteristic
    • H04R2430/21Direction finding using differential microphone array [DMA]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2430/00Signal processing covered by H04R, not provided for in its groups
    • H04R2430/20Processing of the output signals of the acoustic transducers of an array for obtaining a desired directivity characteristic
    • H04R2430/23Direction finding using a sum-delay beam-former
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2499/00Aspects covered by H04R or H04S not otherwise provided for in their subgroups
    • H04R2499/10General applications
    • H04R2499/13Acoustic transducers and sound field adaptation in vehicles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/005Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones

Definitions

  • the embodiments of the present invention relate to the field of mobile communications, and in particular, to a method and a terminal device for locating a sound emitting position.
  • Speech recognition is the core technology point of the human-computer interaction interface of intelligent information systems.
  • a sound acquisition sensor is generally used to collect sound signals, and sound signal acquisition and speech recognition are performed for the position of the sound.
  • the scheme for improving the success rate of speech recognition can only extract the sound signal emitted by one position, and the sound signal emitted from other positions can only be extracted as noise and cannot accurately extract and locate the sound issuing position, and can not perform speech recognition.
  • a sound signal collected in an in-vehicle system can be used to collect sound signals in the surrounding environment, extracting sound signals emitted from the main cab, and extracting the extracted slaves from the main
  • the sound signal from the cab is used for voice recognition, and the in-vehicle system can respond to sound signals from the main cab.
  • the sound signal emitted from the passenger cab or the sound signal emitted from the rear seat of the vehicle is filtered by the in-vehicle system as noise, and the sound emitting position cannot be accurately extracted and positioned, and speech recognition is impossible.
  • the in-vehicle system can extract and voice-recognize the voice command of the "open skylight” issued from the main cab, but cannot extract the voice command of the "open skylight” from the passenger's cab or other positions such as the rear seat of the car. Position the position of other sound signals in the vehicle system.
  • the in-vehicle system cannot efficiently and accurately locate the emitting position of other sound signals in the car, thereby reducing the efficiency of the position where the positioning sound signal is emitted, and the user experience is low.
  • Embodiments of the present invention provide a method and a terminal device for locating a sound emitting position, so as to solve the problem that only sound information emitted from a single location can be located and extracted, and sound signals emitted from other locations cannot be located and extracted.
  • a first aspect of the present invention provides a method for locating a sound emitting position, comprising: acquiring K first sound signals; wherein K is an integer greater than or equal to 2; corresponding to N different positions N position parameters, extracting M second sound signals from the K first sound signals; wherein M is less than or equal to N, N is an integer greater than or equal to 2; determining a position corresponding to each second sound signal.
  • the extracting the M second sound signals from the K first sound signals according to the N position parameters corresponding to the N different positions includes: using a beamforming algorithm And extracting M second sound signals from the K first sound signals according to the N position parameters respectively.
  • the determining a location corresponding to each second sound signal specifically: determining, according to a position parameter corresponding to the Lth second sound signal, the Lth a position L corresponding to the second sound signal; wherein the Lth second sound signal is any one of the M second sound signals.
  • the method further includes: And extracting the M second sound signals to perform voice recognition; and acquiring M voice commands corresponding to the M second sound signals.
  • the method further includes: responding to the M voice commands.
  • responding to the M voice commands includes: prior to priority of M different locations corresponding to the M voice commands, Respond to high priority voice commands.
  • a second aspect of the present invention provides a terminal device, comprising: K sound collection sensors, configured to acquire K first sound signals; wherein K is an integer greater than or equal to 2; And for extracting M second sound signals from the K first sound signals according to N position parameters corresponding to the N different positions, and determining a position corresponding to each second sound signal, where M is less than or equal to N, N is an integer greater than or equal to 2.
  • the processor is configured to extract M second sound signals from the K first sound signals according to N position parameters corresponding to the N different positions, specifically including:
  • the processor is configured to extract M second sound signals from the K first sound signals according to the N position parameters by using a beamforming algorithm.
  • the determining, by the processor, the position corresponding to each second sound signal specifically: determining, according to the position parameter corresponding to the Lth second sound signal, a position L corresponding to the Lth second sound signal; wherein, the L The second sound signals are any one of the M second sound signals.
  • the processor is further configured to extract M second sound signals from the K first sound signals. And performing the voice recognition on the extracted M second sound signals, and acquiring the M voice commands corresponding to the M second sound signals.
  • the terminal device further includes an output device, where the output device is configured to acquire the M at the processor After the M voice commands corresponding to the second sound signals, the M voice commands are responded to.
  • the output device is configured to respond to the M voice commands, and specifically includes:
  • the output device is configured to preferentially respond to a command with a high priority according to priorities of M different locations corresponding to the M voice commands.
  • the coordinates of the K sound collection sensors in a three-dimensional space are different.
  • a third aspect of the present invention provides a device for locating a sound emitting position, the device comprising: an acquisition module, an extraction module and a determination module, wherein the acquisition module is configured to acquire K first sound signals; wherein, K An integer that is greater than or equal to 2; the extraction module is configured to extract M second sound signals from the K first sound signals according to N position parameters corresponding to the N different positions; wherein M is less than or equal to N, N is an integer greater than or equal to 2; the Susong determination module is used to determine a position corresponding to each second sound signal.
  • the extracting module is configured to extract M second sound signals from the K first sound signals according to the N position parameters corresponding to the N different positions, and specifically includes: Using the beamforming algorithm, M second sound signals are extracted from the K first sound signals according to the N position parameters, respectively.
  • the determining module is configured to determine a location corresponding to each second sound signal, specifically, the determining module is used by Determining, according to the position parameter corresponding to the Lth second sound signal, a position L corresponding to the Lth second sound signal; wherein, the Lth second sound signal is the M second sound signals Any one of them.
  • the apparatus further includes a voice recognition module and an acquisition module, where the voice recognition module is used by the extraction module After extracting the M second sound signals from the K first sound signals, performing voice recognition on the extracted M second sound signals; the acquiring module is configured to acquire the M second sound signals Corresponding M voice commands.
  • the apparatus further includes a response module, where the response module is configured to acquire the M second sounds in the acquiring module After the M voice commands corresponding to the signal, the response module is configured to respond to the M voice commands.
  • the responding module in response to the M voice commands, includes: prioritizing a high priority according to priorities of M different locations corresponding to the M voice commands Level of voice commands.
  • the embodiment of the present invention has the following advantages: using the beamforming algorithm, extracting M second sound signals from the K first sound signals according to the position parameters, thereby determining each second sound signal. Corresponding issuing position, by this method, it is possible to efficiently extract sound signals from different positions, provide voice recognition capability, and provide users with a higher user experience.
  • the conflicting command is processed by the priority method to reduce the error caused by the on-board central control device responding to multiple commands at the same time.
  • FIG. 1 is a flowchart of a method for positioning a sound emitting position according to an embodiment of the present invention
  • FIG. 2A is a schematic view showing the position of an interior compartment of a vehicle for positioning a sound emitting position according to an embodiment of the present invention
  • 2B is a schematic view showing the position of an interior compartment of a vehicle for positioning a sound emitting position according to another embodiment of the present invention
  • FIG. 3 is a flowchart of a method for positioning a sound emitting position according to another embodiment of the present invention.
  • FIG. 3A is a flowchart of a method for positioning a sound emitting position according to another embodiment of the present invention.
  • FIG. 3B is a flowchart of a method for positioning a sound emitting position according to another embodiment of the present invention.
  • FIG. 4 is a schematic structural diagram of a terminal device 400 according to an embodiment of the present invention.
  • the embodiment of the present invention provides a method for locating a sound emitting position.
  • the terminal device may be an in-vehicle central control device, a smart phone, a tablet computer, or the like.
  • the first sound signal or the second sound signal is only used for distinguishing, and does not represent order or order.
  • the application scenario of the embodiment of the present invention may be any sound collection and voice recognition scenario.
  • the sound in the vehicle system is used. Taking acquisition and speech recognition as an example, the method includes the following steps:
  • K sound collection sensors inside the vehicle system, and the processor can acquire K first sound signals; wherein K is an integer greater than or equal to 2.
  • K can be set to 2, that is, a first sound collection sensor and a second sound collection sensor can be installed in the cab and the passenger cab, respectively.
  • the first sound collection sensor and the second sound collection sensor simultaneously acquire the first sound signal.
  • other sound collection sensors can be installed in the rear seat of the vehicle or other locations in the vehicle.
  • the first sound signal is an ambient sound inside the in-vehicle system, and includes a sound signal emitted from different positions in the vehicle and a sound signal outside the vehicle.
  • the first sound signal may include a sound signal emitted from a cab position (eg, position 1 as shown in FIG. 2A), which is emitted from a passenger's cab position (eg, position 2 as shown in FIG. 2A).
  • the sound signal is at least one of a sound signal emitted from a position of the rear seat of the in-vehicle system (for example, position 3 and position 4 as shown in FIG. 2A) and noise outside the in-vehicle system.
  • the coordinates of the first sound collection sensor and the second sound collection sensor are not coincident in the spatial position, and the first sound collection sensor and the second sound collection sensor are separated by a certain distance.
  • the first sound collection sensor and the second sound collection sensor are respectively disposed on the left and right sides of the center rear view mirror A of the in-vehicle system.
  • the first sound collection sensor is disposed at the position C of the in-vehicle system
  • the second sound collection sensor is disposed at the position B of the in-vehicle system. Therefore, the time of the sound signal collected by the first sound collecting sensor and the second sound collecting sensor is different, so that the sound signal collected by the first sound collecting sensor and the sound signal collected by the second sound collecting sensor form a sound signal. Phase difference.
  • the in-vehicle system includes four sound collection sensors, and at this time, K is 4.
  • the four sound collection sensors are disposed at a central position of the in-vehicle system as shown in FIG. 2B.
  • Extracting M second sound signals from the K first sound signals specifically, extracting M second sound signals from the K first sound signals by using a beamforming algorithm; or using beamforming
  • the algorithm filters out other sound signals from the K first sound signals and extracts M second sound signals.
  • the position where the sound signal is emitted is the cab position
  • the corresponding position parameter is the parameter of the cab position
  • the vehicle-mounted central control device extracts from the K first sound signals according to the position parameter of the cab corresponding to the cab.
  • the second sound signal is extracted according to the position parameter of the cab by using a beamforming algorithm, and the extracted number is determined according to the position parameter corresponding to the second sound signal.
  • the corresponding sounding position of the second sound signal is the cab.
  • the invention provides a method for locating a sound emitting position, which uses a beamforming algorithm to extract M second sound signals from K first sound signals according to positional parameters, thereby determining the corresponding sounding of each second sound signal. position. In this way, the sound signals emitted from different locations can be efficiently extracted, the voice recognition capability is improved, and the user is provided with a higher user experience.
  • FIG. 3 is a flow chart of a method for locating a sound emitting position according to another embodiment of the present invention.
  • the embodiment of the present invention is also applied to an in-vehicle system as an example. As shown in FIG. 3, the method includes the following steps:
  • position 1 is the cab position
  • position 2 is the passenger cab position
  • position 3 is the left side position of the rear seat of the vehicle system
  • position 4 is the rear seat of the vehicle system. The location on the right.
  • the in-vehicle central control device sets the priority of voice commands in response to four different positions in the in-vehicle system according to four different positions.
  • the voice command priority set by an ordinary family car is taken as an example.
  • the command "air conditioner start” and position 4 are simultaneously issued at position 1 and the command “air conditioner off” is taken as an example.
  • K is 2 as an example.
  • the first sound collection sensor and the second sound collection sensor are respectively mounted on the left and right sides of the rear view mirror A.
  • the first sound collection sensor and the second sound collection sensor simultaneously acquire the first sound signal.
  • other sound collection sensors can be installed in the rear seat of the vehicle or other locations in the vehicle.
  • the first sound collecting sensor and the second sound collecting sensor simultaneously collect the command issued by the position 1 "air conditioner.
  • the sound signal is activated, and the first sound collection sensor and the second sound collection sensor simultaneously acquire the sound signal of the command "air conditioner off” issued by the position 4.
  • N is 4 and M is 2 for illustration.
  • the first sound collection sensor and the second sound collection sensor are separated by a certain distance. Therefore, the time of the sound signal collected by the first sound collecting sensor and the second sound collecting sensor is different, so that the sound signal collected by the first sound collecting sensor and the sound signal collected by the second sound collecting sensor form a phase difference.
  • the first sound collection sensor and the second sound collection sensor are disposed on the left and right rear view mirrors.
  • the present invention does not limit the number of sound collection sensors, and the position of the sound collection sensor is not limited.
  • other sound collecting sensors may be placed beside the position where the sound may be emitted, such as at the rear side of the seat at position 1 or position 2 as shown in FIG. 2A.
  • the in-vehicle central control device extracts the position from the position 1 according to the preset position parameter of the position 1. Second sound signal. Using the beamforming algorithm, the in-vehicle central control device extracts the second sound signal emitted from position 1 from the acquired first sound signal according to the positional parameter of the preset position 1.
  • the in-vehicle central control device extracts the second sound signal emitted from the position 4 according to the positional parameter of the preset position 4.
  • the in-vehicle central control device extracts the second sound signal emitted from position 4 from the acquired first sound signal according to the positional parameter of the preset position 4.
  • the in-vehicle central control device utilizes a beamforming algorithm to extract a sound signal that conforms to the positional parameter of the preset position 1 based on the positional parameter of position 1. For example, an "air conditioning start" sound signal sent from the position 1 is acquired; the vehicle central control device uses a beamforming algorithm to extract a sound signal conforming to the preset position parameter of the position 2 according to the position parameter of the position 4. For example, an acoustic signal "air conditioning off" from position 4 is collected.
  • the in-vehicle central control device uses the beamforming algorithm to extract two second sound signals from the two first sound signals according to the four position parameters.
  • the second sound signal emitted from the position 1 is extracted according to the position parameter of the position 1 by using a beamforming algorithm, and determined according to the position parameter corresponding to the second sound signal.
  • the emitted position corresponding to the extracted second sound signal is position 1.
  • the in-vehicle central control device performs speech recognition on the extracted sound signal to identify the extracted sound signal.
  • the vehicle-mounted central control device performs voice recognition on the sound signal extracted from the position 1, and recognizes the extracted sound signal as “air-conditioning start”; the vehicle-mounted central control device performs voice recognition on the sound signal extracted from the position 4, and recognizes and extracts The sound signal is "air conditioner off”.
  • the in-vehicle central control device acquires the voice command corresponding to the extracted M second sound signals.
  • the in-vehicle central control device acquires a voice command corresponding to the sound signal emitted by the extracted position 1 to acquire a voice command of “air conditioning start”; and the in-vehicle central control device acquires a sound signal corresponding to the sound signal emitted by the extracted position 4 Voice command to get the voice command of "air conditioner off”.
  • the in-vehicle central control device responds to the M voice commands according to the acquired voice commands corresponding to the extracted M second sound signals.
  • the air conditioner is activated in response to the voice command.
  • the in-vehicle central control device performs speech recognition on the sound signal extracted at position 1 and the sound signal extracted at position 4, and recognizes the extracted sound signal.
  • the in-vehicle central control device performs speech recognition on the extracted sound signal emitted from the position 1 and the sound signal emitted from the position 4, and recognizes the extracted sound signal.
  • the in-vehicle central control device responds to the two voice commands according to the acquired "air conditioning start” issued by the extracted position 1 and the "air conditioning off” voice command issued by the position 4.
  • the voice command of the two different positions corresponding to the two voice commands is preferentially responded to the high priority voice command, for example, the location.
  • the priority of 1 is higher than the priority of position 4.
  • the in-vehicle central control device first responds to the voice command "air conditioning start" of position 1, thereby turning on the air conditioner.
  • the in-vehicle central control device responds to the voice command "air conditioner off" of position 4, at this time, because the voice command of the vehicle-mounted central control device responding to position 1 is “air conditioner on”, the voice command of position 4 is “air conditioner off", the voice of position 1
  • the command and the voice command of position 4 are conflict commands, and the in-vehicle central control device cannot simultaneously respond to the voice command of position 1 and the voice command of position 4. Therefore, after the vehicle-mounted central control device performs voice recognition on the sound signal of the position 4, the voice command corresponding to the extracted sound signal is acquired, and the voice command of the position 4 is not responded.
  • the conflict command is processed by the priority method, when the vehicle-mounted central control device responds to multiple conflict commands, the in-vehicle central control device cannot make a correct response due to the command conflict, thereby reducing the error caused by the response error.
  • the conflicting command is specifically: if at least two commands use the same resource and execute the at least two commands, the operations of the same resource used are different, and the at least two commands are conflicting commands. .
  • the vehicle-mounted central control device when the acquired two voice commands collide, the time judgment factor is added, and the vehicle-mounted central control device recognizes the conflict command command within the preset time T1 after the high-priority command is recognized. However, when the priority of the identified conflict command is low, the command command with lower priority is ignored. If the in-vehicle central control device recognizes the conflict command after the preset time T1 after the high priority command is recognized, the in-vehicle central control device sequentially responds to the acquired voice command according to the time sequence of the recognized voice command.
  • FIG. 3A is a flowchart of a method for locating a sound emitting position according to another embodiment of the present invention.
  • the following steps may be performed:
  • S401 Determine whether at least one seat of the in-vehicle system is sitting.
  • the in-vehicle system can determine whether the seat of the in-vehicle system is seated by means of gravity sensing.
  • whether the seat of the in-vehicle system in FIG. 2A is seated is determined by gravity sensing. For example, it is judged whether position 1, position 2, position 3 or position 4 in Fig. 2A is sitting.
  • step S301 is not performed.
  • step S301 is performed.
  • the sound signal Before collecting the sound signal, it is first determined whether at least one seat of the vehicle system is sitting, and only sitting on the seat of the vehicle system, and then positioning the position of the sound, improving the efficiency of sound collection and improving the efficiency of determining the position of the sound.
  • step S305a may be performed: identifying the extracted voiceprints of the M second sound signals.
  • S305b Measure the weight of the user on the seat of the vehicle system.
  • S305c determining the identity of the user in combination with the measured weight of the user and the voiceprint of the identified second sound signal.
  • S305d Determine, according to the determined identity of the user, a priority of a voice command corresponding to the second sound signal sent by the user.
  • S305e respond to the voice command corresponding to the second sound signal according to a priority of the voice command corresponding to the second sound signal sent by the user.
  • the priority of the user's identity and the voice command corresponding to the voice signal sent by the user is determined.
  • the priority of responding to the plurality of voice commands is determined in conjunction with the priority of the voice command corresponding to the voice signal sent by the user.
  • the invention provides a method for locating a sound emitting position, which uses a beamforming algorithm to extract M second sound signals from K first sound signals according to positional parameters, thereby determining the corresponding sounding of each second sound signal. position. Further, set the priority of the voice command, through the excellent First, the high-priority method is processed to process the conflict command, and the conflict caused by the in-vehicle central control device responding to multiple conflicting commands is reduced, the error caused by the response error is reduced, and the user experience is improved.
  • FIG. 4 is a schematic diagram of a terminal device 400 according to an embodiment of the present invention. It can be used to perform the aforementioned methods of the embodiments of the present invention.
  • the terminal device 400 may be a terminal device including a mobile phone, a tablet computer, a PDA (Personal Digital Assistant), a POS (Point of Sale), an in-vehicle central control terminal device, and a terminal device.
  • the 400 includes an RF (Radio Frequency) circuit 410, a memory 420, an input device 430, a display device 440, a sensor 450, an audio circuit 460, a WiFi (wireless fidelity) module 470, a processor 480, and a power source 490. component.
  • RF Radio Frequency
  • FIG. 4 is only an example of implementation, and does not constitute a limitation on the terminal device, and may include more or less components than those illustrated, or may combine some components. , or different parts layout.
  • the RF circuit 410 can be used for transmitting and receiving information or during a call, and receiving and transmitting the signal. Specifically, after receiving the downlink information of the base station, the processor 480 processes the data. In addition, the uplink data is designed to be sent to the base station.
  • the RF circuit 410 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, an LNA (Low Noise Amplifier), a duplexer, and the like.
  • the RF circuit 410 can also communicate with the network and other terminal devices through wireless communication.
  • the wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication), GPRS (General Packet Radio Service), CDMA (Code Division Multiple Access). , Code Division Multiple Access), WCDMA (Wideband Code Division Multiple Access), LTE (Long Term Evolution), e-mail, SMS (Short Messaging Service), and the like.
  • GSM Global System of Mobile communication
  • GPRS General Packet Radio Service
  • CDMA Code Division Multiple Access
  • WCDMA Wideband Code Division Multiple Access
  • LTE Long Term Evolution
  • e-mail Short Messaging Service
  • the memory 420 can be used to store software programs and modules, and the processor 480 executes various functional applications and data processing of the terminal device 400 by running software programs and modules stored in the memory 420.
  • the memory 420 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application required for at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may be stored according to The data created by the use of the terminal device 400 (such as audio data, phone book, etc.) and the like.
  • memory 420 can include high speed random access memory, and can also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
  • the display device 440 can be used to display information input by the user or information provided to the user and various menus of the terminal device 400.
  • the display device 440 may include a display panel 441.
  • the display panel 441 may be configured in the form of an LCD (Liquid Crystal Display), an OLED (Organic Light-Emitting Diode), or the like.
  • the touch panel 431 can cover the display panel 441. When the touch panel 431 detects a touch operation on or near the touch panel 431, it transmits to the processor 480 to determine the type of the touch event, and then the processor 480 according to the touch event. The type provides a corresponding visual output on display panel 441.
  • the touch panel 431 and the display panel 441 function as two separate components to implement input and input functions of the terminal device 400.
  • the touch panel 431 and the display panel 441 can be integrated to implement the input and output functions of the terminal device 400.
  • the touch panel 431 and the display panel 441 can be integrated into a touch screen to implement the terminal device 400. Input and output functions.
  • Terminal device 400 may also include at least one type of sensor 450, such as a light sensor, motion sensor, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor, wherein the ambient light sensor may adjust the brightness of the display panel 441 according to the brightness of the ambient light, and the proximity sensor may close the display panel 441 when the terminal device 400 moves to the ear. Or backlight.
  • the accelerometer sensor can detect the magnitude of acceleration in all directions (usually three axes). When it is stationary, it can detect the magnitude and direction of gravity.
  • gesture of the mobile phone such as horizontal and vertical screen switching, related Game, magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; as for the gyroscope, barometer, hygrometer, thermometer, infrared sensor and other sensors that can be configured in the terminal device 400, here No longer.
  • the audio circuit 460, the speaker 461, and the microphone 462 can provide an audio interface between the user and the terminal device 400.
  • the audio circuit 460 can transmit the converted electrical data of the received audio data to the speaker 461 for conversion to the sound signal output by the speaker 461; on the other hand, the microphone 462 converts the collected sound signal into an electrical signal by the audio circuit 460. After receiving, it is converted into audio data, and then processed by the audio data output processor 480, sent to the other mobile phone via the RF circuit 410, or outputted to the memory 420 for further processing.
  • the terminal device 400 can help the user to send and receive emails, browse web pages, access streaming media, etc. through the WiFi module 470, which provides wireless broadband Internet access to the user.
  • FIG. 4 shows the WiFi module 470, it can be understood that it does not belong to the essential configuration of the terminal device 400, and may be omitted as needed within the scope of not changing the essence of the invention.
  • Processor 480 is the control center of terminal device 400, which connects various portions of the entire handset using various interfaces and lines, by running or executing software programs and/or modules stored in memory 420, and recalling data stored in memory 420.
  • the various functions and processing data of the terminal device 400 are executed to perform overall monitoring of the terminal device.
  • the processor 480 may include one or more processing units; preferably, the processor 480 may integrate an application processor and a modem processor, where the application processor mainly processes an operating system, a user interface, an application, and the like.
  • the modem processor primarily handles wireless communications. It can be understood that the above modem processor may not be integrated into the processor 480.
  • the processor 480 may specifically be a central processing unit (CPU).
  • the terminal device 400 further includes a power source 490 (such as a battery) for supplying power to the various components.
  • a power source 490 such as a battery
  • the power source can be logically connected to the processor 480 through the power management system to manage functions such as charging, discharging, and power management through the power management system. .
  • the terminal device 400 includes: K sound collection sensors 450 and a processor 480 having the following functions:
  • the sound collection sensor 450 is configured to acquire K first sound signals; wherein K is an integer greater than or equal to 2.
  • the coordinates of the K sound collection sensors in the three-dimensional space are different.
  • the processor 480 is configured to extract M second sound signals from the K first sound signals according to N position parameters corresponding to the N different positions, and determine a position corresponding to each second sound signal, where M is less than or equal to N, and N is an integer greater than or equal to 2.
  • the processor 480 is configured to determine a location corresponding to each second sound signal, specifically: determining, according to a position parameter corresponding to the Lth second sound signal, the Lth a position L corresponding to the two sound signals; wherein the Lth second sound signal is any one of the M second sound signals.
  • the processor 480 is further configured to perform voice recognition on the extracted M second sound signals after extracting M second sound signals from the K first sound signals. And used to acquire M voice commands corresponding to the M second sound signals.
  • the terminal device 400 further includes: an output device 510, configured to respond after the processor acquires M voice commands corresponding to the M second sound signals.
  • the M voice commands are configured to respond after the processor acquires M voice commands corresponding to the M second sound signals. The M voice commands.
  • the output device 510 is configured to respond to the M voice commands, and specifically includes: the output device The priority is used to respond to the priority of the M different voice positions corresponding to the M voice commands.
  • the output device 510 may specifically be the audio circuit 460 or the display device 440.
  • a method and a terminal device for locating a sound emitting position are provided.
  • M second sound signals are extracted from K first sound signals according to position parameters, thereby determining each The corresponding sounding position of the second sound signal, by which the sound signal emitted from different positions can be efficiently extracted, and the voice recognition capability is provided, thereby providing a higher user experience for the user.
  • the disclosed server and method may be implemented in other manners.
  • the server embodiment described above is merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the embodiments of the present invention.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit. in.
  • the foregoing steps include the steps of the foregoing method embodiments; and the foregoing storage medium includes: a medium that can store program codes, such as a ROM, a RAM, a magnetic disk, or an optical disk.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Otolaryngology (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Quality & Reliability (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mechanical Engineering (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)
  • Telephone Function (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

本发明实施例提供了一种定位声音发出位置的方法和终端设备,采集K个第一声音信号;其中,K为大于等于2的整数;根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号;其中M小于等于N,N为大于等于2的整数;确定每个第二声音信号对应的位置。本发明实施例中,利用波束成型算法,根据位置参数,从K个第一声音信号中提取M个第二声音信号,从而确定每个第二声音信号对应的发出位置,通过这种方法,可以高效的提取不同位置发出的声音信号,提供语音识别能力,为用户提供了更高的用户体验。

Description

一种定位声音发出位置的方法和终端设备 技术领域
本发明实施例涉及移动通信领域,尤其涉及一种定位声音发出位置的方法和终端设备。
背景技术
语音识别是目前智能信息系统人机交互界面的核心技术点。为了提升语音识别的成功率,普遍采用声音采集传感器采集声音信号的方案,针对声音发出位置进行声音信号采集和语音识别。
目前,提高语音识别成功率的方案,只能提取一个位置发出的声音信号,对于其他位置发出的声音信号只能被当做噪声滤除而无法准确地提取和定位声音发出位置,也无法进行语音识别。以安装在汽车内的车载系统为例,目前可以通过安装在车载系统上的声音采集传感器采集周围环境中的声音信号,提取出从主驾驶室发出的声音信号,并对所述提取的从主驾驶室发出的声音信号进行语音识别,车载系统可以响应从主驾驶室发出的声音信号。但是,从副驾驶室发出的声音信号或者从车载后排座位发出的声音信号都被车载系统当做噪声滤除而无法准确地提取和定位声音发出位置,更无法进行语音识别。如,车载系统可以提取并语音识别从主驾驶室发出的“开天窗”的语音命令,但是,无法提取从副驾驶室或者从车载后排座位等其他位置发出的“开天窗”的语音命令,也无法定位车载系统内其他声音信号的发出位置。因此,在汽车内的车载系统这样的应用场景下,车载系统无法高效准确地定位汽车内其他声音信号的发出位置,降低了定位声音信号的发出位置的效率,用户体验低下。
发明内容
本发明实施例提供了一种定位声音发出位置的方法和终端设备,以解决只能定位和提取从单一位置发出的声音信息,无法定位和提取从其他位置发出的声音信号的问题。
本发明第一方面,提供了一种定位声音发出位置的方法,包括:采集K个第一声音信号;其中,K为大于等于2的整数;根据与N个不同位置对应 的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号;其中M小于等于N,N为大于等于2的整数;确定每个第二声音信号对应的位置。
在第一种可能的实现方式中,所述根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,具体包括:利用波束成型算法,分别根据所述N个位置参数,从所述K个第一声音信号中提取M个第二声音信号。
结合第一方面,在第二种可能的实现方式中,所述确定每个第二声音信号对应的位置,具体包括:根据第L个第二声音信号对应的位置参数,确定所述第L个第二声音信号对应的位置L;其中,第L个第二声音信号为所述M个第二声音信号中的任意一个。
结合第一方面和上述任一种可能的实现方式,在第三种可能的实现方式中,从所述K个第一声音信号中提取M个第二声音信号后,所述方法还包括:对提取的所述M个第二声音信号进行语音识别;获取所述M个第二声音信号对应的M个语音命令。
结合第一方面和第三种可能的实现方式,在第四种可能的实现方式中,在获取所述M个第二声音信号对应的M个语音命令之后,所述方法还包括:响应所述M个语音命令。
结合第一方面和第四种可能的实现方式,在第五种可能的实现方式中,响应所述M个语音命令包括:根据所述M个语音命令对应的M个不同位置的优先级,优先响应高优先级的语音命令。
本发明第二方面,提供了一种终端设备,其特征在于,所述终端设备包括:K个声音采集传感器,用于采集K个第一声音信号;其中K为大于等于2的整数;处理器,用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,并确定每个第二声音信号对应的位置,其中M小于等于N,N为大于等于2的整数。
在第一种可能的实现方式中,所述处理器用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,具体包括:所述处理器用于利用波束成型算法,分别根据所述N个位置参数,从所述K个第一声音信号中提取M个第二声音信号。
结合第二方面和第一种可能的实现方式,在第二种可能的实现方式中,所 述处理器用于确定每个第二声音信号对应的位置,具体包括:根据第L个第二声音信号对应的位置参数,确定所述第L个第二声音信号对应的位置L;其中,第L个第二声音信号为所述M个第二声音信号中的任意一个。
结合第二方面和上述任意的一种可能的实现方式,在第三种可能的实现方式中,所述处理器还用于从所述K个第一声音信号中提取M个第二声音信号后,对提取的所述M个第二声音信号进行语音识别,并用于获取所述M个第二声音信号对应的M个语音命令。
结合第二方面和上述任意的一种可能的实现方式,在第四种可能的实现方式中,所述终端设备还包括输出装置;所述输出装置,用于在所述处理器获取所述M个第二声音信号对应的M个语音命令之后,响应所述M个语音命令。
结合第二方面和第四种可能的实现方式中,在第五种可能的实现方式中,所述输出装置用于响应所述M个语音命令,具体包括:
所述输出装置用于根据所述M个语音命令对应的M个不同位置的优先级,优先响应优先级高的命令。
结合第二方面和上述任意的一种可能的实现方式,所述K个声音采集传感器在三维空间内的坐标不同。
本发明第三方面,提供了一种定位声音发出位置的装置,所述装置包括:采集模块,提取模块和确定模块,其中,所述采集模块用于采集K个第一声音信号;其中,K为大于等于2的整数;所述提取模块用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号;其中M小于等于N,N为大于等于2的整数;宿松确定模块用于确定每个第二声音信号对应的位置。
在第一种可能的实现方式中,所述提取模块用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,具体包括:利用波束成型算法,分别根据所述N个位置参数,从所述K个第一声音信号中提取M个第二声音信号。
结合第三方面和第一种可能的实现方式中,在第二种可能的实现方式中,所述确定模块用于确定每个第二声音信号对应的位置,具体包括,所述确定模块用于根据第L个第二声音信号对应的位置参数,确定所述第L个第二声音信号对应的位置L;其中,第L个第二声音信号为所述M个第二声音信号中 的任意一个。
结合第三方面和上述任意的一种可能的实现方式,在第三种可能的实现方式中,所述装置还包括语音识别模块和获取模块,所述语音识别模块用于在所述提取模块用于从所述K个第一声音信号中提取M个第二声音信号后,对提取的所述M个第二声音信号进行语音识别;所述获取模块用于获取所述M个第二声音信号对应的M个语音命令。
结合第三方面和第三种可能的实现方式,在第四种可能的实现方式中,所述装置还包括响应模块,所述响应模块用于在所述获取模块获取所述M个第二声音信号对应的M个语音命令之后,所述响应模块用于响应所述M个语音命令。
结合第三方面和第五中可能的实现方式中,所述响应模块用于响应所述M个语音命令包括:根据所述M个语音命令对应的M个不同位置的优先级,优先响应高优先级的语音命令。
从以上技术方案可以看出,本发明实施例具有以下优点:利用波束成型算法,根据位置参数,从K个第一声音信号中提取M个第二声音信号,从而可以确定每个第二声音信号对应的发出位置,通过这种方法,可以高效的提取不同位置发出的声音信号,提供语音识别能力,为用户提供了更高的用户体验。并通过优先级的方法来处理冲突命令,减少车载中控设备同时响应多个命令带来的误差。
附图说明
为了更清楚地说明本发明实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1是本发明实施例提供的一种定位声音发出位置的方法流程图;
图2A是本发明实施例提供了定位声音发出位置的汽车内室位置示意图;
图2B是本发明另一实施例提供了定位声音发出位置的汽车内室位置示意图;
图3是本发明另一实施例提供了一种定位声音发出位置的方法流程图;
图3A是本发明另一实施例提供了一种定位声音发出位置的方法流程图;
图3B是本发明另一实施例提供了一种定位声音发出位置的方法流程图;
图4是本发明实施例提供的终端设备400结构示意图。
具体实施方式
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。
本发明实施例提供了一种定位声音发出位置的方法,在本发明实施例中,涉及到的终端设备可以为车载中控设备、智能手机、平板电脑等。
在现有技术中,声音采集传感器采集声音信号的方案结合波束成型算法被应用于采集声音信号和语音识别,通过这种方式已经大大提高了语音识别的成功率,但是通过这种方式只能识别单一的一个声音发出位置发出的声音信号,当有多个位置的声音发出位置发出声音信号时,语音识别系统无法同时识别多个声音信号。
在本发明实施例中,第一声音信号或者第二声音信号仅用于区分,并不代表次序或者顺序。
图1是本发明实施例提供的一种定位声音发出位置的方法流程图,本发明实施例的应用场景可以为任意声音采集和语音识别的场景,本发明实施例中,以车载系统中的声音采集和语音识别为例,所述方法包括以下步骤:
S101,采集K个第一声音信号;其中,K为大于等于2的整数。
在车载系统中,车载系统内部有K个声音采集传感器,处理器则可以采集K个第一声音信号;其中,K为大于等于2的整数。
例如,在车载系统中,可以设置K为2,即可以在驾驶室和副驾驶室里分别安装了第一声音采集传感器和第二声音采集传感器。
第一声音采集传感器和第二声音采集传感器同时采集第一声音信号。可选的,在车载系统中,还可以在车载后排座位或车载内其他位置安装其他声音采集传感器。
在本发明实施例中,第一声音信号为车载系统内部的环境声音,包括从车载内不同位置发出的声音信号和车外的声音信号。其中,所述第一声音信号可以包括从驾驶室位置(例如,如图2A所示,位置①)发出的声音信号,从副驾驶室位置(例如,如图2A所示,位置②)发出的声音信号,从车载系统后排座位的位置(例如,如图2A所示,位置③和位置④)发出的声音信号和车载系统外部的噪音中的至少一个。
S102,根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号;其中M小于等于N,N为大于等于2的整数。
同样以车载系统的情况为例进行说明,由于第一声音采集传感器和第二声音采集传感器在空间位置上坐标不重合,而且第一声音采集传感器和第二声音采集传感器拉开了一定的距离。如图2A所示,第一声音采集传感器和第二声音采集传感器分别设置在车载系统的中后视镜A的左右两侧。第一声音采集传感器设置在车载系统的位置C,第二声音采集传感器设置在车载系统的位置B。因此,第一声音采集传感器和第二声音采集传感器采集到的声音信号的时间不相同,从而,对于第一声音采集传感器采集到的声音信号和第二声音采集传感器采集到的声音信号会形成一个相位差。
在本发明另一实施例中,如图2B所示,车载系统包括4个声音采集传感器,此时,K为4。所述4个声音采集传感器设置在车载系统的中央位置,如图2B所示。
从所述K个第一声音信号中提取M个第二声音信号,具体可以为利用波束成型算法,从K个第一声音信号中,提取出M个第二声音信号;也可以为利用波束成型算法,从K个第一声音信号中,过滤其他声音信号,提取M个第二声音信号。
例如,声音信号发出的位置为驾驶室位置,对应的位置参数为驾驶室位置的参数,车载中控设备根据与驾驶室对应的驾驶室的位置参数,从所述K个第一声音信号中提取从驾驶室发出的第二声音信号。
S103,确定每个第二声音信号对应的位置。利用波束成型算法,所述车 载中控设备分别根据N个位置参数,从所述K个第一声音信号中提取M个第二声音信号。
例如,当位置参数是驾驶室的位置参数时,利用波束成型算法,根据驾驶室的位置参数,提取第二声音信号,并根据所述第二声音信号对应的位置参数,确定所述提取的第二声音信号对应的发出位置为驾驶室。
本发明提供了一种定位声音发出位置的方法,利用波束成型算法,根据位置参数,从K个第一声音信号中提取M个第二声音信号,从而可以确定每个第二声音信号对应的发出位置。通过这种方法,可以高效的提取不同位置发出的声音信号,提高了语音识别能力,为用户提供了更高的用户体验。
图3是本发明另一实施例提供了一种定位声音发出位置的方法流程图。本发明实施例同样以应用于车载系统为例进行说明,如图3所示,所述方法包括以下步骤:
S301a,设置响应N个不同位置的语音命令的优先级。
同样以图2A的位置示意图为例,图2A中,位置①为驾驶室位置,位置②为副驾驶室位置,位置③为车载系统后排座位左侧的位置,位置④为车载系统后排座位右侧的位置。
在本发明实施例中,以车载系统为例,假设K为2,N为4,M为2。
车载中控设备在车载系统内根据4个不同位置,设置响应4个不同位置的语音命令的优先级。
例如,以普通家用轿车设置的语音命令优先级为例。
命令 开天窗 关天窗 开收音机 播音乐
位置① 1 1 1 1
位置② 1 1 2 2
位置③ 2 2 3 3
位置④ 2 2 4 4
表(1)  普通家用轿车设置的语音命令优先级
从表(1)中可以看出来,当位置①发出如“开天窗”、“关天窗”、“开收音机”或者“播音乐”等命令时,位置①发出命令的优先级高于其他位置发出的同样语义的命令的优先级。
在本发明另一实施例中,设置响应N个不同位置的语音命令的优先级时, 加入儿童和成人声音判断因素。设置语音命令为儿童声音的优先级为低优先级或者设置当语音命令为儿童声音时,屏蔽儿童声音的语音命令,设置语音命令为成人声音的优先级为高优先级。
在本发明实施例中以位置①发出命令“空调启动”和位置④同时发出命令“空调关闭”为例。
S301,采集K个第一声音信号。
在本发明实施例中,以K为2举例说明。
其中,在车载系统中,在中后视镜A的左右两侧分别安装了第一声音采集传感器和第二声音采集传感器。
第一声音采集传感器和第二声音采集传感器同时采集第一声音信号。可选的,在车载系统中,还可以在车载后排座位或车载内其他位置安装其他声音采集传感器。
例如,当位置①发出命令“空调启动”的声音信号和位置④同时发出命令“空调关闭”的声音信号时,第一声音采集传感器和第二声音采集传感器同时采集到位置①发出的命令“空调启动”的声音信号,同样第一声音采集传感器和第二声音采集传感器同时采集到位置④发出的命令“空调关闭”的声音信号。
S302,根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号;其中M小于等于N,N为大于等于2的整数。
在本发明实施例中,以N为4,M为2进行举例说明。
由于第一声音采集传感器和第二声音采集传感器在空间位置上坐标不重合,而且第一声音采集传感器和第二声音采集传感器拉开了一定的距离。因此,第一声音采集传感器和第二声音采集传感器采集到的声音信号的时间不相同,从而,对于第一声音采集传感器采集到的声音信号和第二声音采集传感器采集的声音信号会形成一个相位差。
本发明以第一声音采集传感器和第二声音采集传感器设置在中后视镜左右为例,本发明对声音采集传感器的数量不做限制,对声音采集传感器的位置也不做限制。例如,还可以将其他声音采集传感器设置在可能发出声音的位置旁边,如安装在如图2A所示的位置①或位置②的座位的后侧。
例如,车载中控设备根据预置的位置①的位置参数,提取从位置①发出的 第二声音信号。利用波束成型算法,根据预置的位置①的位置参数,车载中控设备从采集的第一声音信号中提取从位置①发出的第二声音信号。
同时,车载中控设备根据预置的位置④的位置参数,提取从位置④发出的第二声音信号。利用波束成型算法,根据预置的位置④的位置参数,车载中控设备从采集的第一声音信号中提取从位置④发出的第二声音信号。
例如,车载中控设备利用波束成型算法,根据位置①的位置参数,提取符合预置的位置①的位置参数的声音信号。如,采集到从位置①发出的“空调启动”的声音信号;车载中控设备利用波束成型算法,根据位置④的位置参数,提取符合预置的位置②的位置参数的声音信号。如,采集到从位置④发出的“空调关闭”的声音信号。
S303,确定每个第二声音信号对应的位置。
利用波束成型算法,所述车载中控设备分别根据4个位置参数,从所述2个第一声音信号中提取2个第二声音信号。
例如,当位置参数是位置①的位置参数时,利用波束成型算法,根据位置①的位置参数,提取从位置①发出的第二声音信号,并根据所述第二声音信号对应的位置参数,确定所述提取的第二声音信号对应的发出位置为位置①。
S304,对提取的所述M个第二声音信号进行语音识别。
车载中控设备对所述提取的声音信号进行语音识别,识别提取的声音信号。
例如,车载中控设备对从位置①提取的声音信号进行语音识别,识别提取出的声音信号为“空调启动”;车载中控设备对从位置④提取的声音信号进行语音识别,识别提取出的声音信号为“空调关闭”。
S305,获取所述M个第二声音信号对应的语音命令。
车载中控设备获取所述提取的M个第二声音信号对应的语音命令。
例如,车载中控设备获取所述提取的位置①发出的声音信号对应的语音命令,获取“空调启动”的语音命令;并且,车载中控设备获取所述提取的位置④发出的声音信号对应的语音命令,获取“空调关闭”的语音命令。
S306,响应所述M个语音命令。
车载中控设备根据获取的所述提取的M个第二声音信号对应的语音命令,响应所述M个语音命令。
例如,车载中控设备获取到从位置①发出的“空调启动”的语音命令后,响应所述语音命令,启动空调。
在本发明另一实施例中,车载中控设备对位置①提取的声音信号和位置④提取的声音信号进行语音识别,识别提取的声音信号。车载中控设备对提取的从位置①发出的声音信号和从位置④发出的声音信号进行语音识别,识别提取的声音信号。获取所述提取的位置①发出的声音信号对应的语音命令,和获取所述提取的位置④发出的声音信号对应的语音命令,例如,获取位置①发出的“空调启动”和位置④发出的“空调关闭”的语音命令。车载中控设备根据获取的提取的位置①发出的“空调启动”和位置④发出的“空调关闭”的语音命令,响应所述2个语音命令。可选的,当车载中控设备语音识别到两个位置的语音命令时,会根据所述2个语音命令对应的2个不同位置的优先级,优先响应高优先级的语音命令,例如,位置①的优先级高于位置④的优先级,车载中控设备先响应位置①的语音命令“空调启动”,从而开启空调。车载中控设备再响应位置④的语音命令“空调关闭”,此时由于车载中控设备响应位置①的语音命令为“空调开启”,位置④的语音命令为“空调关闭”,位置①的语音命令和位置④的语音命令为冲突命令,车载中控设备无法同时响应位置①的语音命令和位置④的语音命令。因此,车载中控设备对位置④的声音信号进行语音识别后,获取提取的声音信号对应的语音命令,不响应位置④的语音命令。通过优先级的方法来处理冲突命令,减少车载中控设备响应多个冲突命令时,由于命令冲突带来的车载中控设备无法做出正确的响应,减少响应错误带来的误差。
其中,所述冲突命令具体为,如果至少两个命令会使用同一资源、且执行所述至少两个命令时,对所述使用的同一资源的操作不同,则所述至少两个命令为冲突命令。
在本发明另一实施例中,当获取的两个语音命令冲突时,增加时间判断因素,当高优先级的命令被识别之后的预设时间T1时间内,车载中控设备识别到冲突命令命令,但是识别到的冲突命令的优先级较低时,则忽略优先级较低的命令命令。若高优先级命令被识别之后的预设时间T1时间之后,车载中控设备识别到冲突命令时,车载中控设备则按照识别到的语音命令的时间顺序来顺序响应获取的语音命令。
图3A是本发明另一实施例提供了一种定位声音发出位置的方法流程图,在本发明另一实施例中,在执行步骤S301之前,还可以执行下列步骤:
S401,判断车载系统的至少一个座位是否坐人。
具体的,车载系统可以通过重力感应的方式判断车载系统的座位是否坐人。
例如,通过重力感应的方式判断图2A中的车载系统的座位是否坐人。如判断图2A中的位置①,位置②,位置③或者位置④是否坐人。
当所述车载系统判断所述车载系统的所有座位都没有坐人时,不执行步骤S301。
当所述车载系统判断所述车载系统的至少一个座位坐人时,执行步骤S301。
在采集声音信号之前,首先判断车载系统的至少一个座位上是否坐人,只有在车载系统的座位上坐人,再来定位声音发出的位置,提高声音采集的效率和提高确定声音发出位置的效率。
在本发明另一实施例中,如图3B所示,可以在执行步骤S305后,执行步骤S305a:识别提取的所述M个第二声音信号的声纹。
S305b:测量车载系统的坐人的座位上,用户的重量。
S305c:结合测量的用户的重量和识别的第二声音信号的声纹,确定所述用户的身份。
S305d:根据所述确定的所述用户的身份,确定所述用户发出的第二声音信号对应的语音命令的优先级。
S305e:根据所述用户发出的第二声音信号对应的语音命令的优先级,响应所述第二声音信号对应的语音命令。
通过结合重力感应和声纹识别的方式,确定用户的身份和用户发出的声音信号对应的语音命令的优先级。结合用户发出的声音信号对应的语音命令的优先级,确定响应多个语音命令的优先级。减小由于需要响应多个语音命令,给车载中控设备系统带来的误差和错误。
本发明提供了一种定位声音发出位置的方法,利用波束成型算法,根据位置参数,从K个第一声音信号中提取M个第二声音信号,从而可以确定每个第二声音信号对应的发出位置。更进一步地,设置语音命令的优先级,通过优 先处理高优先级的方法来处理冲突命令,减少车载中控设备响应多个冲突命令带来的冲突,减少响应错误带来的误差,提高了用户体验。
图4是本发明实施例提供了一种终端设备400。可用于执行本发明实施例前述的方法。如图4所示,该终端设备400可以为包括手机、平板电脑、PDA(Personal Digital Assistant,个人数字助理)、POS(Point of Sale,销售终端)、车载中控终端设备等终端设备,终端设备400包括RF(Radio Frequency,射频)电路410、存储器420、输入装置430、显示装置440、传感器450、音频电路460、WiFi(wireless fidelity,无线保真)模块470、处理器480、以及电源490等部件。本领域技术人员可以理解,图4中示出的终端设备结构只做实现方式的举例,并不构成对终端设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。
RF电路410可用于收发信息或通话过程中,信号的接收和发送,特别地,将基站的下行信息接收后,给处理器480处理;另外,将设计上行的数据发送给基站。通常,RF电路410包括但不限于天线、至少一个放大器、收发信机、耦合器、LNA(Low Noise Amplifier,低噪声放大器)、双工器等。此外,RF电路410还可以通过无线通信与网络和其他终端设备通信。所述无线通信可以使用任一通信标准或协议,包括但不限于GSM(Global System of Mobile communication,全球移动通讯系统)、GPRS(General Packet Radio Service,通用分组无线服务)、CDMA(Code Division Multiple Access,码分多址)、WCDMA(Wideband Code Division Multiple Access,宽带码分多址)、LTE(Long Term Evolution,长期演进)、电子邮件、SMS(Short Messaging Service,短消息服务)等。
存储器420可用于存储软件程序以及模块,处理器480通过运行存储在存储器420的软件程序以及模块,从而执行终端设备400的各种功能应用以及数据处理。存储器420可主要包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序(比如声音播放功能、图像播放功能等)等;存储数据区可存储根据终端设备400的使用所创建的数据(比如音频数据、电话本等)等。此外,存储器420可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件、闪存器件、或其他易失性固态存储器件。
显示装置440可用于显示由用户输入的信息或提供给用户的信息以及终端设备400的各种菜单。显示装置440可包括显示面板441,可选的,可以采用LCD(Liquid Crystal Display,液晶显示器)、OLED(Organic Light-Emitting Diode,有机发光二极管)等形式来配置显示面板441。进一步的,触控面板431可覆盖显示面板441,当触控面板431检测到在其上或附近的触摸操作后,传送给处理器480以确定触摸事件的类型,随后处理器480根据触摸事件的类型在显示面板441上提供相应的视觉输出。虽然在图4中,触控面板431与显示面板441是作为两个独立的部件来实现终端设备400的输入和输入功能。但是在某些实施例中,可以将触控面板431与显示面板441集成而实现终端设备400的输入和输出功能,例如,触控面板431与显示面板441可以集成为触摸屏,实现终端设备400的输入和输出功能。
终端设备400还可包括至少一种传感器450,比如光传感器、运动传感器以及其他传感器。具体地,光传感器可包括环境光传感器及接近传感器,其中,环境光传感器可根据环境光线的明暗来调节显示面板441的亮度,接近传感器可在终端设备400移动到耳边时,关闭显示面板441或者背光。作为运动传感器的一种,加速计传感器可检测各个方向上(一般为三轴)加速度的大小,静止时可检测出重力的大小及方向,可用于识别手机姿态的应用(比如横竖屏切换、相关游戏、磁力计姿态校准)、振动识别相关功能(比如计步器、敲击)等;至于终端设备400还可配置的陀螺仪、气压计、湿度计、温度计、红外线传感器等其他传感器,在此不再赘述。
音频电路460、扬声器461,传声器462可提供用户与终端设备400之间的音频接口。音频电路460可将接收到的音频数据转换后的电信号,传输到扬声器461,由扬声器461转换为声音信号输出;另一方面,传声器462将收集的声音信号转换为电信号,由音频电路460接收后转换为音频数据,再将音频数据输出处理器480处理后,经RF电路410以发送给比如另一手机,或者将音频数据输出至存储器420以便进一步处理。
终端设备400通过WiFi模块470可以帮助用户收发电子邮件、浏览网页和访问流式媒体等,它为用户提供了无线的宽带互联网访问。虽然图4示出了WiFi模块470,但是可以理解的是,其并不属于终端设备400的必须构成,完全可以根据需要在不改变发明的本质的范围内而省略。
处理器480是终端设备400的控制中心,利用各种接口和线路连接整个手机的各个部分,通过运行或执行存储在存储器420内的软件程序和/或模块,以及调用存储在存储器420内的数据,执行终端设备400的各种功能和处理数据,从而对终端设备进行整体监控。可选的,处理器480可包括一个或多个处理单元;优选的,处理器480可集成应用处理器和调制解调处理器,其中,应用处理器主要处理操作系统、用户界面和应用程序等,调制解调处理器主要处理无线通信。可以理解的是,上述调制解调处理器也可以不集成到处理器480中。处理器480具体可以为中央处理器(Central Processing Unit,CPU)。
终端设备400还包括给各个部件供电的电源490(比如电池),优选的,电源可以通过电源管理系统与处理器480逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。
在本发明实施例中,终端设备400包括:K个声音采集传感器450和处理器480具有以下功能:
声音采集传感器450,用于采集K个第一声音信号;其中K为大于等于2的整数。
具体的,所述K个声音采集传感器在三维空间内的坐标不同。
处理器480,用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,并确定每个第二声音信号对应的位置,其中M小于等于N,N为大于等于2的整数。
在本发明另一实施例中,所述处理器480用于确定每个第二声音信号对应的位置,具体包括:根据第L个第二声音信号对应的位置参数,确定所述第L个第二声音信号对应的位置L;其中,第L个第二声音信号为所述M个第二声音信号中的任意一个。
在本发明另一实施例中,所述处理器480还用于从所述K个第一声音信号中提取M个第二声音信号后,对提取的所述M个第二声音信号进行语音识别,并用于获取所述M个第二声音信号对应的M个语音命令。
在本发明另一实施例中,终端设备400还包括:输出装置510,所述输出装置510,用于在所述处理器获取所述M个第二声音信号对应的M个语音命令之后,响应所述M个语音命令。
所述输出装置510用于响应所述M个语音命令,具体包括:所述输出装 置用于根据所述M个语音命令对应的M个不同位置的优先级,优先响应优先级高的命令。
在本发明实施例中,所述输出装置510具体可以为音频电路460或者显示装置440。
本发明实施例中,提供了一种定位声音发出位置的方法和终端设备,利用波束成型算法,根据位置参数,从K个第一声音信号中提取M个第二声音信号,从而可以确定每个第二声音信号对应的发出位置,通过这种方法,可以高效的提取不同位置发出的声音信号,提供语音识别能力,为用户提供了更高的用户体验。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元、算法及方法步骤,能够以计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的移动终端和接入终端的分别执行了本发明中所述的方法实施例里的工作过程,具体工作可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,所揭露的服务器和方法,可以通过其它的方式实现。例如,以上所描述的服务器实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本发明实施例方案的目的。
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元 中。
本领域普通技术人员可以理解:实现上述方法实施例的全部或部分步骤可以通过程序命令相关的硬件来完成,前述的程序可以存储于一计算机可读取存储介质中,该程序在执行时,执行包括上述方法实施例的步骤;而前述的存储介质包括:ROM、RAM、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以所述权利要求的保护范围为准。

Claims (13)

  1. 一种定位声音发出位置的方法,其特征在于,所述方法包括:
    采集K个第一声音信号;其中,K为大于等于2的整数;
    根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号;其中M小于等于N,N为大于等于2的整数;
    确定每个第二声音信号对应的位置。
  2. 根据权利要求1所述的方法,其特征在于,所述根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,具体包括:
    利用波束成型算法,分别根据所述N个位置参数,从所述K个第一声音信号中提取M个第二声音信号。
  3. 根据权利要求1或2所述的方法,其特征在于,所述确定每个第二声音信号对应的位置,具体包括:
    根据第L个第二声音信号对应的位置参数,确定所述第L个第二声音信号对应的位置L;其中,第L个第二声音信号为所述M个第二声音信号中的任意一个。
  4. 根据权利要求1-3任一所述的方法,其特征在于,从所述K个第一声音信号中提取M个第二声音信号后,所述方法还包括:
    对提取的所述M个第二声音信号进行语音识别;
    获取所述M个第二声音信号对应的M个语音命令。
  5. 根据权利要求4所述的方法,其特征在于,在获取所述M个第二声音信号对应的M个语音命令之后,所述方法还包括:
    响应所述M个语音命令。
  6. 根据权利要求5所述的方法,其特征在于,所述响应所述M个语音命令包括:根据所述M个语音命令对应的M个不同位置的优先级,优先响应高优先级的语音命令。
  7. 一种终端设备,其特征在于,所述终端设备包括:
    K个声音采集传感器,用于采集K个第一声音信号;其中K为大于等于2的整数;
    处理器,用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,并确定每个第二声音信号对应的位置,其中M小于等于N,N为大于等于2的整数。
  8. 根据权利要求7所述的终端设备,其特征在于,所述处理器用于根据与N个不同位置对应的N个位置参数,从所述K个第一声音信号中提取M个第二声音信号,具体包括:
    所述处理器用于利用波束成型算法,分别根据所述N个位置参数,从所述K个第一声音信号中提取M个第二声音信号。
  9. 根据权利要求7或8所述的终端设备,其特征在于,所述处理器用于确定每个第二声音信号对应的位置,具体包括:
    根据第L个第二声音信号对应的位置参数,确定所述第L个第二声音信号对应的位置L;其中,第L个第二声音信号为所述M个第二声音信号中的任意一个。
  10. 根据权利要求7-9任一所述的终端设备,其特征在于,所述处理器还用于从所述K个第一声音信号中提取M个第二声音信号后,对提取的所述M个第二声音信号进行语音识别,并用于获取所述M个第二声音信号对应的M个语音命令。
  11. 根据权利要求7-10任一所述的终端设备,其特征在于,所述终端设备还包括输出装置;
    所述输出装置,用于在所述处理器获取所述M个第二声音信号对应的M个语音命令之后,响应所述M个语音命令。
  12. 根据权利要求11所述的设备,其特征在于,所述输出装置用于响应所述M个语音命令,具体包括:
    所述输出装置用于根据所述M个语音命令对应的M个不同位置的优先级,优先响应优先级高的命令。
  13. 根据权利要求7-12所述的设备,其特征在于,所述K个声音采集传感器在三维空间内的坐标不同。
PCT/CN2015/079391 2015-05-20 2015-05-20 一种定位声音发出位置的方法和终端设备 WO2016183825A1 (zh)

Priority Applications (6)

Application Number Priority Date Filing Date Title
EP15892204.7A EP3264266B1 (en) 2015-05-20 2015-05-20 Method for positioning sounding location, and terminal device
CN201580076714.3A CN107430524B (zh) 2015-05-20 2015-05-20 一种定位声音发出位置的方法和终端设备
KR1020177030167A KR102098668B1 (ko) 2015-05-20 2015-05-20 발음 위치 및 단말 장치 위치를 결정하는 방법
PCT/CN2015/079391 WO2016183825A1 (zh) 2015-05-20 2015-05-20 一种定位声音发出位置的方法和终端设备
JP2017557075A JP6615227B2 (ja) 2015-05-20 2015-05-20 音声の発生位置を特定するための方法及び端末デバイス
US15/566,979 US10410650B2 (en) 2015-05-20 2015-05-20 Method for locating sound emitting position and terminal device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2015/079391 WO2016183825A1 (zh) 2015-05-20 2015-05-20 一种定位声音发出位置的方法和终端设备

Publications (1)

Publication Number Publication Date
WO2016183825A1 true WO2016183825A1 (zh) 2016-11-24

Family

ID=57319145

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/079391 WO2016183825A1 (zh) 2015-05-20 2015-05-20 一种定位声音发出位置的方法和终端设备

Country Status (6)

Country Link
US (1) US10410650B2 (zh)
EP (1) EP3264266B1 (zh)
JP (1) JP6615227B2 (zh)
KR (1) KR102098668B1 (zh)
CN (1) CN107430524B (zh)
WO (1) WO2016183825A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019069731A1 (ja) * 2017-10-06 2019-04-11 ソニー株式会社 情報処理装置、情報処理方法、プログラム、および移動体

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110556113A (zh) * 2018-05-15 2019-12-10 上海博泰悦臻网络技术服务有限公司 基于声纹识别的车辆控制方法与云端服务器
DE102018212902A1 (de) 2018-08-02 2020-02-06 Bayerische Motoren Werke Aktiengesellschaft Verfahren zum Bestimmen eines digitalen Assistenten zum Ausführen einer Fahrzeugfunktion aus einer Vielzahl von digitalen Assistenten in einem Fahrzeug, computerlesbares Medium, System, und Fahrzeug
US10944588B2 (en) * 2018-11-29 2021-03-09 International Business Machines Corporation Resolving conflicting commands received by an electronic device
US11468886B2 (en) * 2019-03-12 2022-10-11 Lg Electronics Inc. Artificial intelligence apparatus for performing voice control using voice extraction filter and method for the same
CN110297702B (zh) * 2019-05-27 2021-06-18 北京蓦然认知科技有限公司 一种多任务并行处理方法和装置
JP7198741B2 (ja) * 2019-12-27 2023-01-04 本田技研工業株式会社 車両操作権管理装置、車両操作権管理方法及びプログラム
KR20210133600A (ko) * 2020-04-29 2021-11-08 현대자동차주식회사 차량 음성 인식 방법 및 장치
CN111786860B (zh) * 2020-06-29 2022-04-01 广东美的制冷设备有限公司 家电及其控制方法和计算机可读存储介质
CN115503639A (zh) * 2022-10-13 2022-12-23 广州小鹏汽车科技有限公司 语音处理方法、语音交互方法、服务器及存储介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140294195A1 (en) * 2013-03-28 2014-10-02 Jvis-Usa, Llc Speaker system such as a sound bar assembly having improved sound quality
CN104442622A (zh) * 2013-09-25 2015-03-25 现代自动车株式会社 用于车辆的声音控制系统和方法
CN104464739A (zh) * 2013-09-18 2015-03-25 华为技术有限公司 音频信号处理方法及装置、差分波束形成方法及装置
CN104572258A (zh) * 2013-10-18 2015-04-29 通用汽车环球科技运作有限责任公司 用于在车载计算机系统处处理多个音频流的方法和设备

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0418831A (ja) 1990-05-14 1992-01-23 Sony Corp 遠隔制御装置
JP3863306B2 (ja) * 1998-10-28 2006-12-27 富士通株式会社 マイクロホンアレイ装置
JP3715584B2 (ja) 2002-03-28 2005-11-09 富士通株式会社 機器制御装置および機器制御方法
JP4327510B2 (ja) 2003-06-05 2009-09-09 コニカミノルタビジネステクノロジーズ株式会社 リモート操作システム
CN1815556A (zh) 2005-02-01 2006-08-09 松下电器产业株式会社 可利用语音命令操控车辆的方法及系统
US8214219B2 (en) * 2006-09-15 2012-07-03 Volkswagen Of America, Inc. Speech communications system for a vehicle and method of operating a speech communications system for a vehicle
US20090055180A1 (en) * 2007-08-23 2009-02-26 Coon Bradley S System and method for optimizing speech recognition in a vehicle
JP4547721B2 (ja) 2008-05-21 2010-09-22 株式会社デンソー 自動車用情報提供システム
US8141115B2 (en) 2008-12-17 2012-03-20 At&T Labs, Inc. Systems and methods for multiple media coordination
US8660782B2 (en) 2010-03-31 2014-02-25 Denso International America, Inc. Method of displaying traffic information and displaying traffic camera view for vehicle systems
KR101987966B1 (ko) 2012-09-03 2019-06-11 현대모비스 주식회사 차량용 어레이 마이크의 음성 인식 향상 시스템 및 그 방법
TWI598774B (zh) * 2013-10-25 2017-09-11 和冠股份有限公司 電磁書寫單元及兼具墨水與電磁書寫功能的電磁式手寫筆
US20160012827A1 (en) * 2014-07-10 2016-01-14 Cambridge Silicon Radio Limited Smart speakerphone
US20160080861A1 (en) * 2014-09-16 2016-03-17 Toyota Motor Engineering & Manufacturing North America, Inc. Dynamic microphone switching
DE102015220400A1 (de) * 2014-12-11 2016-06-16 Hyundai Motor Company Sprachempfangssystem im fahrzeug mittels audio-beamforming und verfahren zum steuern desselben
US10304463B2 (en) * 2016-10-03 2019-05-28 Google Llc Multi-user personalization at a voice interface device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140294195A1 (en) * 2013-03-28 2014-10-02 Jvis-Usa, Llc Speaker system such as a sound bar assembly having improved sound quality
CN104464739A (zh) * 2013-09-18 2015-03-25 华为技术有限公司 音频信号处理方法及装置、差分波束形成方法及装置
CN104442622A (zh) * 2013-09-25 2015-03-25 现代自动车株式会社 用于车辆的声音控制系统和方法
CN104572258A (zh) * 2013-10-18 2015-04-29 通用汽车环球科技运作有限责任公司 用于在车载计算机系统处处理多个音频流的方法和设备

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3264266A4 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019069731A1 (ja) * 2017-10-06 2019-04-11 ソニー株式会社 情報処理装置、情報処理方法、プログラム、および移動体

Also Published As

Publication number Publication date
EP3264266B1 (en) 2020-08-05
KR20170129249A (ko) 2017-11-24
CN107430524A (zh) 2017-12-01
US20180108368A1 (en) 2018-04-19
EP3264266A1 (en) 2018-01-03
EP3264266A4 (en) 2018-03-28
KR102098668B1 (ko) 2020-04-08
CN107430524B (zh) 2020-10-27
JP2018524620A (ja) 2018-08-30
JP6615227B2 (ja) 2019-12-04
US10410650B2 (en) 2019-09-10

Similar Documents

Publication Publication Date Title
WO2016183825A1 (zh) 一种定位声音发出位置的方法和终端设备
US10834237B2 (en) Method, apparatus, and storage medium for controlling cooperation of multiple intelligent devices with social application platform
US11314389B2 (en) Method for presenting content based on checking of passenger equipment and distraction
US10183680B2 (en) Mobile terminal and method for controlling application for vehicle
US9743222B2 (en) Method for controlling and an electronic device thereof
KR102301880B1 (ko) 전자 장치 및 이의 음성 대화 방법
US9578668B2 (en) Bluetooth pairing system and method
US20170243578A1 (en) Voice processing method and device
US20150281430A1 (en) Method and apparatus for providing information based on movement of an electronic device
CN109672775B (zh) 调节唤醒灵敏度的方法、装置及终端
KR20170061489A (ko) 전자 기기 및 이의 운송 기기 제어 방법
CN108068846B (zh) 一种地铁乘车调度方法及移动终端
CN113314120B (zh) 处理方法、处理设备及存储介质
KR102611775B1 (ko) 그룹 메시지의 전달 방법 및 이를 수행하는 전자 장치
CN108702410B (zh) 一种情景模式控制方法及移动终端
CN113254092B (zh) 处理方法、设备及存储介质
EP3945706A1 (en) Systems and methods for bluetooth authentication using communication fingerprinting
CN113742027A (zh) 交互方法、智能终端及可读存储介质
CN112738730A (zh) 搜救定位方法、装置及存储介质
CN114021002A (zh) 信息展示方法、移动终端及存储介质
CN114974246A (zh) 处理方法、智能终端及存储介质
CN114201584A (zh) 模板校验方法及相关装置
CN113296678A (zh) 处理方法、移动终端及存储介质
CN116682423A (zh) 一种语音意图匹配方法、装置、智能座舱和电子设备
CN109874124A (zh) 用于无线通信抑制的方法和设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15892204

Country of ref document: EP

Kind code of ref document: A1

REEP Request for entry into the european phase

Ref document number: 2015892204

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 15566979

Country of ref document: US

ENP Entry into the national phase

Ref document number: 20177030167

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2017557075

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE