CN109817214B - Interaction method and device applied to vehicle - Google Patents

Interaction method and device applied to vehicle Download PDF

Info

Publication number
CN109817214B
CN109817214B CN201910184912.1A CN201910184912A CN109817214B CN 109817214 B CN109817214 B CN 109817214B CN 201910184912 A CN201910184912 A CN 201910184912A CN 109817214 B CN109817214 B CN 109817214B
Authority
CN
China
Prior art keywords
playing
target
voice
user
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910184912.1A
Other languages
Chinese (zh)
Other versions
CN109817214A (en
Inventor
彭汉迎
欧阳能钧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Apollo Zhilian Beijing Technology Co Ltd
Original Assignee
Apollo Zhilian Beijing Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Apollo Zhilian Beijing Technology Co Ltd filed Critical Apollo Zhilian Beijing Technology Co Ltd
Priority to CN202111119722.5A priority Critical patent/CN113763956A/en
Priority to CN201910184912.1A priority patent/CN109817214B/en
Priority to CN202111120001.6A priority patent/CN113763957A/en
Publication of CN109817214A publication Critical patent/CN109817214A/en
Application granted granted Critical
Publication of CN109817214B publication Critical patent/CN109817214B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R16/00Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for
    • B60R16/02Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements
    • B60R16/037Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements for occupant comfort, e.g. for automatic adjustment of appliances according to personal settings, e.g. seats, mirrors, steering wheel
    • B60R16/0373Voice control
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/21Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being power information

Abstract

The embodiment of the application discloses an interaction method and device applied to a vehicle. One embodiment of the method comprises: if a user awakening voice is received, determining a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user awakening voice received by the at least two sound pickups; and receiving first user voice through the target sound pickup, and sending a playing instruction for playing the multimedia file to the target playing device for playing in response to the fact that the first user voice comprises information for indicating the identification of the multimedia file. The method provided by the embodiment of the application can not only receive the voice of the driver position, but also receive the voice of other positions such as the rear row position, so that the passenger at the rear row position can independently perform voice interaction with the vehicle, see multimedia content and well realize that other people in the vehicle are not influenced.

Description

Interaction method and device applied to vehicle
Technical Field
The embodiment of the application relates to the technical field of computers, in particular to the technical field of internet, and particularly relates to an interaction method and device applied to vehicles.
Background
In a conventional car-mounted voice system, which serves only one person around a driver, a microphone for receiving voice is also provided at the front end of the vehicle interior. Also, conventional navigation, music, telephone, etc. are designed around the driver.
Therefore, for passengers in the vehicle interior, the vehicle-mounted voice system is difficult to use, voice interaction with the vehicle is difficult, and multimedia playing cannot be performed.
Disclosure of Invention
The embodiment of the application provides an interaction method and device applied to a vehicle.
In a first aspect, an embodiment of the present application provides an interaction method applied to a vehicle, including: if the user awakening voice is received, determining a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user awakening voice received by the at least two sound pickups; and receiving the first user voice through the target sound pickup, and sending a playing instruction for playing the multimedia file to the target playing device for playing in response to the fact that the first user voice comprises the information for indicating the identification of the multimedia file.
In some embodiments, after sending a play instruction to the target playback device to play the multimedia file, the method further includes: receiving a second user voice through the target sound pickup, and adjusting the playing state of the target playing apparatus using the playing adjustment information in response to determining that the second user voice includes adjustment information for indicating an adjustment of the playing state.
In some embodiments, in response to determining that the second user speech includes adjustment information for indicating an adjustment to the play state, adjusting the play state of the target playback device using the play adjustment information includes: determining whether the target playback device is in playback in response to determining that the second user speech includes adjustment information indicating an adjustment to a playback state; in response to determining that the target playback device is playing, the playback state of the target playback device is adjusted using the playback adjustment information.
In some embodiments, determining a target microphone and a target playing device corresponding to a user awakening voice from at least two microphones and at least two multimedia playing devices based on voice signal amplitudes of the user awakening voice received by the at least two microphones includes: determining a target sound pickup from the at least two sound pickups based on the amplitude of the voice signal of the user wake-up voice received by the at least two sound pickups; and determining a target playing device corresponding to the target sound pickup.
In some embodiments, the rear row sound pick-up in the at least two sound pick-up devices is used for receiving the voice of the rear row of the vehicle, and the rear row playing device in the at least two multimedia playing devices is used for playing the multimedia file to the rear row of the vehicle; receiving a first user voice through a target microphone, comprising: if the target sound pickup is a back row sound pickup, receiving the first user voice through the back row sound pickup; and playing the multimedia file by using the target playing device based on the file address, comprising: and if the target playing device is a back-row multimedia playing device, playing the multimedia file by using the back-row multimedia playing device based on the file address.
In a second aspect, the present application provides an interaction device applied to a vehicle, where the vehicle includes at least two microphones and at least two multimedia playing devices for an interior of the vehicle, and the device includes: the determining unit is configured to determine a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user awakening voice received by the at least two sound pickups if the user awakening voice is received; and the sending unit is configured to receive the first user voice through the target sound pickup, and in response to determining that the first user voice comprises the information indicating the identification of the multimedia file, send a playing instruction indicating the playing of the multimedia file to the target playing device for playing.
In some embodiments, the apparatus further comprises: an adjusting unit configured to receive the second user voice through the target sound collector, and adjust the play state of the target play device using the play adjustment information in response to determining that the second user voice includes adjustment information indicating an adjustment of the play state.
In some embodiments, the adjusting unit is further configured to perform adjusting the play state of the target playback device using the play adjustment information in response to determining that the second user voice includes adjustment information indicating to adjust the play state, as follows: determining whether the target playback device is in playback in response to determining that the second user speech includes adjustment information indicating an adjustment to a playback state; in response to determining that the target playback device is playing, the playback state of the target playback device is adjusted using the playback adjustment information.
In some embodiments, the determining unit is further configured to determine, from the at least two microphones and the at least two multimedia playing apparatuses, a target microphone and a target playing apparatus corresponding to the user-awakening voice, based on the voice signal amplitude of the user-awakening voice received by the at least two microphones, as follows: determining a target sound pickup from the at least two sound pickups based on the amplitude of the voice signal of the user wake-up voice received by the at least two sound pickups; and determining a target playing device corresponding to the target sound pickup.
In some embodiments, the rear row sound pick-up in the at least two sound pick-up devices is used for receiving the voice of the rear row of the vehicle, and the rear row playing device in the at least two multimedia playing devices is used for playing the multimedia file to the rear row of the vehicle; a transmitting unit further configured to perform receiving the first user voice through the target microphone as follows: if the target sound pickup is a back row sound pickup, receiving the first user voice through the back row sound pickup; and playing the multimedia file by using the target playing device based on the file address, comprising: and if the target playing device is a back-row multimedia playing device, playing the multimedia file by using the back-row multimedia playing device based on the file address.
In a third aspect, the embodiment of the present application provides an interactive system applied to a vehicle, where the system includes a processor, at least two sound collectors for an interior of the vehicle, and at least two multimedia playing devices; at least two sound pick-up devices for receiving user voice; a processor for performing the method of the first aspect; and the target playing device is used for playing the multimedia file.
In a fourth aspect, an embodiment of the present application provides an electronic device, including: one or more processors; a storage device for storing one or more programs which, when executed by one or more processors, cause the one or more processors to implement a method as in any embodiment of the interaction method as applied to the vehicle.
In a fifth aspect, embodiments of the present application provide a computer-readable storage medium, on which a computer program is stored, which when executed by a processor, implements the method of any of the embodiments as applied to an interaction method for a vehicle.
According to the interaction scheme applied to the vehicle, firstly, if the user awakening voice is received, the target sound pickup and the target playing device corresponding to the user awakening voice are determined from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user voice received by the at least two sound pickups. And then, receiving the voice of the first user through the target sound pickup, and if the sentence corresponding to the voice of the first user is determined to comprise the identification of the multimedia file, determining the file address of the multimedia file based on the identification. And finally, playing the multimedia file by using the target playing device based on the file address. The scheme provided by the embodiment of the application can not only receive the voice of the position of a driver, but also can receive the voice of other positions and even the back row, so that passengers at other positions can independently perform voice interaction with the vehicle, receive and see multimedia content and can better realize that other people in the vehicle are not influenced.
Drawings
Other features, objects and advantages of the present application will become more apparent upon reading of the following detailed description of non-limiting embodiments thereof, made with reference to the accompanying drawings in which:
FIG. 1 is an exemplary system architecture diagram in which the present application may be applied;
FIG. 2 is a flow chart of one embodiment of an interaction method applied to a vehicle according to the present application;
FIG. 3 is a schematic diagram of an application scenario of an interaction method applied to a vehicle according to the present application;
FIG. 4 is a flow chart of yet another embodiment of an interaction method applied to a vehicle according to the present application;
FIG. 5 is a schematic structural diagram of one embodiment of an interaction device applied to a vehicle according to the present application;
FIG. 6 is a schematic structural diagram of one embodiment of an interactive system for a vehicle according to the present application;
FIG. 7 is a block diagram of a computer system suitable for use in implementing the electronic device of an embodiment of the present application.
Detailed Description
The present application will be described in further detail with reference to the following drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the relevant invention and not restrictive of the invention. It should be noted that, for convenience of description, only the portions related to the related invention are shown in the drawings.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
Fig. 1 shows an exemplary system architecture 100 to which embodiments of the vehicle-applied interaction method or vehicle-applied interaction device of the present application may be applied.
As shown in fig. 1, system architecture 100 may include a vehicle 101, a network 102, and a server 103. Network 102 is the medium used to provide a communication link between vehicle 101 and server 103. Network 102 may include various connection types, such as wired, wireless communication links, or fiber optic cables, to name a few.
The user may use the vehicle 101 to interact with the server 103 over the network 102 to receive or send messages or the like. The vehicle 101 may have installed thereon various communication client applications, such as an interactive application applied to the vehicle, a video-type application, a live application, an instant messaging tool, a mailbox client, social platform software, and the like.
Here, the vehicle 101 may be hardware or software. When the vehicle 101 is hardware, it may be various electronic devices with a display screen, including but not limited to smart phones, tablet computers, e-book readers, laptop portable computers, desktop computers, and the like. When the vehicle 101 is software, it can be installed in the electronic devices listed above. It may be implemented as multiple pieces of software or software modules (e.g., multiple pieces of software or software modules to provide distributed services) or as a single piece of software or software module. And is not particularly limited herein.
The server 103 may be a server that provides various services, such as a background server that provides support for the vehicle 101. The background server can analyze and process the received data such as the user awakening voice and the first user voice, and feed back the processing result to the terminal equipment.
It should be noted that the interaction method applied to the vehicle provided in the embodiment of the present application may be executed by the server 103 or the vehicle 101, and accordingly, the interaction device applied to the vehicle may be disposed in the server 103 or the vehicle 101.
It should be understood that the number of terminal devices, networks, and servers in fig. 1 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation.
With continued reference to FIG. 2, a flow 200 of one embodiment of an interaction method applied to a vehicle according to the present application is shown. The interaction method applied to the vehicle comprises the following steps:
step 201, if a user awakening voice is received, determining a target sound pickup and a target playing device corresponding to the user awakening voice from at least two sound pickups and at least two multimedia playing devices based on the voice signal amplitude of the user awakening voice received by the at least two sound pickups.
In this embodiment, the vehicle includes at least two microphones for the vehicle interior and at least two multimedia players for the vehicle interior. An executing subject (for example, a processor in the vehicle shown in fig. 1) of the interaction method applied to the vehicle may determine, when the user-awakening voice is received, a target microphone and a target playback device corresponding to the user-awakening voice based on the voice signal amplitude (An) of the user-awakening voice received by each microphone of the at least two microphones of the vehicle. The multimedia file may be at least one of audio, video, and the like. The multimedia playing device may be a display, a loudspeaker for outputting audio, and the like. Here, the user wake-up voice is a voice including information indicating a preset wake-up word. The preset awakening words are preset words which can be used for awakening the execution main body to carry out interaction. Such as "hello, zhang san". And if the received voice is determined to include the information indicating the preset awakening word, determining that the user awakening voice is received. The user speaks the user awakening voice at a certain seat in the car, and the sound pickup mainly used for receiving the voice of the seat can be used as a target sound pickup of the user awakening voice, namely the awakened target sound pickup.
The vehicle in this embodiment includes a sound pickup whose main pickup object is a passenger on a first row seat of the vehicle, and further includes a sound pickup whose main pickup object is a passenger on a rear row seat of the vehicle. The rear vehicle row or rear vehicle seat here may refer to a seat behind the first vehicle seat row. Each seat in the car can correspond to one sound pick-up and one multimedia playing device, or each row of seats corresponds to two sound pick-ups. For example, four seats are provided in the car, and four microphones may be provided in the car. In the car inner room, the position of each sound pick-up is convenient for it to receive the pronunciation of main pickup object. The setting position of the playing device is convenient for the corresponding user to watch. For example, a second row of microphones (rear row microphones) and a second row of playback devices (rear row playback devices) may be disposed on the side of the back of the first row seat facing the second row.
The execution main body can determine the target sound pickup and the target playing device corresponding to the user awakening voice in various modes. For example, the execution subject may determine the amplitude of the voice received by each sound pickup, set the sound pickup that receives the voice with the largest amplitude as the target sound pickup, and set the multimedia playing apparatus that displays the voice to the seat corresponding to the target sound pickup as the target playing apparatus.
In some optional implementations of this embodiment, step 201 may include:
determining a target sound pickup from the at least two sound pickups based on the amplitude of the voice signal of the user awakening voice received by the at least two sound pickups; and determining a target playing device corresponding to the target sound pickup.
In these alternative implementations, the executing body may determine the target microphone based on the amplitude of the voice signal of the user wake-up voice received by the at least two microphones. And determines a target playback device corresponding to the target sound pickup. Specifically, the target pickup may be determined in a variety of ways. The sound pickup receiving the largest amplitude may be taken as the target sound pickup. It is also possible to determine the microphone that receives the largest amplitude and the next amplitude, and assume these two microphones to be the target microphones, respectively. In the case where one of the sound pickup is a target sound pickup, the amplitude of the other sound pickup after the positional deviation is eliminated is calculated from the isolation. The difference between the amplitude of the audio received by one of the microphones and the amplitude of the audio with the position offset removed by the other microphone is determined for each of the assumed cases. And the target sound pickup assumed as a result of obtaining a small difference is taken as the target sound pickup.
These implementations may first accurately determine the pickup that was awakened, and then determine the playback device that played the multimedia file to the user. Then the user can interact with the vehicle by voice through the sound pick-up and watch the multimedia content by using the playing device.
Step 202, receiving the first user voice through the target sound pickup, and in response to determining that the first user voice includes information indicating an identifier of the multimedia file, sending a play instruction indicating to play the multimedia file to the target playing device for playing.
In this embodiment, the target sound pickup may collect a first user voice and transmit the first user voice to the execution main body, so that the execution main body receives the first user voice. In response to determining that the first user speech includes information indicating an identifier of a multimedia file, a play instruction indicating to play the multimedia file may be sent to the target playback device so that the target playback device may play the multimedia file. In practice, the target playback device may play back in real time, or may play back after downloading the multimedia file through the download address.
The execution main body can determine the identifier through the information of the indication identifier, and acquire the multimedia file by using the identifier. Specifically, the execution body may determine a file address of the multimedia file based on the identification. The executing agent may then use the identifier to search for the file address of the multimedia file locally or on the internet. In addition, the corresponding multimedia file can be found locally by directly utilizing the identification. The identifier is capable of indicating a multimedia file for distinguishing the contents of different multimedia files. For example, the identity may be a name, abbreviation, or part of a name. The information indicating the identity may be the identity itself, or identify a corresponding code or character, etc. The file address here may be a play address for real-time play or a download address for downloading a multimedia file.
In some optional implementations of this embodiment, a rear row sound pickup of the at least two sound pickups is configured to receive a voice of a rear row of the vehicle, and a rear row playing device of the at least two multimedia playing devices is configured to play a multimedia file to the rear row of the vehicle; receiving the first user voice through the target microphone in step 202 may include: if the target sound pickup is a back row sound pickup, receiving the first user voice through the back row sound pickup; and step 203 may comprise: and if the target playing device is a back-row multimedia playing device, playing the multimedia file by using the back-row multimedia playing device based on the file address.
In these alternative implementations, the execution main body may receive the first user voice through the rear line microphone in a case where the target microphone is determined to be the rear line microphone. The execution main body can play the multimedia file by using the back-row multimedia playing device under the condition that the target playing device is the back-row multimedia playing device. Therefore, the user sitting on the rear row of the vehicle can wake up the multimedia playing device corresponding to the position of the user and watch the multimedia content played by the playing device.
With continued reference to fig. 3, fig. 3 is a schematic diagram of an application scenario of the interaction method applied to the vehicle according to the present embodiment. In the application scenario of fig. 3, if the execution main body 301 receives the user wake-up voice 302, based on the voice signal amplitudes 303 of the user wake-up voice 302 received by at least two microphones, a target microphone 304 and a target playback device 305 corresponding to the user wake-up voice 302 are determined from the at least two microphones and the at least two multimedia playback devices; the first user speech 306 is received by the target microphone 304 and, in response to determining that the first user speech 306 includes information 307 indicating an identification of a multimedia file, a play instruction 308 is sent to the target playback device to play back the multimedia file.
The method provided by the embodiment of the application can not only receive the voice of the position of the driver, but also receive the voice of other positions such as the rear row position, so that the passenger at the rear row position can independently perform voice interaction with the vehicle, see multimedia content and well realize that other people in the vehicle are not influenced.
With further reference to fig. 4, a flow 400 of yet another embodiment of an interaction method applied to a vehicle is shown. The flow 400 of the interaction method applied to the vehicle comprises the following steps:
step 401, if a user awakening voice is received, determining a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude of the user awakening voice received by the at least two sound pickups.
In this embodiment, the vehicle includes at least two sound collectors and at least two multimedia playing devices. Wherein the at least two microphones may comprise a front row microphone and a rear row microphone. The at least two multimedia playing devices may include a front playing device and a rear playing device. The back row sound pick-up in at least two sound pick-up is used for receiving the pronunciation of vehicle back row, and the back row play device in at least two multimedia play devices is used for playing the multimedia file to the vehicle back row. An executing subject (for example, a processor in the vehicle shown in fig. 1) of the interaction method applied to the vehicle may determine, when the user-awakening voice is received, a target microphone and a target playback device corresponding to the user-awakening voice based on the voice signal amplitude (An) of the user-awakening voice received by each microphone of the at least two microphones of the vehicle. The multimedia file may be at least one of audio, video, and the like. The multimedia playing device may be a display, a loudspeaker for outputting audio, and the like.
Step 402, receiving a first user voice through a target sound pickup, and in response to determining that the first user voice includes information indicating an identifier of a multimedia file, sending a play instruction indicating to play the multimedia file to a target playing device for playing.
In this embodiment, the target sound pickup may collect a first user voice and transmit the first user voice to the execution main body, so that the execution main body receives the first user voice. In response to determining that the first user speech includes information indicating an identifier of a multimedia file, a play instruction indicating to play the multimedia file may be sent to the target playback device so that the target playback device may play the multimedia file. In practice, the target playback device may play back in real time, or may play back after downloading the multimedia file through the download address.
In step 403, a second user voice is received through the target microphone, and the playing state of the target playing device is adjusted using the playing adjustment information in response to determining that the second user voice includes adjustment information for indicating an adjustment of the playing state.
In this embodiment, the executing entity may determine the target sound pickup and receive the second user voice through the target sound pickup after the target sound pickup is awakened. If the sentence corresponding to the second user voice is determined to include the adjustment information, the playing state of the target playing device may be adjusted by using the playing adjustment information. After voice recognition is carried out on the audio, a sentence corresponding to the audio can be obtained. The adjusting information is a word and a sentence which appear in the sentence and are about adjusting the playing, and the adjusting information can be a complete sentence or a part of the sentence. For example, the statement may be "fast forward" or "fast forward video". The play adjustment information in both statements is "fast forward". The playing status can be presented in various ways, such as playing progress, playing volume, whether to play in full screen, and the like.
In practice, the adjustment information may be any of various contents for adjusting the playing state, such as playing, pausing, stopping, fast-forwarding, fast-rewinding, volume-up or full-screen playing, etc.
In some optional implementations of this embodiment, the adjusting the playing state of the target playing apparatus in response to the second user voice including the adjustment information for instructing to adjust the playing state, using the playing adjustment information, may include:
judging whether the target playback apparatus is in playback in response to the second user voice received through the target sound collector including adjustment information indicating an adjustment of a playback state; in response to determining that the target playback device is playing, the playback state of the target playback device is adjusted using the playback adjustment information.
In these optional implementations, the executing entity may need to determine whether the target playback device is playing before adjusting the playback status each time, and if so, may adjust the playback status of the target playback device.
These implementations can play back when it is determined that the playback apparatus is playing back, thereby reducing the execution of ineffective adjustments of the playback apparatus by the main body.
In this embodiment, the user can utilize pronunciation and target adapter to interact to adjust target play device's broadcast state, and then richened the passenger of sitting in each seat and the interactive mode of on-vehicle system, also let the accommodation process more convenient.
With further reference to fig. 5, as an implementation of the method shown in the above figures, the present application provides an embodiment of an interaction device applied to a vehicle, which corresponds to the method embodiment shown in fig. 2, and which can be applied to various electronic devices.
As shown in fig. 5, the interaction device 500 applied to a vehicle of the present embodiment includes: a determination unit 501 and a transmission unit 502. The determining unit 501 is configured as a determining unit, and configured to determine, if a user awakening voice is received, a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on voice signal amplitudes of the user awakening voice received by the at least two sound pickups; a sending unit 502 configured to receive the first user voice through the target sound pickup, and in response to determining that the first user voice includes information indicating an identification of the multimedia file, send a play instruction indicating to play the multimedia file to the target playing device for playing.
In some embodiments, the determining unit 501 of the interaction device 500 applied to the vehicle may determine, when the user-awakening voice is received, the target microphone and the target playback device corresponding to the user-awakening voice based on the voice signal amplitude of the user-awakening voice received by each microphone of the vehicle. The multimedia file may be at least one of audio, video, and the like. The multimedia playing device may be a display, a loudspeaker for outputting audio, and the like.
In some embodiments, the transmitting unit 502 may receive the first user speech. In response to determining that the first user speech includes information indicating an identifier of a multimedia file, a play instruction indicating to play the multimedia file may be sent to the target playback device so that the target playback device may play the multimedia file. In practice, the target playback device may play back in real time, or may play back after downloading the multimedia file through the download address.
In some optional implementations of this embodiment, the apparatus further includes: an adjusting unit configured to receive the second user voice through the target sound collector, and adjust the play state of the target play device using the play adjustment information in response to determining that the second user voice includes adjustment information indicating an adjustment of the play state.
In some optional implementations of this embodiment, the adjusting unit is further configured to perform adjusting the playing state of the target playing device using the playing adjustment information in response to determining that the second user voice includes the adjustment information for indicating to adjust the playing state, as follows: determining whether the target playback device is in playback in response to determining that the second user speech includes adjustment information indicating an adjustment to a playback state; in response to determining that the target playback device is playing, the playback state of the target playback device is adjusted using the playback adjustment information.
In some optional implementations of this embodiment, the determining unit is further configured to determine, based on the voice signal amplitude of the user-awakening voice received by the at least two microphones, a target microphone and a target playing apparatus corresponding to the user-awakening voice from the at least two microphones and the at least two multimedia playing apparatuses, as follows: determining a target sound pickup from the at least two sound pickups based on the amplitude of the voice signal of the user wake-up voice received by the at least two sound pickups; and determining a target playing device corresponding to the target sound pickup.
In some optional implementations of this embodiment, a rear row sound pickup of the at least two sound pickups is configured to receive a voice of a rear row of the vehicle, and a rear row playing device of the at least two multimedia playing devices is configured to play a multimedia file to the rear row of the vehicle; a transmitting unit further configured to perform receiving the first user voice through the target microphone as follows: if the target sound pickup is a back row sound pickup, receiving the first user voice through the back row sound pickup; and playing the multimedia file by using the target playing device based on the file address, comprising: and if the target playing device is a back-row multimedia playing device, playing the multimedia file by using the back-row multimedia playing device based on the file address.
As shown in fig. 6, as an implementation of the methods shown in the above figures, the present application provides an interactive system applied to a vehicle, the system includes a processor, at least two sound collectors for an interior of the vehicle, and at least two multimedia playing devices, the processor is used for executing the methods shown in the above figures;
the at least two sound pickups are used for receiving user awakening voice;
the processor for performing the method according to any of the above embodiments;
the target playing device is used for playing the multimedia file.
The system provided by the embodiment of the application can not only receive the voice of the position of the driver, but also receive the voice of other positions such as the rear row, so that the passengers at the rear row can independently perform voice interaction with the vehicle, see multimedia content and well realize that other people in the vehicle are not influenced.
As shown in fig. 7, electronic device 700 may include a processing means (e.g., central processing unit, graphics processor, etc.) 701 that may perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM)702 or a program loaded from storage 708 into a Random Access Memory (RAM) 703. In the RAM703, various programs and data necessary for the operation of the electronic apparatus 700 are also stored. The processing device 701, the ROM 702, and the RAM703 are connected to each other by a bus 704. An input/output (I/O) interface 705 is also connected to bus 704.
Generally, the following devices may be connected to the I/O interface 705: including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone 706; an output device 707; storage 708 including, for example, magnetic tape, hard disk, etc.; and a communication device 709. The communication means 709 may allow the electronic device 700 to communicate wirelessly or by wire with other devices to exchange data. While fig. 7 illustrates an electronic device 700 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may alternatively be implemented or provided. Each block shown in fig. 7 may represent one device or may represent multiple devices as desired. It should be noted that, if the electronic device is a terminal device, the input device 706 may further include an accelerometer, a gyroscope, or the like; the output devices 707 may include, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, etc.
In particular, according to an embodiment of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such embodiments, the computer program may be downloaded and installed from a network via the communication means 709, or may be installed from the storage means 708, or may be installed from the ROM 702. The computer program, when executed by the processing device 701, performs the above-described functions defined in the methods of embodiments of the present disclosure. It should be noted that the computer readable medium of the embodiments of the present disclosure may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In embodiments of the disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In embodiments of the present disclosure, however, a computer readable signal medium may comprise a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present application may be implemented by software or hardware. The described units may also be provided in a processor, and may be described as: a processor includes a determination unit and a transmission unit. Where the names of these units do not constitute a limitation on the units themselves in some cases, for example, the sending unit may also be described as "a unit that receives a first user voice through a target sound pickup, sends a play instruction instructing to play a multimedia file to a target playback apparatus for playback in response to determining that the first user voice includes information indicating an identification of the multimedia file".
As another aspect, the present application also provides a computer-readable medium, which may be contained in the apparatus described in the above embodiments; or may be present separately and not assembled into the device. The computer readable medium carries one or more programs which, when executed by the apparatus, cause the apparatus to: if the user awakening voice is received, determining a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user awakening voice received by the at least two sound pickups; and receiving the first user voice through the target sound pickup, and sending a playing instruction for playing the multimedia file to the target playing device for playing in response to the fact that the first user voice comprises the information for indicating the identification of the multimedia file.
The above description is only a preferred embodiment of the application and is illustrative of the principles of the technology employed. It will be appreciated by those skilled in the art that the scope of the invention herein disclosed is not limited to the particular combination of features described above, but also encompasses other arrangements formed by any combination of the above features or their equivalents without departing from the spirit of the invention. For example, the above features may be replaced with (but not limited to) features having similar functions disclosed in the present application.

Claims (11)

1. An interaction method applied to a vehicle including at least two microphones and at least two multimedia playing devices for an interior compartment of the vehicle, one microphone for each seat in the vehicle and one multimedia playing device, the method comprising:
if a user awakening voice is received, determining a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user awakening voice received by the at least two sound pickups;
receiving first user voice through the target sound pickup, and responding to the fact that the first user voice comprises information used for indicating identification of a multimedia file, sending a playing instruction for indicating playing of the multimedia file to the target playing device to play;
the determining, from the at least two microphones and the at least two multimedia playing devices, a target microphone and a target playing device corresponding to a user awakening voice based on the voice signal amplitude of the user awakening voice received by the at least two microphones includes:
determining a target sound pickup from the at least two sound pickups based on the voice signal amplitude of the user wake-up voice received by the at least two sound pickups; determining a target playing device corresponding to the target sound pickup;
the determining step of the target sound pickup includes:
determining the sound pick-up which receives the maximum amplitude and the next amplitude, and determining the difference value of the audio amplitude received by the sound pick-up and the audio amplitude of the other sound pick-up after the position deviation is eliminated for each sound pick-up in the two sound pick-ups;
the pickup indicated by the smaller difference is taken as the target pickup.
2. The method of claim 1, wherein after the sending a playback instruction to the target playback device to play back the multimedia file, the method further comprises:
receiving a second user voice through the target sound pickup, and adjusting a playing state of the target playing device by using the playing adjustment information in response to determining that the second user voice includes adjustment information for indicating an adjustment of the playing state.
3. The method of claim 2, wherein said adjusting the play state of the target playback device using the play adjustment information in response to determining that the second user speech includes adjustment information indicating an adjustment to the play state comprises:
in response to determining that the second user speech includes adjustment information indicating an adjustment to a play state, determining whether the target playback device is playing;
and in response to determining that the target playing device is playing, adjusting the playing state of the target playing device by using the playing adjustment information.
4. The method of claim 1, wherein a rear row microphone of the at least two microphones is used for receiving voice of a rear row of a vehicle, and a rear row playing device of the at least two multimedia playing devices is used for playing a multimedia file to the rear row of the vehicle;
the receiving, by the target microphone, a first user voice includes:
if the target sound pickup is a back row sound pickup, receiving the first user voice through the back row sound pickup; and
the method further comprises the following steps: determining a file address of the multimedia file based on the identifier of the multimedia file;
playing the multimedia file by using the target playing device based on the file address, comprising:
and if the target playing device is a back-row multimedia playing device, playing the multimedia file by using the back-row multimedia playing device based on the file address.
5. An interactive device for application to a vehicle, the vehicle including at least two microphones and at least two multimedia playing devices for an interior compartment of the vehicle, one microphone for each seat in the vehicle and one multimedia playing device, the device comprising:
the determining unit is configured to determine a target sound pickup and a target playing device corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing devices based on the voice signal amplitude values of the user awakening voice received by the at least two sound pickups if the user awakening voice is received;
a sending unit configured to receive a first user voice through the target sound pickup, and in response to determining that the first user voice includes information indicating an identifier of a multimedia file, send a play instruction indicating to play the multimedia file to the target playing device for playing;
the determining unit is further configured to perform the following steps of determining a target sound pickup and a target playing apparatus corresponding to the user awakening voice from the at least two sound pickups and the at least two multimedia playing apparatuses based on the voice signal amplitude of the user awakening voice received by the at least two sound pickups:
determining a target sound pickup from the at least two sound pickups based on the voice signal amplitude of the user wake-up voice received by the at least two sound pickups; determining a target playing device corresponding to the target sound pickup;
the determining step of the target sound pickup includes:
determining the sound pick-up which receives the maximum amplitude and the next amplitude, and determining the difference value of the audio amplitude received by the sound pick-up and the audio amplitude of the other sound pick-up after the position deviation is eliminated for each sound pick-up in the two sound pick-ups;
the pickup indicated by the smaller difference is taken as the target pickup.
6. The apparatus of claim 5, wherein the apparatus further comprises:
an adjusting unit configured to receive a second user voice through the target sound collector, and adjust a play state of the target playback device using the play adjustment information in response to determining that the second user voice includes adjustment information indicating an adjustment of a play state.
7. The apparatus of claim 6, wherein the adjusting unit is further configured to perform the adjusting the play state of the target playback apparatus using the play adjustment information in response to determining that the second user voice includes adjustment information indicating an adjustment to the play state, as follows:
in response to determining that the second user speech includes adjustment information indicating an adjustment to a play state, determining whether the target playback device is playing;
and in response to determining that the target playing device is playing, adjusting the playing state of the target playing device by using the playing adjustment information.
8. The apparatus of claim 5, wherein a rear row microphone of the at least two microphones is configured to receive a voice of a rear row of a vehicle, and a rear row playing device of the at least two multimedia playing devices is configured to play a multimedia file to the rear row of the vehicle;
the transmitting unit is further configured to perform the receiving of the first user voice through the target sound collector as follows:
if the target sound pickup is a back row sound pickup, receiving the first user voice through the back row sound pickup; and
the apparatus is further configured to determine a file address of the multimedia file based on the identification of the multimedia file;
playing the multimedia file by using the target playing device based on the file address, comprising:
and if the target playing device is a back-row multimedia playing device, playing the multimedia file by using the back-row multimedia playing device based on the file address.
9. An interactive system for application to a vehicle, the system comprising a processor, at least two microphones for an interior compartment of the vehicle, and at least two multimedia playing devices;
the at least two sound pickups are used for receiving user voice;
the processor is configured to perform the method of any one of claims 1-4;
and the target playing device is used for playing the multimedia file.
10. An electronic device, comprising:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement the method of any one of claims 1-4.
11. A computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, carries out the method according to any one of claims 1-4.
CN201910184912.1A 2019-03-12 2019-03-12 Interaction method and device applied to vehicle Active CN109817214B (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN202111119722.5A CN113763956A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle
CN201910184912.1A CN109817214B (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle
CN202111120001.6A CN113763957A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910184912.1A CN109817214B (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle

Related Child Applications (2)

Application Number Title Priority Date Filing Date
CN202111120001.6A Division CN113763957A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle
CN202111119722.5A Division CN113763956A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle

Publications (2)

Publication Number Publication Date
CN109817214A CN109817214A (en) 2019-05-28
CN109817214B true CN109817214B (en) 2021-11-23

Family

ID=66608773

Family Applications (3)

Application Number Title Priority Date Filing Date
CN202111120001.6A Pending CN113763957A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle
CN201910184912.1A Active CN109817214B (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle
CN202111119722.5A Pending CN113763956A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN202111120001.6A Pending CN113763957A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle

Family Applications After (1)

Application Number Title Priority Date Filing Date
CN202111119722.5A Pending CN113763956A (en) 2019-03-12 2019-03-12 Interaction method and device applied to vehicle

Country Status (1)

Country Link
CN (3) CN113763957A (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111462744B (en) * 2020-04-02 2024-01-30 深圳创维-Rgb电子有限公司 Voice interaction method and device, electronic equipment and storage medium
CN112802468B (en) * 2020-12-24 2023-07-11 合创汽车科技有限公司 Interaction method and device of automobile intelligent terminal, computer equipment and storage medium
CN115547327A (en) * 2022-09-23 2022-12-30 中国第一汽车股份有限公司 Data transmission method and device, storage medium and target vehicle

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106953961A (en) * 2017-04-28 2017-07-14 苏州科技大学 The mobile phone speech application apparatus and its application process of a kind of dual microphone
CN108399916A (en) * 2018-01-08 2018-08-14 蔚来汽车有限公司 Vehicle intelligent voice interactive system and method, processing unit and storage device

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1815556A (en) * 2005-02-01 2006-08-09 松下电器产业株式会社 Method and system capable of operating and controlling vehicle using voice instruction
FR2945696B1 (en) * 2009-05-14 2012-02-24 Parrot METHOD FOR SELECTING A MICROPHONE AMONG TWO OR MORE MICROPHONES, FOR A SPEECH PROCESSING SYSTEM SUCH AS A "HANDS-FREE" TELEPHONE DEVICE OPERATING IN A NOISE ENVIRONMENT.
CN204189450U (en) * 2014-10-30 2015-03-04 上海修源网络科技有限公司 A kind of vehicle-mounted voice operating system human-computer interaction device
CN105355213B (en) * 2015-11-11 2019-09-24 Oppo广东移动通信有限公司 A kind of method and device of directional recording
CN107303909B (en) * 2016-04-20 2020-06-23 斑马网络技术有限公司 Voice call-up method, device and equipment
CN106095380A (en) * 2016-05-31 2016-11-09 广东欧珀移动通信有限公司 Acoustical signal acquisition methods and device
US10448150B2 (en) * 2016-06-03 2019-10-15 Faraday & Future Inc. Method and apparatus to detect and isolate audio in a vehicle using multiple microphones
CN106357871A (en) * 2016-09-29 2017-01-25 维沃移动通信有限公司 Voice amplifying method and mobile terminal
JP6896540B2 (en) * 2017-07-18 2021-06-30 アルパイン株式会社 In-vehicle system
CN109257682B (en) * 2018-09-29 2020-04-24 歌尔科技有限公司 Sound pickup adjusting method, control terminal and computer readable storage medium
CN109300471B (en) * 2018-10-23 2021-09-14 中冶东方工程技术有限公司 Intelligent video monitoring method, device and system for field area integrating sound collection and identification

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106953961A (en) * 2017-04-28 2017-07-14 苏州科技大学 The mobile phone speech application apparatus and its application process of a kind of dual microphone
CN108399916A (en) * 2018-01-08 2018-08-14 蔚来汽车有限公司 Vehicle intelligent voice interactive system and method, processing unit and storage device

Also Published As

Publication number Publication date
CN113763956A (en) 2021-12-07
CN109817214A (en) 2019-05-28
CN113763957A (en) 2021-12-07

Similar Documents

Publication Publication Date Title
CN109817214B (en) Interaction method and device applied to vehicle
CN106910510A (en) Vehicle-mounted power amplifying device, vehicle and its audio play handling method
CN107004423A (en) Feedback for sharing the communication of the enhanced conversation type in acoustic space is eliminated
CN110278543B (en) Method and device for updating control system of automobile and storage medium
CN107004425A (en) Enhanced conversational communication in shared acoustic space
CN111343410A (en) Mute prompt method and device, electronic equipment and storage medium
WO2022042634A1 (en) Audio data processing method and apparatus, and device and storage medium
CN109582274B (en) Volume adjusting method and device, electronic equipment and computer readable storage medium
KR20210041553A (en) Audio stream mixing system and method
CN115038011A (en) Vehicle, control method, control device, control equipment and storage medium
CN111429902A (en) Method and apparatus for waking up a device
CN110166157A (en) The control method and device of in-vehicle multi-media system, electronic equipment, readable medium
CN111045634B (en) Audio processing method and device
CN110601966A (en) Method, electronic device and computer readable medium for playing messages
KR101500177B1 (en) Audio system of vehicle
CN112307161B (en) Method and apparatus for playing audio
CN114038465B (en) Voice processing method and device and electronic equipment
CN114121050A (en) Audio playing method and device, electronic equipment and storage medium
CN109375892B (en) Method and apparatus for playing audio
CN111930229B (en) Man-machine interaction method and device and electronic equipment
CN111741444A (en) Display method, device, equipment and storage medium
CN109445873B (en) Method and device for displaying setting interface
CN116866872A (en) Method, apparatus and computer storage medium for information processing
CN111145792B (en) Audio processing method and device
CN113838488B (en) Audio playing packet generation method and device and audio playing method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20211019

Address after: 100176 Room 101, 1st floor, building 1, yard 7, Ruihe West 2nd Road, economic and Technological Development Zone, Daxing District, Beijing

Applicant after: Apollo Zhilian (Beijing) Technology Co.,Ltd.

Address before: 100085 Baidu Building, 10 Shangdi Tenth Street, Haidian District, Beijing

Applicant before: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) Co.,Ltd.

GR01 Patent grant
GR01 Patent grant