WO2024055566A1 - 一种车载交互方法、系统、控制器和汽车 - Google Patents

一种车载交互方法、系统、控制器和汽车 Download PDF

Info

Publication number
WO2024055566A1
WO2024055566A1 PCT/CN2023/084779 CN2023084779W WO2024055566A1 WO 2024055566 A1 WO2024055566 A1 WO 2024055566A1 CN 2023084779 W CN2023084779 W CN 2023084779W WO 2024055566 A1 WO2024055566 A1 WO 2024055566A1
Authority
WO
WIPO (PCT)
Prior art keywords
voice
display
vehicle
state
container
Prior art date
Application number
PCT/CN2023/084779
Other languages
English (en)
French (fr)
Inventor
张泉辉
罗化
凌云健
张月霞
吴盼盼
Original Assignee
比亚迪股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 比亚迪股份有限公司 filed Critical 比亚迪股份有限公司
Publication of WO2024055566A1 publication Critical patent/WO2024055566A1/zh

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R11/00Arrangements for holding or mounting articles, not otherwise provided for
    • B60R11/02Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/20Arrangements for obtaining desired frequency or directional characteristics
    • H04R1/32Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only

Definitions

  • the present application relates to the field of human-computer interaction, and in particular to a vehicle-mounted interaction method, system, controller and automobile.
  • the voice interaction interface is displayed in a preset fixed position, resulting in users being unable to obtain a good visual and interactive experience.
  • This application provides a vehicle-mounted interaction method and system, which uses a sound source positioning method to determine the user's orientation information in the vehicle.
  • the target display position of the voice interaction interface on the display screen is determined through the orientation information, and the target display position is displayed according to the target display position.
  • the voice interaction interface solves the problem that the voice interaction interface is only displayed in a preset fixed position, which enables different users to display the voice interaction interface in different positions on the screen when they interact with the vehicle-mounted terminal to improve the user's visual experience. and interactive experience to improve interaction efficiency.
  • the first aspect of this application discloses a vehicle-mounted interaction method.
  • the method is applied to a vehicle-mounted terminal.
  • the method includes:
  • the voice container is displayed on the display screen according to the target display position.
  • the second aspect of this application discloses a vehicle-mounted interactive system, the system is applied to a vehicle-mounted terminal, and the system includes:
  • a first determination module configured to respond to receiving the user's first voice input and determine the orientation information of the sound source of the first voice relative to the interior space of the vehicle;
  • a second determination module configured to determine the target display position of the voice container in response to the first voice on the vehicle-mounted terminal according to the orientation information
  • a display module is configured to display the voice container on the display screen according to the target display position.
  • the third aspect of this application discloses a controller, which is used to execute some or all of the steps in any of the vehicle-mounted interaction methods disclosed in the first aspect of this application.
  • the fourth aspect of this application discloses a computer storage medium.
  • the computer storage medium stores computer instructions. When the computer instructions are called, they are used to execute part of any of the vehicle-mounted interaction methods disclosed in the first aspect of this application. or all steps.
  • the fifth aspect of this application discloses a car, which includes the vehicle-mounted interactive system disclosed in the second aspect of this application, or the controller disclosed in the third aspect of this application.
  • This application discloses a vehicle-mounted interaction method and system.
  • the method includes: responding to the user's first voice input and determining the orientation information of the sound source of the first voice relative to the interior space of the vehicle; determining the response according to the orientation information.
  • the voice container of the first voice is at a target display position of the vehicle-mounted terminal; and the voice container is displayed on the display screen according to the target display position.
  • this application uses the sound source positioning method to determine the user's orientation information in the car, determines the target display position of the voice interactive interface on the display screen through the orientation information, and displays the voice interactive interface according to the target display position, which can enable different When users conduct voice interaction with the vehicle-mounted terminal, the voice interaction interface is displayed at different locations on the screen to improve the user's visual experience and interactive experience and improve interaction efficiency.
  • Figure 1 is a schematic flowchart of a vehicle interaction method in an embodiment of the present application
  • Figure 2 is another schematic flowchart of a vehicle interaction method in an embodiment of the present application.
  • Figure 3 is a schematic display diagram of a voice container in an embodiment of the present application.
  • Figure 4 is a schematic diagram of the display of the voice container when the main driver wakes up and the co-driver wakes up in an embodiment of the present application;
  • Figure 5 is a schematic diagram of the sound zone in the car according to an embodiment of the present application.
  • Figure 6 is a schematic diagram of different displays of voice containers when responding to sound sources at different locations in the car in an embodiment of the present application
  • Figure 7 is a schematic diagram of the display of the voice status component in different voice interaction states in an embodiment of the present application.
  • Figure 8 is a schematic diagram of the permanent display of the voice status component on the display screen in an embodiment of the present application.
  • Figure 9 is a schematic diagram showing the display of a voice container when responding to voice under a third-party application in an embodiment of the present application.
  • Figure 10 is a schematic structural diagram of a vehicle interactive system in an embodiment of the present application.
  • an embodiment means that a particular feature, structure or characteristic described in connection with the embodiment can be included in at least one embodiment of the present application.
  • the appearances of this phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those skilled in the art understand, both explicitly and implicitly, that the embodiments described herein may be combined with other embodiments.
  • This application discloses a vehicle-mounted interaction method and system.
  • the vehicle-mounted interaction method and system can enable different users to display the voice interaction interface at different positions on the screen when they interact with the vehicle-mounted terminal, thereby improving the user's visual experience and interactive experience. , improve interaction efficiency. Detailed descriptions are given below.
  • FIG. 1 Please refer to Figures 1 and 3.
  • This application provides a vehicle-mounted interaction method.
  • the vehicle-mounted interaction method is applied to a vehicle-mounted terminal.
  • the vehicle-mounted interaction method may include the following operations:
  • the vehicle-mounted terminal responds to receiving the user's first voice input and determines the orientation information of the sound source of the first voice relative to the space inside the vehicle.
  • a voice collection device needs to be installed in the vehicle to collect voice information in the vehicle.
  • the vehicle-mounted terminal receives the user's first voice collected by the voice collection device, it can determine the direction and distance of the sound source relative to the voice collection device based on the voice information of the first voice, which can then be combined with the position of the seat in the car. and the spatial distance between seats to fit the orientation information of the sound source relative to the interior space of the car.
  • the orientation information may include the left front position, the right front position, the left rear position and the right rear position, which is not limited in the embodiment of this application.
  • the vehicle-mounted terminal can determine that the user's orientation information is the left front position in the car.
  • the vehicle-mounted terminal determines the target display position of the voice container in response to the first voice on the vehicle-mounted terminal according to the orientation information.
  • the above-mentioned voice container is also a voice interactive interface, and the display position of the voice container on the display screen of the vehicle-mounted terminal is determined by the vehicle-mounted terminal based on the orientation information obtained in step S101.
  • the vehicle-mounted terminal determines that the direction information of the sound source is the left front position according to the sound source of the voice command, the text content of the command "How is the weather today" can be included in the user's voice command.
  • the voice container 11 is displayed on the display screen 1, and its target display position is the upper left corner of the display screen.
  • the target display position may include several display marks used to represent the orientation information, and may be the upper left corner, the upper right corner, or the middle position, which is not limited in the embodiment of this application.
  • the vehicle-mounted terminal displays the voice container on the display screen according to the target display position.
  • the method described in the embodiments of this application can use the sound source positioning method to determine the user's orientation information in the car, determine the target display position of the voice interaction interface on the display screen through the orientation information, and display the voice according to the target display position.
  • the interactive interface solves the problem that the voice interactive interface is only displayed in a preset fixed position, thereby enabling users in different locations to display the voice interactive interface in different positions on the screen when they interact with the vehicle-mounted terminal via voice, thereby improving the user's visual perception. experience and interactive experience to improve interaction efficiency.
  • step S102 may include the following operations:
  • the vehicle-mounted terminal obtains the sound zone parameters of the vehicle, and the sound zone parameters include one or more of the left front sound zone, the right front sound zone, the left middle sound zone, the right middle sound zone, the left rear sound zone and the right rear sound zone. a combination;
  • the vehicle-mounted terminal determines the target sound zone information that matches the orientation information according to the sound zone parameters
  • the vehicle-mounted terminal determines the target display position of the voice container in response to the first voice on the vehicle-mounted terminal based on the target sound zone information.
  • the vehicle can be divided into different sound zone parameters based on the distribution of the voice collection devices installed inside the vehicle. As shown in Figure 5, the vehicle is divided into 6 sound zones, including the left front sound zone A and the right front sound zone. Area B, left middle area C, right middle area D, left rear area E and right rear area F.
  • the vehicle-mounted terminal can determine the target sound zone information that matches the sound source location information based on the actual sound zone parameters of the vehicle, and then determine the target display position of the voice container.
  • the vehicle-mounted terminal can determine the user's location. The information is the rear position in the car, and then it can be determined that the target sound zone information matching the position information is other sound zones. Furthermore, the vehicle-mounted terminal can determine that the target display position of the voice container on the vehicle-mounted terminal is the middle position based on the target sound zone information.
  • the method described in the embodiments of the present application can determine the accurate sound zone position of the sound source by matching the sound zone parameters of the vehicle with the orientation information, which is conducive to improving the accuracy of determining the target display position of the speech container, and thus can When users in different locations interact with the vehicle-mounted terminal via voice, the voice interaction interface is displayed at different locations on the screen, thereby improving the user's visual experience and interactive experience, and improving interaction efficiency.
  • the vehicle-mounted terminal determines the target display position of the voice container in response to the first voice on the vehicle-mounted terminal based on the target sound zone information, which may include the following operations:
  • the vehicle-mounted terminal determines the target display area information of the voice container on the display screen and the color information corresponding to the target sound area information based on the corresponding relationship between the preset display area information and the target sound area information. and/or azimuth text information corresponding to the target sound area information;
  • the vehicle-mounted terminal determines the target display area information, and the color information corresponding to the target sound area information and/or the orientation text information corresponding to the target sound area information as the target display position of the voice container.
  • the corresponding relationship between the preset display area information and the target sound area information may be set to a one-to-one relationship or a many-to-one relationship, which is not limited in the embodiment of the present application.
  • the vehicle-mounted terminal can also determine the color information or text information corresponding to the target sound zone information based on the corresponding relationship.
  • the target sound area information is the left front sound area or the right front sound area
  • the corresponding area in the speech container is displayed in red
  • the target sound area information is the left front and back sound areas or the right back sound area
  • the corresponding area in the speech container Areas are shown in green.
  • Figure 6 if a vehicle only contains the left front sound area A, the right front sound area B and other sound areas, when the target sound area information is the left front sound area, the corresponding area in the speech container displays the word "main"; when the target sound area When the sound zone information is the right front sound zone, the corresponding area in the speech container displays the word "Deputy"; when the target sound zone information is other sound zones, the corresponding area in the speech container displays the word "Quan".
  • the method described in the embodiment of the present application can determine the target display area information, color, text and other information of the speech container through the corresponding relationship between the preset display area information and the target sound area information, and use a variety of methods to represent the orientation.
  • the information display logo provides a richer display of the voice container, allowing users in different locations to display different expressions when interacting with the vehicle terminal, clearly expressing the direction of the sound zone, whether it is the main driver, the co-driver, or the entire vehicle wake-up, etc. status to improve the user’s visual experience and interactive experience.
  • the method may further include the following operations:
  • the vehicle-mounted terminal determines whether the position information corresponding to the second voice is consistent with the position information corresponding to the first voice;
  • the vehicle-mounted terminal displays the voice container at a target display position that matches the orientation information corresponding to the second voice.
  • the vehicle-mounted terminal when the vehicle-mounted terminal responds to the first voice input from the user in the main driver's seat, the display position of the voice container on the display screen is the upper left. At this time, if another voice input from the passenger seat is received, During the user's second voice input, the vehicle-mounted terminal determines that the orientation information corresponding to the two voices is inconsistent, and then the vehicle-mounted terminal adjusts the display position of the voice container to the upper right. It should be noted that regardless of whether the current voice interaction mode is a continuous conversation mode or not, the display position of the voice container will be adjusted.
  • the continuous dialogue mode means that when the voice assistant responds to the user's voice command and does not exit to the unawakened state, it continues to issue new voice commands to the voice assistant.
  • the semantic interaction mode at this time is the continuous dialogue mode.
  • the non-continuous dialogue mode means that the voice assistant responds to the user's voice command and does not continue to issue new voice commands to the voice assistant.
  • the semantic interaction mode during this period is the discontinuous dialogue mode.
  • the method described in the embodiment of the present application determines whether to adjust the target display position of the voice container by whether the orientation information corresponding to the input voice is consistent.
  • the target display position of the voice container can be adjusted in real time, thereby enabling
  • the voice interaction interface can be quickly switched to different locations on the screen to improve the user's visual experience and interactive experience, and improve interaction efficiency.
  • the method may further include the following operations:
  • the vehicle-mounted terminal determines whether the currently running application is a third-party application
  • the vehicle-mounted terminal displays the voice container at a preset position on the display screen.
  • the voice container displayed in the upper left or upper right of the screen will not affect the user's operation of the interface of the desktop program.
  • the program running on the desktop is a third-party application such as a music playing application or a video playing application or a preset application
  • the voice container displayed in the upper right corner may affect the operation of the interface.
  • the vehicle-mounted terminal determines that the application currently running on the desktop is a third-party application. Even if it is a voice command issued by the person in the passenger seat, the Keep the speech container displayed in the upper left corner.
  • the method described in the embodiment of the present application determines whether the application running on the desktop is a third-party application, and presents the display of the voice container on the display screen in a unified position, without affecting the user's operation of the interface of the third-party application, thereby greatly Improved user experience.
  • the method may also include the following operations:
  • the vehicle-mounted terminal identifies the timbre information of the first voice
  • the vehicle-mounted terminal determines the target display identification corresponding to the timbre information according to the preset timbre display identification
  • the vehicle-mounted terminal displays the voice container in a display form corresponding to the target display identification.
  • the vehicle-mounted terminal can also identify the timbre information of the user's voice, and then determine the target display logo to be displayed in the voice container based on the timbre information, thereby displaying the voice container in a display form corresponding to the target display logo. shown on the display.
  • the vehicle-mounted terminal distinguishes and identifies the timbre information of girls or children in the back row, and displays a rose voice bar corresponding to a girl's voice or a yellow voice bar corresponding to a child's voice in the voice container.
  • the vehicle-mounted terminal can also dynamically display the above-mentioned voice bar with light effects based on the frequency of the recognized sound and the pitch of the sound.
  • the method described in the embodiments of the present application can distinguish different characteristics through timbre, audio, sound bass, etc., and then display them on the display screen through differentiated display logos, which can improve the user's personalized experience.
  • this application provides another vehicle-mounted interaction method.
  • the vehicle-mounted interaction method may include the following operations:
  • the vehicle-mounted terminal responds to receiving the user's first voice input and determines the orientation information of the sound source of the first voice relative to the interior space of the vehicle.
  • the vehicle-mounted terminal determines the target display position of the voice container in response to the first voice on the vehicle-mounted terminal according to the orientation information.
  • the vehicle-mounted terminal displays the voice container on the display screen according to the target display position.
  • the car The carrier terminal adjusts the display state of the voice container according to the application responding to the first voice and the current voice interaction mode.
  • the voice container 11 includes a voice state component 12.
  • the voice state component 12 is used for the vehicle-mounted terminal to display the current voice interaction state, wherein the voice interaction state includes a normal state. , listening state, speech recognition state, and broadcasting state; the voice container 11 also includes a function card 13, which is used by the vehicle-mounted terminal to display the voice interaction result when the current voice interaction state is the broadcasting state.
  • the vehicle-mounted terminal can adjust the display form of the voice state component according to the current voice interaction state.
  • the four display forms in the figure represent the normal state, listening state, speech recognition state, and broadcast state respectively.
  • the normal state which is the state when the voice is not awakened, can be integrated with the system scene desktop in the form of light effects running along the voice ball from sunrise to sunset, and the breathing state of the voice itself is displayed on the screen interface.
  • the listening state means that when the voice is awakened and the user inputs relevant speech instructions, the voice shape will be in the shape of an ear and the effect will be heard when listening, conveying that the user's voice instructions are being listened to.
  • the speech recognition state is that the user's input speech is being processed in real time to recognize its semantics.
  • the speech morphology here is dynamically like a water drop.
  • the broadcasting state means that after the voice command has been recognized, the voice interaction result needs to be displayed before broadcasting.
  • the working dynamics during the broadcasting are in the speaker breathing state. It should be noted that this application does not limit the specific form of the display form, which can be adjusted according to actual needs.
  • step S205 when the user touches the display area other than the voice container 11 on the display screen, the vehicle-mounted terminal will adjust the display state of the voice container according to the application of the voice response and the current voice interaction mode.
  • the display state of the voice container includes the display state of the voice state component and the display state of the function card. For example, when the user issues the voice command "Hello" to wake up the voice assistant, the voice status component is displayed on the display in the listening state. That is, the current voice interaction mode is discontinuous mode, and there are no voice response applications yet.
  • the display state of the voice status component can be adjusted to the normal state (that is, the above-mentioned state when the voice is not awakened).
  • the method described in the embodiment of the present application can realize the adjustment operation of the display state of the voice container through the combination of the user's voice and the user's touch operation on the screen, which greatly facilitates the interactive operation between the user and the vehicle-mounted terminal, and improves the Improve the user’s interactive experience and efficiency.
  • step S205 may include the following operations:
  • the vehicle-mounted terminal determines whether the application responding to the first voice is a preset application
  • the vehicle-mounted terminal hides the function card and adjusts the display state of the voice state component to the display form corresponding to the listening state;
  • the vehicle-mounted terminal hides the function card and adjusts the display state of the voice status component to the display form corresponding to the normal state.
  • the default application may be an application with a higher display priority such as a navigation application.
  • the display state of the voice container is adjusted only according to whether the current voice interaction mode is a continuous dialogue mode.
  • the function card is hidden and the voice state is adjusted to the listening state. At this time, the voice interaction function will not exit. ;
  • the vehicle-mounted terminal when it is determined that the application is the preset application, maintains the display state of the function card and maintains the display state of the voice status component as the The display form corresponding to the broadcast status;
  • the method further includes:
  • the vehicle-mounted terminal maintains the display state of the function card and adjusts the display state of the voice state component to the display form corresponding to the normal state.
  • the function card displays the navigation list.
  • the voice status component The display state remains the display form corresponding to the broadcast state. After that, when the current voice interaction state changes to the normal state due to timeout or other reasons, the vehicle-mounted terminal still maintains the display state of the function card and does not hide it.
  • the method described in the embodiment of the present application can realize the adjustment operation of the display status of the voice status component and the function card through the combination of the user's voice and the user's touch operation on the screen, thereby improving the humanized operation of voice interaction and greatly improving the humanized operation of voice interaction. It greatly facilitates the user's interaction with the vehicle-mounted terminal and improves the user's interactive experience and efficiency.
  • the vehicle-mounted terminal controls the voice status component to be permanently displayed on the display screen.
  • the voice status component is permanently displayed on the leftmost side of the status bar on the display. Since the status bar will not be easily hidden automatically during application use, it can ensure that smart voice can be used on any interface. It also improves the penetration rate of voice functions and the cockpit intelligent full-scenario interactive experience, making the interface display unified and regular.
  • sequence number of each step in the above embodiments does not mean the order of execution.
  • the execution order of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of the embodiments of the present application. .
  • this application provides a vehicle-mounted interactive system, which is applied to a vehicle-mounted terminal and may include:
  • the first determination module 301 is configured to respond to receiving the user's first voice input and determine the orientation information of the sound source of the first voice relative to the interior space of the vehicle;
  • the second determination module 302 is configured to determine the target display position of the voice container in response to the first voice on the vehicle-mounted terminal according to the orientation information;
  • the display module 303 is used to display the voice container on the display screen according to the target display position.
  • the system described in the embodiments of the present application can use the sound source positioning method to determine the user's position information in the car.
  • Information the target display position of the voice interactive interface on the display screen is determined through the orientation information, and the voice interactive interface is displayed according to the target display position, which solves the problem that the voice interactive interface is only displayed in a preset fixed position, thereby enabling
  • the voice interaction interface is displayed at different locations on the screen to improve the user's visual experience and interactive experience, and improve interaction efficiency.
  • This application also provides a controller for executing the vehicle-mounted interaction method.
  • controller for executing the vehicle-mounted interaction method.
  • Each module in the above controller can be implemented in whole or in part through software, hardware and combinations thereof.
  • Each of the above modules may be embedded in or independent of the processor of the computer device in the form of hardware, or may be stored in the memory of the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
  • This application also provides an automobile, including the above controller.
  • the present application also provides a car, comprising the above-mentioned in-vehicle interactive system.
  • the present application also provides a computer-readable storage medium on which computer instructions are stored, and when executed by the processor, the computer instructions implement the vehicle-mounted interaction method described in the above method embodiment.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • Mechanical Engineering (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

一种车载交互方法、系统、控制器和汽车,涉及人机交互领域。方法应用于车载终端,方法包括:响应于用户的第一语音输入,并确定第一语音的音源相对于车内空间的方位信息(S101);根据方位信息确定响应于第一语音的语音容器(11)在车载终端的目标显示位置(S102);根据目标显示位置将语音容器(11)显示在显示屏(1)上(S103)。

Description

一种车载交互方法、系统、控制器和汽车
相关申请的交叉引用
本申请要求于2022年09月15日提交的申请号为202211127032.9、名称为“一种车载交互方法、系统、控制器和汽车”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及人机交互领域,尤其涉及一种车载交互方法、系统、控制器和汽车。
背景技术
目前,针对汽车智能座舱内的语音交互功能,用户在与车载终端进行语音交互时,语音交互界面均是以预设固定的位置显示,导致用户无法获得良好的视觉和交互体验。
发明内容
本申请提供了一种车载交互方法及系统,采用音源定位方式确定用户在车内的方位信息,通过该方位信息确定出语音交互界面在显示屏上的目标显示位置,按照该目标显示位置显示该语音交互界面,解决了语音交互界面仅以预设固定的位置显示的问题,进而能够使得不同的用户与车载终端进行语音交互时,将语音交互界面显示在屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。
为了解决上述技术问题,本申请第一方面公开了一种车载交互方法,所述方法应用于车载终端,所述方法包括:
响应于用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息;
根据所述方位信息确定响应于所述第一语音的语音容器在车载终端的目标显示位置;
根据所述目标显示位置将所述语音容器显示在显示屏上。
本申请第二方面公开了一种车载交互系统,所述系统应用于车载终端,所述系统包括:
第一确定模块,用于响应于接收到用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息;
第二确定模块,用于根据所述方位信息确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置;
显示模块,用于根据所述目标显示位置将所述语音容器显示在显示屏上。
本申请第三方面公开了一种控制器,所述控制器用于执行本申请第一方面公开的任意一种车载交互方法中的部分或全部步骤。
本申请第四方面公开了一种计算机存储介质,所述计算机存储介质存储有计算机指令,所述计算机指令被调用时,用于执行本申请第一方面公开的任意一种车载交互方法中的部分或全部步骤。
本申请第五方面公开了一种汽车,所述汽车包括本申请第二方面公开的车载交互系统,或本申请第三方面公开的控制器。
与现有技术相比,本申请具有以下有益效果:
本申请公开了一种车载交互方法及系统,该方法包括:响应于用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息;根据所述方位信息确定响应于所述第一语音的语音容器在车载终端的目标显示位置;根据所述目标显示位置将所述语音容器显示在显示屏上。可见,本申请采用音源定位方式确定用户在车内的方位信息,通过该方位信息确定出语音交互界面在显示屏上的目标显示位置,按照该目标显示位置显示该语音交互界面,能够使得不同的用户与车载终端进行语音交互时,将语音交互界面显示在屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。
附图说明
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1是本申请一实施例中车载交互方法的一流程示意图;
图2是本申请一实施例中车载交互方法的另一流程示意图;
图3是本申请一实施例中语音容器的显示示意图;
图4是本申请一实施例中主驾唤醒和副驾唤醒时语音容器的显示示意图;
图5是本申请一实施例中车内音区的示意图;
图6是本申请一实施例中响应车内不同位置音源时语音容器的不同显示示意图;
图7是本申请一实施例中语音状态组件在不同语音交互状态下的显示示意图;
图8是本申请一实施例中语音状态组件在显示屏上常驻显示的示意图;
图9是本申请一实施例中在第三方应用下响应语音时语音容器的显示示意图;
图10是本申请一实施例中车载交互系统的结构示意图。
具体实施方式
为了使本技术领域的人员更好地理解本申请方案,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人 员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别不同对象,而不是用于描述特定顺序。此外,术语“包括”和“具有”以及它们任何变形,意图在于覆盖不排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或端没有限定于已列出的步骤或单元,而是可选地还包括没有列出的步骤或单元,或可选地还包括对于这些过程、方法、产品或端固有的其他步骤或单元。
在本文中提及“实施例”意味着,结合实施例描述的特定特征、结构或特性可以包含在本申请的至少一个实施例中。在说明书中的各个位置出现该短语并不一定均是指相同的实施例,也不是与其它实施例互斥的独立的或备选的实施例。本领域技术人员显式地和隐式地理解的是,本文所描述的实施例可以与其它实施例相结合。
本申请公开了一种车载交互方法及系统,该车载交互方法及系统能够使得不同的用户与车载终端进行语音交互时,将语音交互界面显示在屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。以下分别进行详细的说明。
在一实施例中,请参阅图1和图3,本申请提供一种车载交互方法,该车载交互方法应用于车载终端,如图1所示,该车载交互方法可以包括以下操作:
S101、车载终端响应于接收到用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息。
本申请实施中,车辆中需要安装有语音采集装置,可以采集车内的语音信息。车载终端在接收到语音采集装置采集到的用户的第一语音时,便可根据第一语音的语音信息,确定出音源相对于语音采集装置的方向和距离,进而可以结合车内座椅的位置以及座椅之间的空间距离,拟合出音源相对于车内空间的方位信息。其中,根据车内的座椅设置,比如当有4个座椅时,方位信息可以包括左前位、右前位、左后位以及右后位,本申请实施例不做限定。举例说明,当用户在靠近主驾驶位置向车载终端发出语音指令时,则车载终端可以确定出用户的方位信息为车内的左前位。
S102、所述车载终端根据所述方位信息确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置。
本申请实施例中,上述语音容器也即语音交互界面,该语音容器在车载终端显示屏上的显示位置由车载终端根据步骤S101中获得的方位信息确定。如图3所示,用户发出语音指令“今天天气怎么样”,车载终端根据语音指令的声源确定出音源的方位信息为左前位时,可将包含用户的“今天天气怎么样”指令文字内容的语音容器11显示在显示屏1上,其目标显示位置为显示屏的左上方。需要说明的是,目标显示位置可以包括若干种用于表征所述方位信息的显示标识,可以是左上角,可以是右上角,还可以是中间位置,本申请实施例不做限定。
S103、所述车载终端根据所述目标显示位置将所述语音容器显示在显示屏上。
可见,本申请实施例所描述的方法能够采用音源定位方式确定用户在车内的方位信息,通过该方位信息确定出语音交互界面在显示屏上的目标显示位置,按照该目标显示位置显示该语音交互界面,解决了语音交互界面仅以预设固定的位置显示的问题,进而能够使得不同的位置的用户与车载终端进行语音交互时,将语音交互界面显示在屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。
在一个可选的实施例中,步骤S102可以包括以下操作:
所述车载终端获取车辆的音区参数,所述音区参数包括左前音区、右前音区、左中音区、右中音区、左后音区以及右后音区中的一种或多种组合;
所述车载终端根据所述音区参数确定出与所述方位信息相匹配的目标音区信息;
所述车载终端根据所述目标音区信息,确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置。
本申请实施例中,车辆依据其内部设置的语音采集装置的分布的不同,可以划分出不同的音区参数,如图5所示车辆划分为6个音区,包括左前音区A、右前音区B、左中音区C、右中音区D、左后音区E以及右后音区F。车载终端可以根据车辆的实际音区参数确定出与音源方位信息相匹配的目标音区信息,进而确定出语音容器的目标显示位置。
举例说明,若一车辆只包含左前音区A、右前音区B以及其他音区,此时当用户在左后排座椅位置向车载终端发出语音指令时,则车载终端可以确定出用户的方位信息为车内的后方位,进而可以确定出与该方位信息相匹配的目标音区信息为其他音区。进而,车载终端可以根据该目标音区信息,确定语音容器在车载终端的目标显示位置为中间位置。
可见,本申请实施例所描述的方法能够通过车辆的音区参数与方位信息匹配的方式,确定出音源的准确音区位置,有利于提高确定出语音容器的目标显示位置的精准度,进而能够使得不同的位置的用户与车载终端进行语音交互时,将语音交互界面显示在屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。
在该可选的实施例中,所述车载终端根据所述目标音区信息,确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置,可以包括以下操作:
所述车载终端根据预设显示区域信息与所述目标音区信息的对应关系,确定所述语音容器在所述显示屏上的目标显示区域信息,以及与所述目标音区信息对应的颜色信息和/或与所述目标音区信息对应的方位文字信息;
所述车载终端将所述目标显示区域信息,以及与所述目标音区信息对应的颜色信息和/或与所述目标音区信息对应的方位文字信息确定为所述语音容器的目标显示位置。
本申请实施例中,预设显示区域信息与所述目标音区信息的对应关系可以设置为一一对应的关系,也可以设置为多对一的关系,本申请实施例不做限定。示例性的,预设显示区域信息与所述目标音区信息的对应关系:{预设显示区域,目标音区信息}={(左上角,左前音区),(右上角,右前音区),(左下角,左后音区),(右下角,右后音区)}。此外, 车载终端还可以根据该对应关系确定出与目标音区信息对应的颜色信息或文字信息。举例说明,当目标音区信息为左前音区或右前音区时,语音容器中对应的区域显示为红色;当目标音区信息为左前后音区或右后音区时,语音容器中对应的区域显示为绿色。如图6所示,若一车辆只包含左前音区A、右前音区B以及其他音区,当目标音区信息为左前音区时,语音容器中对应的区域显示“主”字;当目标音区信息为右前音区时,语音容器中对应的区域显示“副”字;当目标音区信息为其他音区时,语音容器中对应的区域显示“全”字。
可见,本申请实施例所描述的方法能够通过预设显示区域信息与所述目标音区信息的对应关系确定出语音容器的目标显示区域信息、颜色以及文字等信息,采用多种表征所述方位信息的显示标识对语音容器进行更加丰富的展示,使得不同的位置的用户与车载终端进行语音交互时展现不同的表现形态,清晰地表达音区的方位,是主驾还是副驾还是全车唤醒等状态,提高用户的视觉体验和交互体验。
在另一个可选的实施例中,在所述车载终端根据所述目标显示位置将所述语音容器显示在显示屏上之后,该方法还可以包括以下操作:
所述车载终端响应于接收到用户的第二语音输入,判断所述第二语音对应的方位信息与所述第一语音对应的方位信息是否一致;
当判断结果为否时,所述车载终端将所述语音容器以与所述第二语音对应的方位信息相匹配的目标显示位置进行显示。
本申请实施例中,如图4所示,当车载终端响应于主驾驶位用户的第一语音输入时,语音容器在显示屏上的显示位置为左上方,此时若又接收到副驾驶位用户的第二语音输入时,车载终端判断出两次语音对应的方位信息不一致,进而车载终端将语音容器的显示位置调整为右上方。需要说明的是,当前的语音交互模式无论是否为连续对话模式,语音容器的显示位置均会被调整。其中,连续对话模式是指在语音助手响应于用户的语音指令,且未退出到未唤醒状态时,继续向语音助手发布新语音指令,此时的语义交互模式为连续对话模式。非连续对话模式是指在语音助手响应于用户的语音指令,未继续向语音助手发布新语音指令,在语音助手未退出到未唤醒状态之前,此段时间的语义交互模式为非连续对话模式。
可见,本申请实施例所描述的方法通过输入语音对应的方位信息是否一致,进而来判断是否对语音容器的目标显示位置进行调整,能够在不一致时,实时调整语音容器目标显示位置,进而能够使得不同的位置的用户与车载终端进行语音交互时,将语音交互界面快速切换到屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。
在又一个可选的实施例中,在所述车载终端根据所述目标显示位置将所述语音容器显示在显示屏上之前,该方法还可以包括以下操作:
所述车载终端判断当前运行的应用是否为第三方应用;
当判断结果为是时,所述车载终端将所述语音容器显示在显示屏的预设位置处。
本申请实施例中,示例性的,当桌面运行的程序是导航界面时,此时语音容器无论显示在屏幕的左上方还是右上方,都不会影响用户对桌面程序的界面进行操作,而当桌面运行的程序是音乐播放应用或者视频播放应用等第三方应用或是预先设定的应用程序时,此时语音容器显示在右上方可能会影响界面的操作。如图9示例的第三方应用下响应语音时语音容器的显示示意图所示,车载终端判断出当前桌面运行的应用程序为第三方应用程序,即使是副驾驶位人员发出的语音指令,此时仍将语音容器保持在左上角显示。
可见,本申请实施例所描述的方法通过判断桌面运行应用是否为第三方应用,将语音容器在显示屏上的显示呈现为统一的位置,不影响用户对第三方应用的界面进行操作,从而大大提升了用户的使用体验。
在又一个可选的实施例中,该方法还可以包括以下操作:
所述车载终端识别所述第一语音的音色信息;
所述车载终端根据预设音色显示标识,确定所述音色信息对应的目标显示标识;
所述车载终端将所述语音容器以与所述目标显示标识相对应的显示形态进行显示。
本申请实施例中,车载终端还可以识别用户语音的音色信息,进而根据音色信息确定出要显示在语音容器中的目标显示标识,从而将语音容器以与所述目标显示标识相对应的显示形态显示在显示屏上。比如,车载终端区分识别女生或后排儿童的音色信息,在语音容器中显示出玫瑰色的语音条对应女生的声音或者,显示出黄色的语音条对应儿童的声音。
在一可选实施例中,车载终端还可以根据识别出声音的频率和音贝的高低,来对上述的语音条进行光效的动态显示。
可见,本申请实施例所描述的方法能够通过音色、音频以及音贝等方面,区分出不同的特征之后,通过差异化的显示标识显示在显示屏上,能够提高用户的个性化体验。
在又一实施例中,如图2所示,本申请提供另一种车载交互方法,该车载交互方法可以包括以下操作:
S201、车载终端响应于接收到用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息。
S202、所述车载终端根据所述方位信息确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置。
S203、所述车载终端根据所述目标显示位置将所述语音容器显示在显示屏上。
本申请实施例中,针对步骤S201-步骤S203的其它描述,请分别对应参照实施例一中针对步骤S101-步骤S103的详细描述,本申请实施例不再赘述。
S204、接收用户的点击指令。
S205、当判断出所述点击指令的点击区域在所述语音容器的显示区域之外时,所述车 载终端根据响应于所述第一语音的应用以及当前语音交互模式,调整所述语音容器的显示状态。
本申请实施例中,如图3所示,语音容器11包括语音状态组件12,所述语音状态组件12用于所述车载终端展示当前的语音交互状态,其中,所述语音交互状态包括常态状态、倾听状态、语音识别状态、播报状态;所述语音容器11还包括功能卡片13,所述功能卡片13用于所述车载终端在当前的语音交互状态为所述播报状态时展示语音交互结果。
如图7所示,车载终端可以根据当前的语音交互状态,调整所述语音状态组件的展示形态,图中4个展示形态分别代表常态状态、倾听状态、语音识别状态、播报状态。常态状态也即语音未唤醒时的状态,可以以光效沿着语音球日出到日落运作作息的形式,与系统场景桌面相融合,具有语音本身的呼吸态展现在屏幕界面上。倾听状态也即语音被唤醒时,用户输入相关话语指令,语音形态呈耳朵状侧耳倾听时动效,传达正在聆听用户语音指令。语音识别状态也即正在实时处理用户输入语音以识别其语义,这里语音形态工作动态呈水滴状。播报状态也即语音指令被识别完成需要展示语音交互结果后进行播报,播报时的工作动态呈喇叭呼吸态。需要说明的是,本申请并不限定展示形态的具体形式,可以根据实际需要进行调整。
在步骤S205中,当用户点触显示屏上除语音容器11之外的显示区域时,车载终端会根据语音响应的应用以及当前的语音交互模式,调整语音容器的显示状态。其中,语音容器的显示状态包括所述语音状态组件的显示状态以及所述功能卡片的显示状态。比如,当用户发出语音指令“你好”唤醒语音助手时,语音状态组件以倾听状态显示在显示屏上,也即当前的语音交互模式为非连续模式,同时语音响应的应用也还未有,此时用户点触屏幕上的语音容器11之外的显示区域时,即可调整语音状态组件的显示状态至常态状态(也即上述的语音未唤醒时的状态)。
可见,本申请实施例所描述的方法能够通过用户语音以及用户对屏幕的点触操作的结合,组合实现语音容器的显示状态的调整操作,极大的方便了用户与车载终端的交互操作,提高了用户的交互体验和交互效率。
在一个可选的实施例中,步骤S205可以包括以下操作:
所述车载终端判断响应于所述第一语音的应用是否为预设应用;
当判断出所述应用不为所述预设应用时,判断当前语音交互模式是否为连续对话模式;
当判断结果为是时,所述车载终端隐藏所述功能卡片,并将所述语音状态组件的显示状态调整为所述倾听状态对应的显示形态;
当判断结果为否时,所述车载终端隐藏所述功能卡片,并将所述语音状态组件的显示状态调整为所述常态状态对应的显示形态。
本申请实施例中,预设应用可以是导航应用等显示优先级较高的应用,当车载终端判 断出响应于第一语音的应用不属于上述预设应用时,则仅根据当前语音交互模式是否为连续对话模式,调整所述语音容器的显示状态。考虑到更好的体现语音交互的人性化体验,本申请示例中可设置,当为连续对话模式时,将功能卡片隐藏的同时将语音状态调整为倾听状态,此时并不会退出语音交互功能;当不为连续对话模式时,将功能卡片隐藏的同时将语音状态调整为常态状态,也即调整到语音未唤醒状态。
在该可选的实施例中,当判断出所述应用为所述预设应用时,所述车载终端维持所述功能卡片的显示状态,并将所述语音状态组件的显示状态维持为所述播报状态对应的显示形态;
在所述车载终端将所述语音状态组件的显示状态维持为所述播报状态对应的显示形态之后,所述方法还包括:
若当前的语音交互状态变更为所述常态状态时,所述车载终端维持所述功能卡片的显示状态,并将所述语音状态组件的显示状态调整为所述常态状态对应的显示形态。
本申请实施例中,举例说明,当语音响应的应用为导航应用时,无论当前语音交互模式是否为连续对话模式,功能卡片显示出导航列表,此时点击列表的外围区域时,语音状态组件的显示状态仍维持为播报状态对应的显示形态。在此之后,因超时等原因当前语音交互状态变更为常态状态时,车载终端仍维持所述功能卡片的显示状态,不将其隐藏。
可见,本申请实施例所描述的方法能够通过用户语音以及用户对屏幕的点触操作的结合,组合实现对语音状态组件以及功能卡片的显示状态的调整操作,提升语音交互的人性化操作,极大的方便了用户与车载终端的交互操作,提高了用户的交互体验和交互效率。
在另一个可选的实施例中,如图8所示,车载终端控制所述语音状态组件在所述显示屏上常驻显示。比如,语音状态组件作为智能语音的入口,常驻显示在显示屏上的状态栏最左侧,由于状态栏在应用使用过程中不会轻易自动隐藏,能够保证了智能语音在任意界面下都常驻显示,同时提升语音功能的渗透率和座舱智能化全场景交互形态体验,使得界面显示统一规整。
应理解,上述各实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。
在又一实施例中,如图10所示,本申请提供一种车载交互系统,该系统应用于车载终端,其可以包括:
第一确定模块301,用于响应于接收到用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息;
第二确定模块302,用于根据所述方位信息确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置;
显示模块303,用于根据所述目标显示位置将所述语音容器显示在显示屏上。
可见,本申请实施例所描述的系统能够采用音源定位方式确定用户在车内的方位信 息,通过该方位信息确定出语音交互界面在显示屏上的目标显示位置,按照该目标显示位置显示该语音交互界面,解决了语音交互界面仅以预设固定的位置显示的问题,进而能够使得不同的位置的用户与车载终端进行语音交互时,将语音交互界面显示在屏幕的不同位置,提高用户的视觉体验和交互体验,提升交互效率。
本申请还提供一种控制器,用于执行所述的车载交互方法。关于控制器的具体限定可以参见上文中对于车载交互方法的限定,在此不再赘述。上述控制器中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。
本申请还提供一种汽车,包括上述的控制器。
本申请还提供一种汽车,包括上述的车载交互系统。
本申请还提供了一种计算机可读存储介质,其上存储有计算机指令,计算机程度在被处理器执行时实现根据上述方法实施例中所述的车载交互方法。
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。

Claims (16)

  1. 一种车载交互方法,其特征在于,所述方法应用于车载终端,所述方法包括:
    响应于用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息;
    根据所述方位信息确定响应于所述第一语音的语音容器在车载终端的目标显示位置;
    根据所述目标显示位置将所述语音容器显示在显示屏上。
  2. 根据权利要求1所述的车载交互方法,其特征在于,所述根据所述方位信息确定响应于所述第一语音的语音容器在车载终端的目标显示位置,包括:
    获取车辆的音区参数,所述音区参数包括左前音区、右前音区、左中音区、右中音区、左后音区以及右后音区中的一种或多种组合;
    根据所述音区参数确定出与所述方位信息相匹配的目标音区信息;
    根据所述目标音区信息,确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置。
  3. 根据权利要求2所述的车载交互方法,其特征在于,所述根据所述目标音区信息,确定响应于所述第一语音的语音容器在车载终端的目标显示位置,包括:
    根据预设显示区域信息与所述目标音区信息的对应关系,确定所述语音容器在所述显示屏上的目标显示区域信息,以及与所述目标音区信息对应的颜色信息和/或与所述目标音区信息对应的方位文字信息;
    将所述目标显示区域信息,以及与所述目标音区信息对应的颜色信息和/或与所述目标音区信息对应的方位文字信息确定为所述语音容器的目标显示位置。
  4. 根据权利要求1-3任一项所述的车载交互方法,其特征在于,所述根据所述目标显示位置将所述语音容器显示在显示屏上之后,所述方法还包括:
    响应于用户的第二语音输入,判断所述第二语音对应的方位信息与所述第一语音对应的方位信息是否一致;
    当判断结果为否时,将所述语音容器以与所述第二语音对应的方位信息相匹配的目标显示位置进行显示。
  5. 根据权利要求1-4任一项所述的车载交互方法,其特征在于,所述语音容器包括语音状态组件,所述语音状态组件用于所述车载终端展示当前的语音交互状态,其中,所述语音交互状态包括常态状态、倾听状态、语音识别状态、播报状态;
    所述语音容器还包括功能卡片,所述功能卡片用于所述车载终端在当前的语音交互状态为所述播报状态时展示语音交互结果。
  6. 根据权利要求5所述的车载交互方法,其特征在于,所述根据所述目标显示位置将所述语音容器显示在显示屏上之后,所述方法还包括:
    接收用户的点击指令;
    当判断出所述点击指令的点击区域在所述语音容器的显示区域之外时,根据响应于所述第一语音的应用以及当前语音交互模式,调整所述语音容器的显示状态,其中,所述语音容器的显示状态包括所述语音状态组件的显示状态以及所述功能卡片的显示状态。
  7. 根据权利要求6所述的车载交互方法,其特征在于,所述根据响应于所述第一语音的应用以及当前语音交互模式,调整所述语音容器的显示状态,包括:
    判断响应于所述第一语音的应用是否为预设应用;
    当判断出所述应用不为所述预设应用时,判断当前语音交互模式是否为连续对话模式;
    当判断结果为是时,隐藏所述功能卡片,并将所述语音状态组件的显示状态调整为所述倾听状态对应的显示形态;
    当判断结果为否时,隐藏所述功能卡片,并将所述语音状态组件的显示状态调整为所述常态状态对应的显示形态。
  8. 根据权利要求7所述的车载交互方法,其特征在于,所述方法还包括:
    当判断出所述应用为所述预设应用时,维持所述功能卡片的显示状态,并将所述语音状态组件的显示状态维持为所述播报状态对应的显示形态;
    在所述将所述语音状态组件的显示状态维持为所述播报状态对应的显示形态之后,所述方法还包括:
    若当前的语音交互状态变更为所述常态状态时,维持所述功能卡片的显示状态,并将所述语音状态组件的显示状态调整为所述常态状态对应的显示形态。
  9. 根据权利要求1-8任一项所述的车载交互方法,其特征在于,在所述根据所述目标显示位置将所述语音容器显示在显示屏上之前,所述方法还包括:
    所判断当前运行的应用是否为第三方应用;
    当判断结果为是时,将所述语音容器显示在显示屏的预设位置处。
  10. 根据权利要求5-8任一项所述的车载交互方法,其特征在于,所述方法还包括:
    根据当前的语音交互状态,调整所述语音状态组件的展示形态;
    其中,所述根据当前的语音交互状态,调整所述语音状态组件的展示形态,包括:
    当判断出当前的语音交互状态处于未唤醒状态时,控制所述语音状态组件展示所述常态状态对应的展示形态;
    当判断出当前的语音交互状态处于唤醒状态时,控制所述语音状态组件展示所述倾听状态对应的展示形态;
    当判断出当前的语音交互状态处于识别话语指令时,控制所述语音状态组件展示所述语音识别状态对应的展示形态;
    当判断出所述话语指令被识别完成需要展示语音交互结果时,控制所述语音状态组件 展示所述播报状态对应的展示形态。
  11. 根据权利要求10所述的车载交互方法,其特征在于,所述方法还包括:
    控制所述语音状态组件在所述显示屏上常驻显示。
  12. 根据权利要求1-11任一项所述的车载交互方法,其特征在于,所述方法还包括:
    识别所述第一语音的音色信息;
    根据预设音色显示标识,确定所述音色信息对应的目标显示标识;
    将所述语音容器以与所述目标显示标识相对应的显示形态进行显示。
  13. 一种车载交互系统,其特征在于,所述系统应用于车载终端,所述系统包括:
    第一确定模块,用于响应于接收到用户的第一语音输入,并确定所述第一语音的音源相对于车内空间的方位信息;
    第二确定模块,用于根据所述方位信息确定响应于所述第一语音的语音容器在所述车载终端的目标显示位置;
    显示模块,用于根据所述目标显示位置将所述语音容器显示在显示屏上。
  14. 一种控制器,其特征在于,所述控制器用于执行如权利要求1-12任一项所述的车载交互方法。
  15. 一种计算机存储介质,其特征在于,所述计算机存储介质存储有计算机指令,所述计算机指令被调用时用于执行如权利要求1-12任一项所述的车载交互方法。
  16. 一种汽车,其特征在于,所述汽车包括:如权利要求13所述的车载交互系统,或者,如权利要求14所述的控制器。
PCT/CN2023/084779 2022-09-15 2023-03-29 一种车载交互方法、系统、控制器和汽车 WO2024055566A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211127032.9A CN117746851A (zh) 2022-09-15 2022-09-15 一种车载交互方法、系统、控制器和汽车
CN202211127032.9 2022-09-15

Publications (1)

Publication Number Publication Date
WO2024055566A1 true WO2024055566A1 (zh) 2024-03-21

Family

ID=90274197

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/084779 WO2024055566A1 (zh) 2022-09-15 2023-03-29 一种车载交互方法、系统、控制器和汽车

Country Status (2)

Country Link
CN (1) CN117746851A (zh)
WO (1) WO2024055566A1 (zh)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110171372A (zh) * 2019-05-27 2019-08-27 广州小鹏汽车科技有限公司 车载终端的界面显示方法、装置及车辆
CN112309380A (zh) * 2019-07-26 2021-02-02 北京新能源汽车股份有限公司 一种语音控制方法、系统、设备及汽车
CN112365891A (zh) * 2020-10-30 2021-02-12 东风汽车有限公司 汽车贯穿屏的车机虚拟语音助手交互方法、电子设备及存储介质
CN112735411A (zh) * 2020-12-25 2021-04-30 广州橙行智动汽车科技有限公司 控制方法、客户端、车辆、语音系统和存储介质
CN112802468A (zh) * 2020-12-24 2021-05-14 广汽蔚来新能源汽车科技有限公司 汽车智能终端的交互方法、装置、计算机设备和存储介质
US20210358492A1 (en) * 2019-03-29 2021-11-18 Lg Electronics Inc. Voice interaction method and vehicle using the same
CN113851126A (zh) * 2021-09-22 2021-12-28 思必驰科技股份有限公司 车内语音交互方法及系统

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210358492A1 (en) * 2019-03-29 2021-11-18 Lg Electronics Inc. Voice interaction method and vehicle using the same
CN110171372A (zh) * 2019-05-27 2019-08-27 广州小鹏汽车科技有限公司 车载终端的界面显示方法、装置及车辆
CN112309380A (zh) * 2019-07-26 2021-02-02 北京新能源汽车股份有限公司 一种语音控制方法、系统、设备及汽车
CN112365891A (zh) * 2020-10-30 2021-02-12 东风汽车有限公司 汽车贯穿屏的车机虚拟语音助手交互方法、电子设备及存储介质
CN112802468A (zh) * 2020-12-24 2021-05-14 广汽蔚来新能源汽车科技有限公司 汽车智能终端的交互方法、装置、计算机设备和存储介质
CN112735411A (zh) * 2020-12-25 2021-04-30 广州橙行智动汽车科技有限公司 控制方法、客户端、车辆、语音系统和存储介质
CN113851126A (zh) * 2021-09-22 2021-12-28 思必驰科技股份有限公司 车内语音交互方法及系统

Also Published As

Publication number Publication date
CN117746851A (zh) 2024-03-22

Similar Documents

Publication Publication Date Title
CN107340991A (zh) 语音角色的切换方法、装置、设备以及存储介质
US10320354B1 (en) Controlling a volume level based on a user profile
KR102669100B1 (ko) 전자 장치 및 그 제어 방법
KR20190083476A (ko) 디스플레이장치 및 그 제어방법
WO2019107145A1 (ja) 情報処理装置、及び情報処理方法
US11861265B2 (en) Providing audio information with a digital assistant
CN112185369B (zh) 一种基于语音控制的音量调节方法、装置、设备和介质
CN116168125A (zh) 一种虚拟形象和场景控制方法、系统、智能座舱及其车辆
WO2024055566A1 (zh) 一种车载交互方法、系统、控制器和汽车
WO2022198365A1 (zh) 一种语音控制方法及装置
JP2002520681A (ja) 自動音声認識方法
WO2023098564A1 (zh) 一种语音助手显示方法及相关装置
WO2023231538A1 (zh) 一种驾驶状态的显示方法、相关装置、设备以及存储介质
JP2020198077A (ja) 電子デバイスの音声制御方法、電子デバイスの音声制御装置、コンピュータ機器及び記憶媒体
JP2020134627A (ja) 情報処理装置、情報処理方法およびプログラム
EP1649451A1 (en) Method of controlling a dialoging process
JP7407047B2 (ja) 音声出力の制御方法および音声出力制御装置
WO2023246894A1 (zh) 语音交互方法及相关装置
WO2024022111A1 (zh) 语音交互的方法及终端设备
CN116844540A (zh) 车载语音助手互动方法、装置、电子设备及存储介质
US20240205628A1 (en) Spatial Audio for Device Assistants
WO2022193735A1 (zh) 显示设备及语音交互方法
CN117809625A (zh) 一种终端设备及双模型校验的唤醒方法
Winter et al. Interaction styles in use of automotive interfaces
CN117253481A (zh) 车载娱乐系统的交互方法、装置、存储介质及车辆

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23864309

Country of ref document: EP

Kind code of ref document: A1