WO2024131126A1 - Procédé et appareil de commande vocale et véhicule - Google Patents

Procédé et appareil de commande vocale et véhicule Download PDF

Info

Publication number
WO2024131126A1
WO2024131126A1 PCT/CN2023/115948 CN2023115948W WO2024131126A1 WO 2024131126 A1 WO2024131126 A1 WO 2024131126A1 CN 2023115948 W CN2023115948 W CN 2023115948W WO 2024131126 A1 WO2024131126 A1 WO 2024131126A1
Authority
WO
WIPO (PCT)
Prior art keywords
vehicle
voice
information
speech
scene information
Prior art date
Application number
PCT/CN2023/115948
Other languages
English (en)
Chinese (zh)
Inventor
梁颖琪
刘权
张俊仪
黄天
柳颖睿
陈力
Original Assignee
广州汽车集团股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 广州汽车集团股份有限公司 filed Critical 广州汽车集团股份有限公司
Publication of WO2024131126A1 publication Critical patent/WO2024131126A1/fr

Links

Definitions

  • the present application relates to the field of vehicle technology, and more specifically, to a voice control method, device and vehicle.
  • the present application proposes a voice control method, device and vehicle to improve the above problems.
  • the present application provides a voice control method applied to a vehicle, the method comprising: in response to the triggering of a voice output task, obtaining current scene information of the vehicle, the scene information characterizing the emotions of a user in the vehicle and the current state of the vehicle; sending the scene information to a voice generation platform so that the voice generation platform synthesizes emotional speech based on the scene information, the emotional speech being a speech that matches the user's emotions and the voice output task; and receiving and outputting the emotional speech.
  • the present application provides a voice control method, which is applied to a voice generation platform, and the method includes: receiving scene information, wherein the scene information is sent by a vehicle in response to a voice output task trigger, and the scene information characterizes the emotions of a user in the vehicle and the current state of the vehicle; based on the scene information, obtaining emotional speech, wherein the emotional speech is a speech that matches the user's emotions and the voice output task; and sending the emotional speech to the vehicle.
  • the present application provides a voice control system, characterized in that the system includes a vehicle and a voice generation platform, wherein: the vehicle is used to obtain current scene information of the vehicle in response to the triggering of a voice output task, the scene information characterizing the emotions of the users in the vehicle and the current state of the vehicle, and sending the scene information to the voice generation platform; the voice generation platform is used to receive the scene information, obtain emotional voice based on the scene information, and send the emotional voice to the vehicle.
  • the present application provides a voice control device, which runs on a vehicle, and the device includes: a scene information acquisition unit, which is used to obtain the current scene information of the vehicle in response to the triggering of the voice output task, wherein the scene information represents the emotion of the user in the vehicle and the current state of the vehicle; a scene information sending unit, which is used to send the scene information to a voice generation platform, so that the voice generation platform synthesizes emotional voice based on the scene information, and the emotional voice is a voice that matches the emotion of the user and the voice output task;
  • the speech receiving unit is used to receive and broadcast the emotional speech.
  • the present application provides a voice control device, which runs on a voice generation platform, and the device includes: a scene information receiving unit, which is used to receive scene information, wherein the scene information represents the emotion of the user in the vehicle and the current state of the vehicle. state; a speech generating unit for obtaining emotional speech based on the scene information; and a speech sending unit for sending the emotional speech to the vehicle.
  • the present application provides a vehicle comprising one or more processors and a memory; one or more programs are stored in the memory and configured to be executed by the one or more processors, and the one or more programs are configured to execute the above-mentioned method.
  • the present application provides a computer-readable storage medium, in which a program code is stored, wherein the above method is executed when the program code is run.
  • the present application provides a voice control method, device, vehicle and storage medium. After responding to the triggering of a voice output task, the scene information of the vehicle that characterizes the emotions of the user in the vehicle and the current state of the vehicle is obtained, and the scene information is sent to a voice generation platform, so that the voice generation platform synthesizes a voice that matches the user's emotions and the voice output task based on the scene information, and receives and outputs the emotional voice.
  • the scene information that characterizes the current state of the vehicle and the emotions of the user in the vehicle can be obtained, and by sending the scene information to the voice generation platform, the emotional voice that matches the user's emotions and the voice output task can be obtained, thereby improving the emotional level of voice control and further improving the user experience.
  • FIG1 shows a flow chart of a voice control method proposed in an embodiment of the present application
  • FIG2 shows a schematic diagram of a vehicle according to an embodiment of the present application
  • FIG3 shows a schematic diagram of a voice output task triggering proposed in an embodiment of the present application
  • FIG4 is a schematic diagram showing another voice output task triggering method proposed in an embodiment of the present application.
  • FIG5 shows a flow chart of a voice control method proposed in another embodiment of the present application.
  • FIG6 shows a schematic diagram of a speech generation platform proposed in the present application.
  • FIG7 shows a schematic diagram of a generic user emotional model proposed in this application.
  • FIG8 shows a flow chart of a voice control method proposed in yet another embodiment of the present application.
  • FIG9 shows a schematic diagram of interaction between a vehicle and a speech generation platform proposed in the present application.
  • FIG10 shows a structural block diagram of a voice control device proposed in an embodiment of the present application.
  • FIG11 shows a structural block diagram of another voice control device proposed in an embodiment of the present application.
  • FIG12 shows a structural block diagram of a vehicle proposed in the present application
  • FIG13 shows a structural block diagram of a voice control system proposed in the present application.
  • FIG. 14 is a storage unit according to an embodiment of the present application for storing or carrying program codes for implementing the voice control method according to an embodiment of the present application.
  • the inventors proposed a voice control method, device and vehicle in the present application. After responding to the triggering of a voice output task, the scene information of the vehicle characterizing the emotions of the user in the vehicle and the current state of the vehicle is obtained, and the scene information is sent to the voice generation platform, so that the voice generation platform synthesizes a voice matching the user's emotions and the voice output task based on the scene information, and receives and outputs the emotional voice.
  • the scene information characterizing the current state of the vehicle and the emotions of the user in the vehicle can be obtained, and by sending the scene information to the voice generation platform, the emotional voice matching the user's emotions and the voice output task can be obtained, thereby improving the emotional level of voice control and further improving the user experience.
  • a voice control method provided by the present application is applied to a vehicle, and the method includes:
  • the current scene data of the vehicle can be obtained, wherein the scene data includes target driving status data, target entertainment information data, and target cabin environment monitoring data; based on the target driving status data, the target entertainment information data and the target cabin environment monitoring data, the scene information is obtained.
  • the target driving state data can characterize the driving state of the vehicle.
  • the target driving state data can include driving speed, driving mode, intelligent driving function activation state, fuel information, power information, in-vehicle ambient light activation state, ambient light color state, and ambient light mode state.
  • the target entertainment information data can characterize the use of other functions of the vehicle in addition to the basic driving function.
  • the target entertainment information data can include vehicle identification, user identification corresponding to the vehicle, navigation information, voice preference information, user speech speed, user voice emotion, time information, trigger service, user schedule information, etc., wherein the navigation information can include road condition information and destination, and the voice preference information can include preset timbre, preset speech speed, etc.
  • the target cabin environment monitoring data can characterize the in-vehicle environment and the situation of the driver and passengers.
  • the target cabin environment monitoring data can include user identity feature information, weather information, etc.
  • the user identity feature information can include the driver's gender, age, emotional state, and the passenger's gender, age, etc.
  • reference driving status data, reference entertainment information data, and reference cabin environment monitoring data can be obtained; the reference driving status data, reference entertainment information data, and reference cabin environment monitoring data are each filtered and analog-to-digital converted to obtain target driving status data, target entertainment information data, and target cabin environment monitoring data.
  • the vehicle may include an in-vehicle information module, a data processing module, and a network module.
  • the vehicle can obtain reference driving state data, reference entertainment information data, and reference cabin environment monitoring data through the in-vehicle information module.
  • the in-vehicle information module can obtain reference driving state data based on multiple ECUs (Electronic Control Units) and sensors in the vehicle; obtain reference entertainment information data based on relevant applications loaded in the vehicle (such as navigation applications, voice recognition applications, etc.); and obtain reference cabin environment monitoring data based on image acquisition devices, audio acquisition devices, and relevant applications (such as weather applications, etc.).
  • ECUs Electronic Control Units
  • relevant applications loaded in the vehicle such as navigation applications, voice recognition applications, etc.
  • relevant applications such as weather applications, etc.
  • the vehicle can filter and perform analog-to-digital conversion operations on the reference driving status data, reference entertainment information data and reference cabin environment monitoring data through a data processing module to obtain target driving status data, target entertainment information data and target cabin environment monitoring data.
  • the vehicle identification, user identification, navigation information, voice preference information, driving status data, and identity feature information may be associated and processed to obtain scene information.
  • the vehicle can use the data processing module to combine the vehicle identification, user identification, navigation information, voice preference information, Driving status data, identity feature information, and weather information are correlated and processed to obtain scene information.
  • the vehicle's voice output task can be triggered by the user sending voice to the vehicle's automatic speech recognition system (Automatic Speech Recognition, ASR).
  • ASR Automatic Speech Recognition
  • the voice recognition system can recognize the user's voice content and trigger the vehicle's voice output task, so that the vehicle can obtain reference driving status data, reference entertainment information data and reference cabin environment monitoring data from multiple ECUs, sensors, and related applications of the vehicle through the in-vehicle module; then the reference driving status data, reference entertainment information data and reference cabin environment monitoring data are processed by the data processing module to obtain scene information; then the scene information is sent to the voice generation platform through the network module, so that the voice generation platform can obtain emotional voice based on the natural language processing engine and the voice synthesis engine.
  • the emotional voice can be a voice that suddenly indicates that the user has planned a route and can set off, so that the vehicle can receive and broadcast the emotional voice.
  • the vehicle can actively trigger the voice output task.
  • the voice output task can be actively triggered, so that the vehicle can obtain reference driving state data, reference entertainment information data and reference cabin environment monitoring data from multiple ECUs, sensors and related applications of the vehicle through the in-vehicle module; then the reference driving state data, reference entertainment information data and reference cabin environment monitoring data are processed by the data processing module to obtain scene information; then the scene information is sent to the voice generation platform through the network module, so that the voice generation platform can obtain emotional voice based on the natural language processing engine and the voice synthesis engine, and the emotional voice can be a voice that seriously reminds the user to slow down, so that the vehicle can receive and broadcast the emotional voice.
  • S120 Send the scene information to a speech generation platform, so that the speech generation platform synthesizes emotional speech based on the scene information, where the emotional speech is a speech that matches the user's emotion and the speech output task.
  • emotional speech can refer to TTS (Text To Speech)
  • emotional speech can correspond to text information
  • the text information can include synthesis rules and speech content of emotional speech
  • the speech content can refer to the content that needs to be output determined by the vehicle based on the current speech output task
  • the synthesis rules can refer to the voice style, intonation coefficient, speaking speed coefficient, and pitch coefficient of the speech.
  • the vehicle can send the scene information to the voice generation platform through the network module.
  • the speech generation platform may be a cloud platform or a platform deployed in a vehicle.
  • S130 Receive and output the emotional speech.
  • the vehicle can obtain emotional voice through the network module and play the emotional voice through an audio playback device (such as a microphone, car audio, etc.) so that the user and the vehicle can achieve information interaction.
  • an audio playback device such as a microphone, car audio, etc.
  • the vehicle can store the text information corresponding to the emotional speech, and in response to the re-triggering of the speech output task, can generate the emotional speech based on the text information and broadcast the emotional speech.
  • the voice control method proposed in this application can be optimized based on the stored text information to generate emotional voice that better fits the user's mood.
  • a voice control method provided in this embodiment obtains scene information of the vehicle representing the emotion of the user in the vehicle and the current state of the vehicle in response to the triggering of the voice output task, and sends the scene information to a voice generation platform so that the voice generation platform synthesizes the scene information based on the scene information and the emotion of the user and the voice output.
  • the above method can obtain scene information representing the current state of the vehicle and the emotions of the user in the vehicle in response to the triggering of the voice output task, and can obtain emotional voice matching the user's emotions and the voice output task by sending the scene information to the voice generation platform, thereby improving the emotional level of voice control and further improving the user experience.
  • a voice control method provided by the present application is applied to a voice generation platform, and the method includes:
  • S210 Receive scene information, where the scene information is sent by the vehicle in response to a voice output task trigger, and the scene information represents an emotion of a user in the vehicle and a current state of the vehicle.
  • the speech generation platform may include an emotion library, a matching library, and a speech synthesis module, and the emotion library may be used to generate speech emotion information.
  • the speech generation platform can receive scene information from the vehicle through the emotion library.
  • S220 Obtain emotional speech based on the scene information, where the emotional speech is speech that matches the user's emotion and the speech output task.
  • the speech emotion information may include speech style, intonation coefficient, speaking speed coefficient, and pitch coefficient; then obtain the output content corresponding to the speech output task; and then obtain emotional speech based on the output content and the speech emotion information.
  • the generic user emotional model can be a convolutional neural network.
  • the output content corresponding to the voice output task can refer to the content that the vehicle needs to feedback to the user based on the current voice output task.
  • the voice style can include happy, cheerful, ordinary, quiet, sad, angry, and serious; the intonation coefficient can be divided into 1 to 5, and the higher the value, the higher the intonation; the speed coefficient can be divided into 1 to 5, and the higher the value, the faster the speed; the pitch coefficient can be divided into 1 to 5, and the higher the value, the louder the sound.
  • training data may be obtained, and the training data may include scene information of multiple vehicles; the initial generic user emotional model is trained based on the training data to obtain a generic user emotional model.
  • training data can be obtained from the network.
  • the emotion library can obtain reference voice emotion information based on scene information and a pre-trained generic user emotional model, wherein the reference voice emotion information may include the user's emotion type, emotion intensity coefficient, speaking speed coefficient, and the vehicle's in-car environment comfort and driving safety index;
  • the matching library can obtain voice emotion information based on the reference voice emotion information and pre-set mapping rules;
  • the speech synthesis module can obtain emotional speech based on the output content and voice emotion information.
  • the emotion types can include happy, cheerful, normal, quiet, sad, angry, and serious; the emotion intensity coefficient can be divided into 1 to 5, and the higher the value, the more intense the emotion; the speech speed coefficient can be divided into 0.5, 1, 1.5, and 2. The higher the value, the faster the speaking speed; the in-car environment comfort level can be divided into 0.5, 1, 1.5, and 2. The higher the value, the more comfortable and faster the environment is; the driving safety index can be divided into 0.5, 1, 1.5, and 2. The higher the value, the safer it is.
  • the preset mapping rule may refer to the correspondence between the reference voice emotion information and the voice emotion information. For example, when the user's emotion type is happy, the voice style of the mapped voice emotion information may be happy; when the user's driving safety index is low, the voice style of the mapped voice emotion information may be serious, and the intonation coefficient is increased, the speech speed is increased, and the pitch is increased.
  • the speech generation platform can send emotional speech to the vehicle through the network module.
  • a voice control method provided in this embodiment can obtain emotional voice that matches the user's emotions and voice output tasks based on the received scene information through the above-mentioned method, thereby improving the emotional level of voice control and further improving the user experience.
  • a voice control system 300 provided in the present application is characterized in that the system includes a vehicle 310 and a voice generation platform 320, wherein:
  • Step S310 The vehicle is used to obtain current scene information of the vehicle in response to triggering of the voice output task, wherein the scene information represents the emotion of the user in the vehicle and the current state of the vehicle, and send the scene information to the voice generation platform.
  • the vehicle may include an in-vehicle information module, a data processing module, and a network module.
  • the vehicle may be used to obtain the current scene information of the vehicle through the in-vehicle information module and the data processing module in response to the triggering of the voice output task, and then send the scene information to the voice generation platform through the network module.
  • Step S320 The speech generation platform is used to receive the scene information, obtain emotional speech based on the scene information, and send the emotional speech to the vehicle.
  • the speech generation platform may include an emotion library, a matching library, and a speech synthesis module.
  • the speech generation platform may receive scene information through the emotion library, obtain emotional speech through the emotion library, the matching library, and the speech synthesis module, and send the emotional speech to the vehicle.
  • a voice control system provided by this embodiment enables the vehicle to respond to the triggering of a voice output task and obtain scene information that characterizes the current state of the vehicle and the emotions of the users in the vehicle through the above-mentioned method, and by sending the scene information to the voice generation platform, emotional voice that matches the user emotions and voice output task can be obtained, thereby improving the emotional level of voice control and further enhancing the user experience.
  • the present application provides a voice control device 400, which is operated in a vehicle, and the device 400 includes:
  • the scene information acquisition unit 410 is used to acquire the current scene information of the vehicle in response to the triggering of the voice output task, wherein the scene information represents the emotion of the user in the vehicle and the current state of the vehicle.
  • the scene information sending unit 420 is used to send the scene information to the speech generation platform so that the speech generation platform synthesizes emotional speech based on the scene information, and the emotional speech is a speech that matches the user's emotion and the speech output task.
  • the voice receiving unit 430 is used to receive and broadcast the emotional voice.
  • the scene information acquisition unit 410 is specifically used to obtain the current scene data of the vehicle in response to the triggering of the voice output task, and the scene data includes target driving status data, target entertainment information data, and target cabin environment monitoring data; based on the target driving status data, the target entertainment information data and the target cabin environment monitoring data, the scene information is obtained.
  • the scene information acquisition unit 410 is specifically used to obtain reference driving status data, reference entertainment information data, and reference cabin environment monitoring data in response to the triggering of the voice output task; and perform filtering and analog-to-digital conversion operations on the reference driving status data, the reference entertainment information data, and the reference cabin environment monitoring data to obtain target driving status data, target entertainment information data, and target cabin environment monitoring data.
  • the target entertainment information data includes a vehicle identification, a user identification corresponding to the vehicle, navigation information, and voice preference information;
  • the target cabin environment monitoring data includes identity characteristic information of the user;
  • the scene information acquisition unit 410 is specifically used to associate the vehicle identification, the user identification with the navigation information, the voice preference information, the driving status data, and the identity characteristic information to obtain the scene information.
  • the voice receiving unit 430 is specifically used to store the text information; in response to the re-triggering of the voice output task, generate the emotional voice based on the text information; and broadcast the emotional voice.
  • a voice control device 600 provided in the present application runs on a voice generation platform, and the device 600 includes:
  • the scene information receiving unit 610 is used to receive scene information, where the scene information represents the emotions of the user in the vehicle and the current state of the vehicle.
  • the speech generation unit 620 is used to obtain emotional speech based on the scene information.
  • the voice sending unit 630 is used to send the emotional voice to the vehicle.
  • the speech generation unit 620 is specifically used to obtain speech emotion information based on the scene information and a pre-trained generic user emotional model, and the speech emotion information includes speech style, intonation coefficient, speaking speed coefficient, and pitch coefficient; obtain the output content corresponding to the speech output task; and obtain the emotional speech based on the output content and the speech emotion information.
  • the speech generation unit 620 is specifically used to obtain reference speech emotion information based on the scene information and a pre-trained generic user emotional model, wherein the reference speech emotion information includes the user's emotion type, emotion intensity coefficient, speaking speed coefficient, and the vehicle's in-vehicle environment comfort and driving safety index; and obtain the speech emotion information based on the reference speech emotion information and pre-set mapping rules.
  • the device 600 further includes:
  • the model training unit 640 is used to obtain training data, where the training data includes scene information of multiple vehicles; and train the initial generic user emotional model based on the training data to obtain the generic user emotional model.
  • the vehicle 100 includes one or more (only one is shown in the figure) processors 102, a memory 104, and a network module 106 that are coupled to each other.
  • the memory 104 stores a program that can execute the content of the above-mentioned embodiment, and the processor 102 can execute the program stored in the memory 104.
  • the processor 102 may include one or more processing cores.
  • the processor 102 uses various interfaces and lines to connect various parts of the entire vehicle 100, and executes various functions and processes data of the vehicle 100 by running or executing instructions, programs, code sets or instruction sets stored in the memory 104, and calling data stored in the memory 104.
  • the processor 102 can be implemented in at least one hardware form of digital signal processing (DSP), field-programmable gate array (FPGA), and programmable logic array (PLA).
  • DSP digital signal processing
  • FPGA field-programmable gate array
  • PDA programmable logic array
  • the processor 102 can integrate one or a combination of a central processing unit (CPU), a graphics processing unit (GPU), and a modem.
  • the CPU mainly processes the operating system, user interface, and application programs; the GPU is responsible for rendering and drawing display content; and the modem is used to process wireless communications. It is understandable that the above-mentioned modem may not be integrated into the processor 102, but may be implemented separately through a communication chip.
  • the memory 104 may include a random access memory (RAM) or a read-only memory (ROM).
  • the memory 104 may be used to store instructions, programs, codes, code sets or instruction sets.
  • the memory 104 may include a program storage area and a data storage area, wherein the program storage area may store instructions for implementing an operating system, instructions for implementing at least one function (such as a touch function, a sound playback function, an image playback function, etc.), instructions for implementing the following various method embodiments, etc.
  • the data storage area may also store data created by the terminal 100 during use (such as a phone book, audio and video data, chat record data), etc.
  • the network module 106 is used to implement information interaction between the vehicle 100 and the speech generation platform, for example, transmission of device control instructions, manipulation request instructions, and status information acquisition instructions, etc.
  • the speech generation platform is specifically a different device, the corresponding network module 106 may be different.
  • the present application provides a voice control system 300 , wherein the system 300 includes a vehicle 310 and a voice generation platform 320 .
  • FIG 14 shows a block diagram of a computer-readable storage medium provided in an embodiment of the present application.
  • the computer-readable storage medium 800 stores program codes, which can be called by a processor to execute the method described in the above method embodiment.
  • the computer-readable storage medium 800 may be an electronic memory such as a flash memory, an EEPROM (electrically erasable programmable read-only memory), an EPROM, a hard disk, or a ROM.
  • the computer-readable storage medium 800 includes a non-transitory computer-readable storage medium.
  • the computer-readable storage medium 800 has storage space for program code 810 for executing any method steps in the above method. These program codes may be read from or written to one or more computer program products.
  • the program code 810 may be compressed, for example, in a suitable form.
  • the present application provides a voice control method, device, and vehicle, which, in response to the triggering of a voice output task, obtains scene information of the vehicle that characterizes the emotions of the user in the vehicle and the current state of the vehicle, and sends the scene information to a voice generation platform, so that the voice generation platform synthesizes a voice that matches the user's emotions and the voice output task based on the scene information, receives and outputs the emotional voice.
  • a voice generation platform By sending scene information to the speech generation platform, emotional speech that matches the user's emotions and speech output tasks can be obtained, thereby improving the emotional level of voice control and further improving the user experience.

Landscapes

  • User Interface Of Digital Computer (AREA)

Abstract

L'invention concerne un procédé et un appareil de commande vocale, et un véhicule. Le procédé consiste : à la suite d'un déclenchement d'une tâche de sortie vocale, à acquérir des informations de scène actuelle d'un véhicule, les informations de scène représentant l'émotion d'un utilisateur dans le véhicule et l'état actuel du véhicule (S110) ; à envoyer les informations de scène à une plate-forme de génération de voix, de telle sorte que la plate-forme de génération de voix synthétise une voix émotionnelle sur la base des informations de scène, la voix émotionnelle étant une voix correspondant à l'émotion de l'utilisateur et à la tâche de sortie vocale (S120) ; et à recevoir et à prononcer la voix émotionnelle (S130). Au moyen du procédé décrit, le niveau émotionnel de commande vocale peut être amélioré, ce qui permet d'améliorer l'expérience de l'utilisateur.
PCT/CN2023/115948 2022-12-23 2023-08-30 Procédé et appareil de commande vocale et véhicule WO2024131126A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211666185.0A CN118248117A (zh) 2022-12-23 2022-12-23 语音控制方法、装置以及车辆
CN202211666185.0 2022-12-23

Publications (1)

Publication Number Publication Date
WO2024131126A1 true WO2024131126A1 (fr) 2024-06-27

Family

ID=91551626

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/115948 WO2024131126A1 (fr) 2022-12-23 2023-08-30 Procédé et appareil de commande vocale et véhicule

Country Status (2)

Country Link
CN (1) CN118248117A (fr)
WO (1) WO2024131126A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20170116441A (ko) * 2016-04-11 2017-10-19 한국전자통신연구원 감정 운전 예방 시스템 및 그 방법
CN111402925A (zh) * 2020-03-12 2020-07-10 北京百度网讯科技有限公司 语音调节的方法、装置、电子设备、车载系统和可读介质
CN113734187A (zh) * 2020-05-29 2021-12-03 逸驾智能科技有限公司 用于与车辆用户进行信息交互的方法、装置及车机

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20170116441A (ko) * 2016-04-11 2017-10-19 한국전자통신연구원 감정 운전 예방 시스템 및 그 방법
CN111402925A (zh) * 2020-03-12 2020-07-10 北京百度网讯科技有限公司 语音调节的方法、装置、电子设备、车载系统和可读介质
CN113734187A (zh) * 2020-05-29 2021-12-03 逸驾智能科技有限公司 用于与车辆用户进行信息交互的方法、装置及车机

Also Published As

Publication number Publication date
CN118248117A (zh) 2024-06-25

Similar Documents

Publication Publication Date Title
US10170111B2 (en) Adaptive infotainment system based on vehicle surrounding and driver mood and/or behavior
EP2140341B1 (fr) Système et procédé d'information à caractère émotionnel
JP7192222B2 (ja) 発話システム
JP2019139354A (ja) 情報提供装置及び情報提供方法
CN109302486B (zh) 一种根据车内环境推送音乐的方法和系统
EP3886086A1 (fr) Assistant personnel virtuel sensible aux émotions
CN112959998A (zh) 一种车载人机交互方法、装置、车辆及电子设备
CN111354359A (zh) 一种车辆语音控制方法、装置、设备、系统及介质
CN110265009B (zh) 一种基于用户身份的主动对话发起方法和装置
CN111192583B (zh) 控制装置、智能体装置以及计算机可读存储介质
WO2024131126A1 (fr) Procédé et appareil de commande vocale et véhicule
CN116483305A (zh) 智能网联汽车数字虚拟人应用系统及其使用方法、车辆
CN115240677A (zh) 用于车辆座舱的语音交互方法、装置和设备
CN115527542A (zh) 车载语音助手的设计方法、装置、终端设备以及存储介质
CN115107672A (zh) 一种车辆控制方法及车辆
CN114734942A (zh) 调节车载音响音效的方法及装置
CN115214696A (zh) 一种车机虚拟形象交互方法、系统、车辆及存储介质
JP2006069358A (ja) 車両の運転支援装置
CN113035181A (zh) 语音数据处理方法、设备和系统
US20240025416A1 (en) In-vehicle soundscape and melody generation system and method using continuously interpreted spatial contextualized information
CN114708864A (zh) 语音助手的信息控制方法、装置、计算机设备和存储介质
CN114291008A (zh) 车辆用智能体装置、车辆用智能体系统以及计算机可读存储介质
CN116343821A (zh) 一种车用基于用户信息进行对话的方法及装置
CN116564301A (zh) 一种基于用户情绪自动推荐使用智能设备的方法和系统
WO2022238126A1 (fr) Procédé et unité de communication pour augmenter une communication