CN106803423B - Man-machine interaction voice control method and device based on user emotion state and vehicle - Google Patents

Man-machine interaction voice control method and device based on user emotion state and vehicle Download PDF

Info

Publication number
CN106803423B
CN106803423B CN201611229157.7A CN201611229157A CN106803423B CN 106803423 B CN106803423 B CN 106803423B CN 201611229157 A CN201611229157 A CN 201611229157A CN 106803423 B CN106803423 B CN 106803423B
Authority
CN
China
Prior art keywords
user
state
voice
set user
emotional state
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201611229157.7A
Other languages
Chinese (zh)
Other versions
CN106803423A (en
Inventor
沈海寅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhicheauto Technology Beijing Co ltd
Original Assignee
Zhicheauto Technology Beijing Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhicheauto Technology Beijing Co ltd filed Critical Zhicheauto Technology Beijing Co ltd
Priority to CN201611229157.7A priority Critical patent/CN106803423B/en
Publication of CN106803423A publication Critical patent/CN106803423A/en
Application granted granted Critical
Publication of CN106803423B publication Critical patent/CN106803423B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L2015/088Word spotting
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/226Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
    • G10L2015/227Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of the speaker; Human-factor methodology

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Child & Adolescent Psychology (AREA)
  • General Health & Medical Sciences (AREA)
  • Hospice & Palliative Care (AREA)
  • Psychiatry (AREA)
  • Signal Processing (AREA)
  • Traffic Control Systems (AREA)
  • Navigation (AREA)
  • Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)

Abstract

The invention discloses a human-computer interaction voice control method, a human-computer interaction voice control device and a vehicle based on the emotional state of a user, wherein the method comprises the following steps: monitoring and setting the expression, voice or action of a user; determining the current emotional state of the set user according to the expression, voice or action of the set user; determining a voice control mode of the vehicle according to the current emotional state of the set user; and performing vehicle man-machine interaction according to the determined voice control mode. The method, the device and the vehicle disclosed by the invention can calculate the current emotion of the user according to the driving behavior, the speaking speed and tone, the facial expression and the like of the user, and the intelligent system can play appropriate music or adjust the change of the navigation voice and the like according to the current emotion state of the user to carry out man-machine interaction with the user, so that the emotion of the user is adjusted, and the driving of the user is safer.

Description

Man-machine interaction voice control method and device based on user emotion state and vehicle
Technical Field
The invention relates to the field of artificial intelligence, in particular to the field of vehicle intelligent control or human-computer interaction, and particularly relates to a human-computer interaction voice control method and device based on a user emotion state and a vehicle.
Background
With the rapid development of society, automobiles are more and more popular in life; although the concept of vehicle autopilot has been long proposed, it has not yet become widespread; currently, the control of the driver is still in a decisive position during the driving of the vehicle. However, a person as a driver may be affected by various emotions during driving, and some emotions may seriously affect driving safety.
Therefore, it is necessary to provide a method or a vehicle capable of analyzing the emotion of a driver.
Disclosure of Invention
The invention aims to solve the technical problem of how to provide a method, an intelligent control system and a vehicle for man-machine interaction based on the emotion of a user aiming at a vehicle environment, so that the emotion of the user can be adjusted, and then the man-machine interaction of the vehicle is controlled based on the emotion state of the user, thereby ensuring the safety of the user in driving the vehicle.
The invention provides a human-computer interaction voice control method based on the emotional state of a user, which comprises the following steps: monitoring and setting the expression, voice or action of a user; determining the current emotional state of the set user according to the expression, voice or action of the set user; determining a voice control mode of the vehicle according to the current emotional state of the set user; and performing vehicle man-machine interaction according to the determined voice control mode.
Further, before determining the current emotional state of the set user according to the expression, voice or motion of the set user, the method includes: counting emotional state data of a plurality of users to form a general user emotional state database, wherein the emotional state database comprises the relation between expressions, voice or actions of the users and emotional states; analyzing the emotional states of a plurality of users in a big data mode according to a general user emotional state database, and calibrating a determination factor of the emotional states of the users according to the actions of the users, wherein the emotional states of the users comprise a cheerful state, an angry state, a sad state, a painful state and an excited state.
Further, the determining the current emotional state of the set user according to the expression, voice or motion of the set user includes: and evaluating and determining the current emotional state of the set user according to the general user emotional state database and the monitored expression, voice or action of the set user.
Further, the method also comprises the step of determining the control mode of the vehicle according to the current emotional state of the set user.
Further, determining the current emotional state of the setting user according to the expression, voice or motion of the setting user further includes: acquiring the change of the speed, tone and sound of the voice of the set user; and determining the current emotional state of the set user according to the average value of the speed, tone and sound of the voice of the set user.
Further, language information contained in the sound in the time period set by the set user is analyzed, and whether specific words and phrases exist in the language expression of the set user is judged; and determining the emotional state of the set user according to the specific words and phrases appearing in the current expression of the set user.
Further, analyzing the action of a set user in a set time period; and judging and setting the current emotional state of the user according to the daily action and the action of the current time of the set user.
Further, the driving behavior of a set user in a set time period is analyzed; and judging and setting the current emotional state of the user according to the daily driving behavior of the set user and the driving behavior at the current time.
Further, analyzing whether the speed, tone and sound of the voice of the set user in the set time period exceed the set threshold; if the speed of speech of the set user is increased to exceed the set speed of speech threshold, and/or the intonation is increased to exceed the set intonation threshold, and/or the voice is increased to exceed the set voice threshold, the set user is in a relative emotional state.
Further, the determining the voice control mode of the vehicle according to the current emotional state of the set user includes: and if the user is set to be in an angry state at present, selecting a gentle and comfortable sound database to perform voice control, and soothing the emotion of the set user.
And further, selecting the set human-computer interaction sound effect according to the emotion state of the set user to interact with the set user.
Further, after the emotional state of the set user is determined, the set music is selected to be played according to the emotional state of the set user so as to sooth the emotion of the set user and ensure the driving safety.
Further, if the driver is currently in the driving navigation state, the set navigation language can be selected to be played according to the set emotion state of the user so as to sooth the emotion of the user and ensure the driving safety.
The invention also provides a human-computer interaction voice control device based on the emotion state of the user, which comprises the following components: the monitoring module is used for monitoring and setting the expression, voice or action of the user; the emotion state analysis module is connected with the monitoring module and used for determining the current emotion state of the set user according to the expression, voice or action of the set user; the processing module is connected with the emotional state analysis module and used for determining a voice control mode of the vehicle according to the current emotional state of the set user; and the execution module is connected with the processing module and is used for carrying out vehicle man-machine interaction according to the determined voice control mode.
The system further comprises an emotional state database module, a general user emotional state database and a user information database, wherein the emotional state database module is used for counting emotional state data of a plurality of users to form the general user emotional state database, and the emotional state database comprises the relation between expressions, voice or actions of the users and emotional states; analyzing the emotional states of a plurality of users in a big data mode according to a general user emotional state database, and calibrating a determination factor of the emotional states of the users according to the actions of the users, wherein the emotional states of the users comprise a cheerful state, an angry state, a sad state, a painful state and an excited state.
Further, the emotion state analysis module is also used for evaluating and determining the emotion state of the set user according to the general user emotion state database and the monitored expression, voice or action of the set user.
Further, the processing module is also used for determining the control mode of the vehicle according to the current emotional state of the set user.
Further, the emotional state analysis module is further configured to: acquiring the change of the speed, tone and sound of the voice of the set user; and determining the current emotional state of the set user according to the average value of the speed, tone and sound of the voice of the set user.
Further, language information contained in the voice in the time period set by the set user is analyzed, and whether specific words and sentences exist in the language expression of the set user is judged; and determining the emotional state of the set user according to the specific words and phrases appearing in the current expression of the set user.
Further, analyzing the action of a set user in a set time period; and judging and setting the current emotional state of the user according to the daily action and the action of the current time of the set user.
Further, the driving behavior of a set user in a set time period is analyzed; and judging and setting the current emotional state of the user according to the daily driving behavior of the set user and the driving behavior at the current time.
Further, analyzing whether the speed, tone and sound of the voice of the set user in the set time period exceed the set threshold; if the speed of speech of the set user is increased to exceed the set speed of speech threshold, and/or the intonation is increased to exceed the set intonation threshold, and/or the voice is increased to exceed the set voice threshold, the set user is in a relative emotional state.
Further, the processing module is also used for selecting a gentle and comforting sound database for voice control and soothing the emotion of the set user if the set user is in an angry state at present.
Furthermore, the execution module is also used for selecting the set human-computer interaction sound effect according to the emotion state of the set user to interact with the set user.
Further, the execution module is further used for selecting the set music to play according to the emotional state of the set user after the emotional state of the set user is determined so as to sooth the emotion of the set user and ensure driving safety.
Further, the execution module is also used for selecting the set navigation language to play according to the set emotional state of the user to placate the emotion of the set user to ensure the driving safety if the driver is currently in the driving navigation state.
The invention provides a vehicle which comprises a man-machine interaction voice control device based on the emotion state of a user.
According to the method, the device and the vehicle provided by the invention, the current emotion of the user can be calculated according to the driving behavior, the speaking speed and tone, the facial expression and the like of the user, such as urgency comparison, excitement comparison, anger comparison, sadness comparison and the like, the intelligent system can play appropriate music or adjust the change of navigation voice and the like according to the current emotion state of the user to perform man-machine interaction with the user, so that the emotion of the user is adjusted, and the driving of the user is safer.
Drawings
Fig. 1 shows a flow chart of a man-machine interaction voice control method based on the emotional state of a user according to an embodiment of the invention.
Fig. 2 shows a block diagram of a human-computer interaction voice control device based on a user emotional state according to an embodiment of the invention.
Fig. 3 shows a block diagram of a vehicle according to an embodiment of the invention.
Detailed Description
The present invention now will be described more fully hereinafter with reference to the accompanying drawings, in which exemplary embodiments of the invention are shown.
Fig. 1 shows a flow chart of a man-machine interaction voice control method based on the emotional state of a user according to an embodiment of the invention, and referring to fig. 1, the method comprises the following steps:
step 101, monitoring and setting the expression, voice or action of a user.
In one embodiment, the expression, voice, or motion of a given user may be monitored or detected by a combination of various sensors.
For example, the expression, the action and the like of a user can be monitored in a vehicle through a fatigue driving camera built in the vehicle; the voice situation of the user can be set through detection of a microphone built in the vehicle.
And step 102, determining the current emotional state of the set user according to the expression, voice or action of the set user.
In one embodiment, emotional state data of a plurality of users can be counted to form a general user emotional state database, wherein the emotional state is determined according to expressions, voice or actions of the users; analyzing the emotional states of a plurality of users in a big data mode according to a general user emotional state database, and calibrating the determination factors of the happy, angry, sad, painful and excited emotional states according to the emotional states of the users; and evaluating the emotional state of the user according to the general user emotional state database and the monitored expression, voice or action of the set user.
In one embodiment, the emotion of the user can be broadly summarized into several emotions of happiness, anger, sadness and happiness. In actual life, the influence of the favorite state or the happy state on the driving of the user is relatively small; however, the emotional state of anger or sadness has a great influence on the driving behavior of the user, for example, the road anger group driving a car may cause a great safety hazard to the current driving user or other users on the road, and in real life, there are frequent accidents such as leaving car, parking hitter, and various traffic accidents caused by the driver's anger rush.
In one embodiment, the emotional state of the user, such as anger or sadness, whose influence on the driving state is greater than the happy or happy emotional state, can be set, the emotional state of the user, such as anger or sadness, is set by emphasis monitoring, and a corresponding control method is adopted in the voice interaction to adjust the emotion of the user.
In one embodiment, aiming at the environment in the vehicle, voice interaction can be an important means for interaction between a person and a vehicle intelligent system, and in the man-machine interaction process, the intelligent system can monitor the voice of a set user through a microphone to obtain the change of the speed, tone and sound size of the voice of the set user; and determining the current emotional state of the set user according to the average value of the speed, tone and sound of the voice of the set user.
In a specific embodiment, whether the speed, tone and sound size of the voice of the set user in the set time period exceed the set threshold value or not can be analyzed; if the speed of speech of the set user is increased to exceed the set speed of speech threshold, or the intonation is increased to exceed the set intonation threshold, and the voice is increased to exceed the set voice threshold, the set user is in a relative emotional state.
In one embodiment, whether a language expression of a set user in a set time period has a specific word or phrase is analyzed; and determining the emotional state of the set user according to the specific words and phrases appearing in the current expression of the set user. For example, it is possible to analyze whether a specific word or phrase, such as an expletive word or a happy word, is present in the language expression of the set user within the set time period.
For example, if a user's linguistic expression frequently has words of dirty, curer, or words of abuse, it may indicate that the user is likely to be in a relatively excited or angry state; for another example, if a sound such as a happy expression sentence or laugh appears in the language expression of the user, it can be said that the user may be currently in a relatively happy state.
In one embodiment, the current emotional state of the user can be determined by analyzing the changes of the speed, tone and sound size of the specific or common words and sentences. For example, in the vehicle human-computer interaction process, there may be some commonly used words and phrases such as a wake-up word, and the intelligent system may analyze changes of voices, tones and sound of the user when expressing the commonly used words according to expression modes of the wake-up words such as "hello, xiaozhi" or "Hi, xiaozhi" expressed by the user, so as to analyze the current emotional state of the user. For example, it is analyzed whether the user has emotional states of joy, anger, sadness, etc. when the user expresses these common words.
In one embodiment, the action of the set user in the set time period is analyzed, and the current emotional state of the set user is judged according to the daily action and the action of the current time of the set user.
In one embodiment, the driving behavior of a set user within a set time period is analyzed; and judging and setting the current emotional state of the user according to the daily driving behavior of the set user and the driving behavior at the current time. For example, the driving behavior of the user is analyzed, and if the user has a violent stepping on the accelerator, a violent tapping of the steering wheel, a leaving behavior, or the like, it can be determined that the user is currently angry.
And 103, determining a voice control mode of the vehicle according to the current emotional state of the set user.
And 104, performing vehicle man-machine interaction according to the determined voice control mode.
In one embodiment, if the user is set to be in an angry state at present, selecting a gentle and comforting sound database for voice control, and soothing the emotion of the set user; and selecting the set human-computer interaction sound effect according to the emotion state of the set user to interact with the set user.
In one embodiment, particularly in a scenario within a vehicle, it is an important function for a user to interact with the vehicle through speech. According to the interactive statistics of daily machines and people, a voice interactive database for user emotion state interaction is obtained, and influence factors of joy, anger, sadness and music can be scientifically analyzed and set when the database is established; and determining a man-machine interaction mode of the vehicle according to the current emotion of the user obtained by analysis so as to encourage, placate or prompt the current driver, so that the driver can calm down, and the driver is prevented from influencing driving safety due to bad emotion in driving.
For example, if the user is currently in an angry state, the voice interaction database of the angry state of the user is selected, and the driver can be soothed by the database, so that the emotion of the driver can be stabilized, and the influence of anger on the driving of the driver is prevented.
For example, if the user is currently in a sad state, the user sad state voice interaction database is selected, and the database is adopted to comfort the driver to prevent the driver from influencing the driving due to sadness.
In one embodiment, after determining the emotion of the user, the intelligent system can select the set music to play according to the emotion state of the user so as to sooth the emotion of the set user and ensure driving safety. For example, the user is currently in a cheerful emotional state, and may choose to play some happy and graceful music; if the user is in the angry state, some light music for soothing the nerves can be played.
In one embodiment, after determining the emotion of the user, if the user is currently in the driving navigation state, the intelligent system may select the set navigation language to play according to the set emotion state of the user to sooth the emotion of the set user to ensure driving safety.
For example, if the current emotional state of the user is a relative collapse sadness, for the male driver, the vehicle intelligence system may select a gentle female voice to interact with the user; for the female driver, the vehicle may select a magnetic male voice for human-machine interaction with the user.
As an example, when performing facial expression recognition, the information in the existing facial recognition database may be used in combination with the specific camera position in the vehicle to perform facial recognition. Firstly, the first step is face detection, namely determining the position in a vehicle lens to find the face position; secondly, detecting key points, namely accurately finding key points of facial contours such as eyebrows, ears and noses at the determined face position, and identifying the face so as to identify a set user; thirdly, determining the information of the set user based on the face recognition of the large-scale data; and fourthly, finding out the facial expression database of the set user, and identifying the current emotional state of the user based on the facial expression information database of the set user.
In one embodiment, the intelligent system determines the control mode of the vehicle based on the current emotional state of the set user. For example, the method can include increasing the warning value of the ADAS system, such as automatically increasing the safe following distance, automatically controlling the vehicle to keep a longer safe distance with vehicles in other lanes, shortening the preset reaction time of the user in the vehicle active safety system, and simultaneously automatically tightening a safety belt and warning the user through the vibration of a steering wheel, prompting the user to pay attention to the driving safety, so that the safety of the system can be improved.
In one embodiment, the control mode of the vehicle may be selected based on the current emotional state of the user after the emotional state of the user is determined. For example, if the user is currently in an angry state, the vehicle can automatically tighten the safety belt, and give an alarm through a vehicle voice system, so that the user is pacified and is prompted to pay attention to driving safety; if the user is still in an angry state, even an action of getting out of the vehicle or trying to bump others by the vehicle occurs, the vehicle intelligent operation system can temporarily take over the control of the vehicle and enable the vehicle to stop at the side; prevent the driver from being injured or self-injured due to anger. Even if the vehicle automatically selects to park at the side, the vehicle can be comprehensively analyzed according to the current situation to determine how to park, so that the danger caused by automatic driving is prevented.
In one embodiment, the action of the set user may include a set driving behavior, the driving behavior of the set user may be extracted and compared with relevant data in a driving behavior evaluation database, the current behavior of the user may be compared with a normal driving behavior of the user, and whether a civilization index, a technical index, an love car index, a stability index, and a power saving index in the current driving behavior of the user are seriously deviated from an average value of the set user may be specifically analyzed, and if the average value is seriously deviated, it may be stated that an emotional state of the current driving of the user may have a problem.
Specifically, the driving behavior that needs to be analyzed may include: the driving behaviors such as trampling force of a switch, an accelerator pedal and a brake, whether a behavior of hurrying a steering wheel exists, pressing a horn randomly, opening a high beam randomly, turning without lighting, changing lanes frequently during driving, forcibly merging lanes during driving, robbing lanes, running a red light, driving without using a safety belt, and the like; and comprehensively analyzing and evaluating the emotional state of the set user by comparing the current driving behaviors of the user with the daily driving behaviors of the user.
Generally speaking, if a person is in an angry state relatively, the person may slam the horn, the action is rough or abnormal relatively, and if the vehicle intelligent system monitors that the user has the behavior of slamming the horn and pressing the horn randomly, the user is judged to be in an excited, angry or bad emotion state; at this time, the intelligent system needs to adjust the voice interaction mode of the vehicle and interact with the user to placate the driver.
In one embodiment, if the user has the driving behaviors of hurrying the steering wheel, pressing a horn at random, opening a high beam at random, turning without lighting, changing lanes frequently during driving, forcibly merging lanes during driving, robbing lanes, running a red light and driving without using a safety belt, the vehicle can also automatically prompt the driver through voice so as to correct the bad habits of the driver and educate the user to form good driving habits.
By way of example, if a person drives normally, the steering wheel is relatively gentle and normal, and the action of stepping on an electric door, an accelerator or a brake is relatively moderate, but the behavior of suddenly stepping on the steering wheel, the accelerator or the brake suddenly exists within a set time during driving, which may indicate that the user is currently in an abnormal state, the current emotional state of the user may be determined according to the current action of the user, and the vehicle may be controlled to change a voice interaction mode or play music according to the set emotional state to sooth the user.
According to the man-machine interaction voice control method based on the emotion state of the user, provided by the embodiment of the invention, the current emotion of the user can be calculated according to the driving behavior (such as different from the daily driving behavior) of the user, the speech speed and the intonation of speaking, even the facial expression and the like, such as the condition of urgency comparison, excitement comparison, anger comparison, sadness comparison and the like, the intelligent system can play appropriate music or adjust the change of the navigation voice and the like according to the current emotion state of the user to carry out man-machine interaction with the user, the emotion of the user is adjusted, and the driving of the user is safer.
Fig. 2 is a block diagram illustrating a configuration of a human-computer interaction voice control device based on an emotional state of a user according to an embodiment of the present invention, and referring to fig. 2, the device 200 includes: a monitoring module 201, configured to monitor and set expressions, voices or actions of a user; the emotion state analysis module 202 is connected with the monitoring module 201 and is used for determining the current emotion state of the set user according to the expression, voice or action of the set user; the processing module 203 is connected with the emotional state analysis module 202 and is used for determining a voice control mode of the vehicle according to the current emotional state of the set user; and the execution module 204 is connected with the processing module 203 and is used for performing vehicle human-computer interaction according to the determined voice control mode.
In one embodiment, the apparatus further comprises: the emotion state database module 205 is used for counting emotion state data of a plurality of users to form a general user emotion state database, wherein the emotion state database comprises the relationship between expressions, voices or actions of the users and emotion states; analyzing emotional states of a plurality of users in a big data mode according to a general user emotional state database, and calibrating a determination factor of the emotional states of the users according to actions of the users, wherein the emotional states of the users comprise a cheerful state, an angry state, a sad state, a painful state and an excited state;
in one embodiment, the emotional state analysis module is further configured to evaluate and determine the emotional state of the set user according to the general user emotional state database and the monitored expression, voice or motion of the set user.
In one embodiment, the emotional state analysis module is further to: acquiring the change of the speed, tone and sound of the voice of the set user; and determining the current emotional state of the set user according to the average value of the speed, tone and sound of the voice of the set user.
In one embodiment, the emotional state analysis module is further to: analyzing language information contained in the voice in the set time period of the set user, and judging whether specific words and sentences exist in the language expression of the set user; and determining the emotional state of the set user according to the specific words and phrases appearing in the current expression of the set user.
In one embodiment, the emotional state analysis module is further to: analyzing the action of a set user in a set time period; and judging and setting the current emotional state of the user according to the daily action and the action of the current time of the set user.
In one embodiment, the emotional state analysis module is further to: analyzing the driving behavior of a set user in a set time period; and judging and setting the current emotional state of the user according to the daily driving behavior of the set user and the driving behavior at the current time.
In one embodiment, the emotional state analysis module is further to: analyzing whether the speed, tone and sound of the voice of a set user in a set time period exceed a set threshold value or not; if the speed of speech of the set user is increased to exceed the set speed of speech threshold, and/or the intonation is increased to exceed the set intonation threshold, and/or the voice is increased to exceed the set voice threshold, the set user is in a relative emotional state.
In one embodiment, the processing module is further configured to select a gentle, soothing sound database for voice control to soothe the mood of the set user if the set user is currently at an angry state.
In one embodiment, the execution module is further used for selecting the set human-computer interaction sound effect according to the emotional state of the set user to interact with the set user.
In one embodiment, the execution module is further configured to select the set music to be played according to the emotional state of the set user after determining the emotional state of the set user, so as to sooth the emotion of the set user and ensure driving safety.
In one embodiment, the execution module is further used for selecting the set navigation language to play according to the set emotional state of the user to sooth the emotion of the set user to ensure driving safety if the driving navigation state is currently in the driving navigation state.
The invention also provides a vehicle which comprises the human-computer interaction voice control device based on the emotional state of the user.
In one embodiment, the vehicle can be connected with a server located at the cloud end in real time, and data of the server is called in time to analyze the emotional state of the user.
Fig. 3 shows a block diagram of a vehicle according to an embodiment of the present invention, which may include, as shown in fig. 3: the intelligent driving system comprises a central control module, an instrument panel 310, a driving recorder 311, a HUD (Head Up Display) Head-Up Display 312, an intelligent vehicle-mounted infotainment system 313 and an intelligent driving module 313.
Dashboard 310 has a 12.3 inch LCD display device, which may be implemented with a TI J6 CPU; the operating system of the dashboard may be based on a QNX embedded system, and the dashboard may be used to display vehicle status including speed, rotational speed, electrical quantity, tire pressure, vehicle parking, gear, etc., a map, vehicle navigation information, vehicle music play, etc. The HUD heads-up display 312 may display GPS navigation information, navigation path information, time information, and the like.
In one embodiment, smart driving module 313 may be used to process operations related to smart driving, and smart driving module 313 may include Advanced Driver Assistance Systems (ADAS), active safety Systems, Attention Assistance Systems (AAS), Fatigue Warning Systems (FWS), Vehicle intelligent Acoustic warning Systems (AVAS), and the like. The vehicle can be intelligently driven by combining an ADAS system and the like, and the intelligent driving can be completely unmanned driving or can be advanced auxiliary driving functions such as auxiliary merging and lane deviation for driving control of a driver.
The central control device can be composed of a plurality of modules, and mainly comprises: a main board 301; a Serial Advanced Technology Attachment (SATA) module 302, connected to a storage device such as SSD303, may be used to store data information; an AM (Amplitude Modulation)/FM (Frequency Modulation) module 304, which provides the vehicle with the function of a radio; a power amplifier module 305 for processing sound; a WIFI (Wireless-Fidelity)/Bluetooth module 306 for providing WIFI/Bluetooth services to the vehicle; an LTE (long term Evolution) communication module 307 that provides a communication function with a telecommunications carrier for the vehicle; a power module 308, wherein the power module 308 provides power for the central control device; the Switch switching module 309, the Switch switching module 309 can be used as an expandable interface to connect with various sensors, for example, if a night vision function sensor or a PM2.5 function sensor is required to be added, the Switch switching module 309 can be connected to the main board of the central control device, so that the processor of the central control device can process data and transmit the data to the central control display.
In one embodiment, the vehicle further includes a look-around camera, an ADAS camera, a night vision camera, a millimeter wave radar, an ultrasonic radar, an ESR radar, or the like. After the vehicle hardware is produced, the intelligent driving related hardware is mounted, and in the later stage, the automatic driving related functions can be completed by using the hardware through OTA (over the air) upgrading.
The description of the present invention has been presented for purposes of illustration and description, and is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to practitioners skilled in this art. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.

Claims (11)

1. A man-machine interaction voice control method based on user emotion states is characterized by comprising the following steps:
monitoring and setting the expression, voice or action of a user;
determining the current emotional state of the set user according to the expression, voice or action of the set user;
determining a control mode of the vehicle according to the current emotional state of the set user, wherein the control mode comprises the steps of increasing the warning value of an ADAS system, automatically tightening a safety belt and giving an alarm to the user through the vibration of a steering wheel; the warning value comprises increasing the safe following distance;
determining a voice control mode of the vehicle according to the current emotional state of the set user;
performing vehicle man-machine interaction according to the determined voice control mode;
and obtaining a voice interaction database interacting with the emotion state of the user according to daily human-computer interaction statistics.
2. The method of claim 1, wherein:
the determining the current emotional state of the set user according to the expression, voice or action of the set user comprises:
counting emotional state data of a plurality of users to form a general user emotional state database, wherein the emotional state database comprises the relation between expressions, voice or actions of the users and emotional states;
analyzing emotional states of a plurality of users in a big data mode according to a general user emotional state database, and calibrating a determination factor of the emotional states of the users according to actions of the users, wherein the emotional states of the users comprise a cheerful state, an angry state, a sad state, a painful state and an excited state;
and/or
The determining the current emotional state of the set user according to the expression, voice or action of the set user comprises:
and evaluating and determining the current emotional state of the set user according to the general user emotional state database and the monitored expression, voice or action of the set user.
3. The method of claim 1 or 2, wherein determining the current emotional state of the set user based on the expression, voice, or motion of the set user further comprises:
acquiring the change of the speed, tone and sound of the voice of the set user; determining the current emotional state of the set user according to the average value of the speed, tone and sound of the voice of the set user;
and/or
Analyzing language information contained in sound in a set time period of a set user, and judging whether specific words and phrases exist in language expression of the set user;
determining the emotional state of the set user according to the specific words and phrases appearing in the current expression of the set user;
and/or
Analyzing the action of a set user in a set time period;
judging and setting the current emotional state of the user according to the daily action and the action of the current time of the set user;
and/or
Analyzing the driving behavior of a set user in a set time period;
and judging and setting the current emotional state of the user according to the daily driving behavior of the set user and the driving behavior at the current time.
4. The method of claim 3, further comprising:
analyzing whether the speed, tone and sound of the voice of a set user in a set time period exceed a set threshold value or not;
if the speed of speech of the set user is increased to exceed the set speed of speech threshold, and/or the intonation is increased to exceed the set intonation threshold, and/or the voice is increased to exceed the set voice threshold, the set user is in a relative emotional state.
5. The method of claim 1, wherein determining the voice control mode of the vehicle according to the current emotional state of the set user comprises:
if the user is set to be in an angry state at present;
selecting a soft and consolation sound database to perform voice control, and pacifying and setting the emotion of the user;
and/or
Selecting a set human-computer interaction sound effect according to the emotion state of a set user to interact with the set user;
and/or
After determining to set the emotional state of the user;
the set music is selected to be played according to the emotion state of the set user so as to sooth the emotion of the set user and ensure the driving safety;
and/or
If the vehicle is currently in a driving navigation state;
and selecting the set navigation language to play according to the emotional state of the set user so as to soothe the emotion of the set user and ensure the driving safety.
6. A human-computer interaction voice control device based on user emotion states is characterized by comprising:
the monitoring module is used for monitoring and setting the expression, voice or action of the user;
the emotion state analysis module is connected with the monitoring module and used for determining the current emotion state of the set user according to the expression, voice or action of the set user;
the processing module is connected with the emotional state analysis module and used for determining a voice control mode of the vehicle according to the current emotional state of the set user; the processing module is also used for determining a control mode of the vehicle according to the current emotion state of the set user, wherein the control mode comprises the steps of increasing the warning value of an ADAS system, automatically tightening a safety belt and giving an alarm to the user through the vibration of a steering wheel; the warning value comprises increasing the safe following distance;
the execution module is connected with the processing module and is used for carrying out vehicle man-machine interaction according to the determined voice control mode; and obtaining a voice interaction database interacting with the emotion state of the user according to daily human-computer interaction statistics.
7. The apparatus of claim 6, wherein:
the system also comprises an emotional state database module which is used for counting emotional state data of a plurality of users to form a general user emotional state database, wherein the emotional state database comprises the relation between the expression, voice or action of the users and the emotional state; analyzing emotional states of a plurality of users in a big data mode according to a general user emotional state database, and calibrating a determination factor of the emotional states of the users according to actions of the users, wherein the emotional states of the users comprise a cheerful state, an angry state, a sad state, a painful state and an excited state;
and/or
And the emotion state analysis module is also used for evaluating and determining the emotion state of the set user according to the general user emotion state database and the monitored expression, voice or action of the set user.
8. The apparatus of claim 6 or 7, wherein the emotional state analysis module is further configured to:
acquiring the change of the speed, tone and sound of the voice of the set user; determining the current emotional state of the set user according to the average value of the speed, tone and sound of the voice of the set user;
and/or
Analyzing language information contained in the voice in the set time period of the set user, and judging whether specific words and sentences exist in the language expression of the set user; determining the emotional state of the set user according to the specific words and phrases appearing in the current expression of the set user;
and/or
Analyzing the action of a set user in a set time period;
judging and setting the current emotional state of the user according to the daily action and the action of the current time of the set user;
and/or
Analyzing the driving behavior of a set user in a set time period; and judging and setting the current emotional state of the user according to the daily driving behavior of the set user and the driving behavior at the current time.
9. The apparatus of claim 8, comprising:
analyzing whether the speed, tone and sound of the voice of a set user in a set time period exceed a set threshold value or not;
if the speed of speech of the set user is increased to exceed the set speed of speech threshold, and/or the intonation is increased to exceed the set intonation threshold, and/or the voice is increased to exceed the set voice threshold, the set user is in a relative emotional state.
10. The apparatus of claim 6, comprising:
the processing module is also used for selecting a soft and comforting sound database to perform voice control if the user is set to be in an angry state at present, and soothing the emotion of the set user;
and/or
The execution module is also used for selecting the set human-computer interaction sound effect according to the emotion state of the set user to interact with the set user;
and/or
The execution module is also used for selecting the set music to play according to the emotion state of the set user after the emotion state of the set user is determined so as to sooth the emotion of the set user and ensure the driving safety;
and/or
The execution module is also used for judging whether the vehicle is in a driving navigation state or not;
and selecting the set navigation language to play according to the emotional state of the set user so as to soothe the emotion of the set user and ensure the driving safety.
11. A vehicle comprising a human-computer interaction voice control device based on the emotional state of a user according to any of claims 6 to 10.
CN201611229157.7A 2016-12-27 2016-12-27 Man-machine interaction voice control method and device based on user emotion state and vehicle Active CN106803423B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611229157.7A CN106803423B (en) 2016-12-27 2016-12-27 Man-machine interaction voice control method and device based on user emotion state and vehicle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611229157.7A CN106803423B (en) 2016-12-27 2016-12-27 Man-machine interaction voice control method and device based on user emotion state and vehicle

Publications (2)

Publication Number Publication Date
CN106803423A CN106803423A (en) 2017-06-06
CN106803423B true CN106803423B (en) 2020-09-04

Family

ID=58985118

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611229157.7A Active CN106803423B (en) 2016-12-27 2016-12-27 Man-machine interaction voice control method and device based on user emotion state and vehicle

Country Status (1)

Country Link
CN (1) CN106803423B (en)

Families Citing this family (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107230384B (en) * 2017-06-21 2020-09-25 深圳市盛路物联通讯技术有限公司 Parking guidance system and method based on expected parking duration and weather information
US10381005B2 (en) 2017-11-28 2019-08-13 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods for determining user frustration when using voice control
CN108010512B (en) * 2017-12-05 2021-04-30 广东小天才科技有限公司 Sound effect acquisition method and recording terminal
CN108664123A (en) * 2017-12-15 2018-10-16 蔚来汽车有限公司 People's car mutual method, apparatus, vehicle intelligent controller and system
CN108052016A (en) * 2017-12-29 2018-05-18 南京工程学院 A kind of interactive intelligent mirror
CN110164427A (en) * 2018-02-13 2019-08-23 阿里巴巴集团控股有限公司 Voice interactive method, device, equipment and storage medium
CN108682419A (en) * 2018-03-30 2018-10-19 京东方科技集团股份有限公司 Sound control method and equipment, computer readable storage medium and equipment
CN108710821A (en) * 2018-03-30 2018-10-26 斑马网络技术有限公司 Vehicle user state recognition system and its recognition methods
CN108549720A (en) 2018-04-24 2018-09-18 京东方科技集团股份有限公司 It is a kind of that method, apparatus and equipment, storage medium are pacified based on Emotion identification
CN108896061A (en) * 2018-05-11 2018-11-27 京东方科技集团股份有限公司 A kind of man-machine interaction method and onboard navigation system based on onboard navigation system
CN109190459A (en) * 2018-07-20 2019-01-11 上海博泰悦臻电子设备制造有限公司 A kind of car owner's Emotion identification and adjusting method, storage medium and onboard system
CN108984229B (en) * 2018-07-24 2021-11-26 广东小天才科技有限公司 Application program starting control method and family education equipment
CN110825216A (en) * 2018-08-10 2020-02-21 北京魔门塔科技有限公司 Method and system for man-machine interaction of driver during driving
CN109243438B (en) * 2018-08-24 2023-09-26 上海擎感智能科技有限公司 Method, system and storage medium for regulating emotion of vehicle owner
CN109532653B (en) 2018-10-11 2020-01-21 百度在线网络技术(北京)有限公司 Method, device, computer equipment and storage medium for communicating with rear vehicle
CN109346079A (en) * 2018-12-04 2019-02-15 北京羽扇智信息科技有限公司 Voice interactive method and device based on Application on Voiceprint Recognition
CN109616109B (en) * 2018-12-04 2020-05-19 北京蓦然认知科技有限公司 Voice awakening method, device and system
CN109599094A (en) * 2018-12-17 2019-04-09 海南大学 The method of sound beauty and emotion modification
CN109669661A (en) * 2018-12-20 2019-04-23 广东小天才科技有限公司 Control method of dictation progress and electronic equipment
US11170240B2 (en) * 2019-01-04 2021-11-09 Cerence Operating Company Interaction system and method
CN109712646A (en) * 2019-02-20 2019-05-03 百度在线网络技术(北京)有限公司 Voice broadcast method, device and terminal
CN110085225B (en) * 2019-04-24 2024-01-02 北京百度网讯科技有限公司 Voice interaction method and device, intelligent robot and computer readable storage medium
CN111976732A (en) * 2019-05-23 2020-11-24 上海博泰悦臻网络技术服务有限公司 Vehicle control method and system based on vehicle owner emotion and vehicle-mounted terminal
CN112009395A (en) * 2019-05-28 2020-12-01 北京车和家信息技术有限公司 Interaction control method, vehicle-mounted terminal and vehicle
CN110334669B (en) * 2019-07-10 2021-06-08 深圳市华腾物联科技有限公司 Morphological feature recognition method and equipment
CN110215683A (en) * 2019-07-11 2019-09-10 龙马智芯(珠海横琴)科技有限公司 A kind of electronic game system of role playing game
CN110534091A (en) * 2019-08-16 2019-12-03 广州威尔森信息科技有限公司 A kind of people-car interaction method identified based on microserver and intelligent sound
CN110641476A (en) * 2019-08-16 2020-01-03 广汽蔚来新能源汽车科技有限公司 Interaction method and device based on vehicle-mounted robot, controller and storage medium
CN112562661A (en) * 2019-09-25 2021-03-26 上海汽车集团股份有限公司 Vehicle-mounted man-machine interaction system and motor vehicle
CN110534135A (en) * 2019-10-18 2019-12-03 四川大学华西医院 A method of emotional characteristics are assessed with heart rate response based on language guidance
CN110689906A (en) * 2019-11-05 2020-01-14 江苏网进科技股份有限公司 Law enforcement detection method and system based on voice processing technology
JP7264079B2 (en) * 2020-02-05 2023-04-25 トヨタ自動車株式会社 Emotion estimation device, method, program and vehicle
CN111329498A (en) * 2020-03-09 2020-06-26 郑州大学 Multi-modal driver emotion auxiliary adjusting method
CN111402925B (en) * 2020-03-12 2023-10-10 阿波罗智联(北京)科技有限公司 Voice adjustment method, device, electronic equipment, vehicle-mounted system and readable medium
CN111666444B (en) * 2020-06-02 2021-04-27 中国科学院计算技术研究所 Audio push method and system based on artificial intelligence, and related method and equipment
CN111605556B (en) * 2020-06-05 2022-06-07 吉林大学 Road rage prevention recognition and control system
CN111724880A (en) * 2020-06-09 2020-09-29 百度在线网络技术(北京)有限公司 User emotion adjusting method, device, equipment and readable storage medium
CN112035034B (en) * 2020-08-27 2024-05-03 芜湖盟博科技有限公司 Vehicle-mounted robot interaction method
CN112185422B (en) * 2020-09-14 2022-11-08 五邑大学 Prompt message generation method and voice robot thereof
CN112446306A (en) * 2020-11-13 2021-03-05 阳光保险集团股份有限公司 Emotion intensity identification method and device, electronic equipment and storage medium
CN113012717A (en) * 2021-02-22 2021-06-22 上海埃阿智能科技有限公司 Emotional feedback information recommendation system and method based on voice recognition
CN113658580B (en) * 2021-06-24 2024-06-18 大众问问(北京)信息科技有限公司 Voice prompt method, device, computer equipment and storage medium
CN113780062A (en) * 2021-07-26 2021-12-10 岚图汽车科技有限公司 Vehicle-mounted intelligent interaction method based on emotion recognition, storage medium and chip
CN116061821A (en) * 2021-11-01 2023-05-05 华人运通(江苏)技术有限公司 Control method, device and system of vehicle, mechanical arm and vehicle
CN114049677B (en) * 2021-12-06 2023-08-25 中南大学 Vehicle ADAS control method and system based on driver emotion index
DE102022133244A1 (en) 2022-12-14 2024-06-20 Bayerische Motoren Werke Aktiengesellschaft Navigation system of a vehicle and method for operating a navigation system of a vehicle

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005239117A (en) * 2004-01-26 2005-09-08 Nissan Motor Co Ltd Driver feeling guiding device
CN102874259A (en) * 2012-06-15 2013-01-16 浙江吉利汽车研究院有限公司杭州分公司 Automobile driver emotion monitoring and automobile control system
CN105700682A (en) * 2016-01-08 2016-06-22 北京乐驾科技有限公司 Intelligent gender and emotion recognition detection system and method based on vision and voice
CN106114516A (en) * 2016-08-31 2016-11-16 合肥工业大学 The angry driver behavior modeling of a kind of drive automatically people's characteristic and tampering devic
CN206049658U (en) * 2016-08-31 2017-03-29 合肥工业大学 Angry driver behavior modeling and tampering devic based on drive automatically people's characteristic

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5088669B2 (en) * 2007-03-23 2012-12-05 株式会社デンソー Vehicle periphery monitoring device
CN101669090A (en) * 2007-04-26 2010-03-10 福特全球技术公司 Emotive advisory system and method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005239117A (en) * 2004-01-26 2005-09-08 Nissan Motor Co Ltd Driver feeling guiding device
CN102874259A (en) * 2012-06-15 2013-01-16 浙江吉利汽车研究院有限公司杭州分公司 Automobile driver emotion monitoring and automobile control system
CN105700682A (en) * 2016-01-08 2016-06-22 北京乐驾科技有限公司 Intelligent gender and emotion recognition detection system and method based on vision and voice
CN106114516A (en) * 2016-08-31 2016-11-16 合肥工业大学 The angry driver behavior modeling of a kind of drive automatically people's characteristic and tampering devic
CN206049658U (en) * 2016-08-31 2017-03-29 合肥工业大学 Angry driver behavior modeling and tampering devic based on drive automatically people's characteristic

Also Published As

Publication number Publication date
CN106803423A (en) 2017-06-06

Similar Documents

Publication Publication Date Title
CN106803423B (en) Man-machine interaction voice control method and device based on user emotion state and vehicle
US12032730B2 (en) Methods and systems for using artificial intelligence to evaluate, correct, and monitor user attentiveness
EP3675121B1 (en) Computer-implemented interaction with a user
EP3384475B1 (en) System and method of conversational adjustment based on user's cognitive state
US10192171B2 (en) Method and system using machine learning to determine an automotive driver's emotional state
CN112735440A (en) Vehicle-mounted intelligent robot interaction method, robot and vehicle
EP3647147B1 (en) Method and apparatus for eveluating vehicle, device and computer readable storage medium
CN111402925A (en) Voice adjusting method and device, electronic equipment, vehicle-mounted system and readable medium
KR20100062145A (en) System and method for controlling sensibility of driver
Kashevnik et al. Multimodal corpus design for audio-visual speech recognition in vehicle cabin
CN112215097A (en) Method for monitoring driving state of vehicle, vehicle and computer readable storage medium
EP4042322A1 (en) Methods and systems for using artificial intelligence to evaluate, correct, and monitor user attentiveness
CN113448433A (en) Emotion responsive virtual personal assistant
CN113771859A (en) Intelligent driving intervention method, device and equipment and computer readable storage medium
CN113879235A (en) Method, system, equipment and storage medium for multi-screen control of automobile
Sanghavi et al. Exploring the influence of driver affective state and auditory display urgency on takeover performance in semi-automated vehicles: Experiment and modelling
CN113879315A (en) Driving support device and driving support method
CN113538852A (en) Fatigue driving reminding method and vehicle-mounted terminal
CN116834749A (en) Fatigue driving early warning method, device, equipment and storage medium
CN113450788B (en) Sound output control method and sound output control device
JP6819529B2 (en) Information processing equipment, information processing system, and information processing method
CN115830724A (en) Vehicle-mounted recognition interaction method and system based on multi-mode recognition
CN114037981A (en) Fatigue driving warning method, fatigue driving warning device, storage medium and electronic equipment
KR102132058B1 (en) Interactive voice communication system embedded in a car
CN105575392A (en) System and method for user interaction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant