CN106933345A - For the multi-modal exchange method and device of intelligent robot - Google Patents

For the multi-modal exchange method and device of intelligent robot Download PDF

Info

Publication number
CN106933345A
CN106933345A CN201710034038.4A CN201710034038A CN106933345A CN 106933345 A CN106933345 A CN 106933345A CN 201710034038 A CN201710034038 A CN 201710034038A CN 106933345 A CN106933345 A CN 106933345A
Authority
CN
China
Prior art keywords
modal
information
user
brain wave
topic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710034038.4A
Other languages
Chinese (zh)
Other versions
CN106933345B (en
Inventor
魏彬彬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Guangnian Wuxian Technology Co Ltd
Original Assignee
Beijing Guangnian Wuxian Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Guangnian Wuxian Technology Co Ltd filed Critical Beijing Guangnian Wuxian Technology Co Ltd
Priority to CN201710034038.4A priority Critical patent/CN106933345B/en
Publication of CN106933345A publication Critical patent/CN106933345A/en
Application granted granted Critical
Publication of CN106933345B publication Critical patent/CN106933345B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/015Input arrangements based on nervous system activity detection, e.g. brain waves [EEG] detection, electromyograms [EMG] detection, electrodermal response detection

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Neurosurgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Neurology (AREA)
  • Health & Medical Sciences (AREA)
  • Dermatology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Biomedical Technology (AREA)
  • Manipulator (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The present invention provides a kind of multi-modal exchange method and device for intelligent robot.Wherein, the method is comprised the following steps:Obtain the multi-modal input information of user;Parse involved object or topic in multi-modal input information;The brain wave information that the user produces to the object or the topic is obtained, multi-modal man-machine interaction strategy is determined, corresponding multi-modal output data is exported, the brain wave information carries user emotion information.The present invention is interacted by the man-machine interaction strategy determined based on brain wave with reference to the strategy, can increase interactive topic, improves user's unhealthy emotion, improves the intelligent and class human nature of intelligent robot, enhancing user's viscosity.

Description

For the multi-modal exchange method and device of intelligent robot
Technical field
The present invention relates to field in intelligent robotics, specifically, it is related to a kind of multi-modal exchange method of intelligent robot And device.
Background technology
With continuing to develop for science and technology, the introducing of information technology, computer technology and artificial intelligence technology, machine Industrial circle is progressively walked out in the research of people, gradually extend to the neck such as medical treatment, health care, family, amusement and service industry Domain.And people for the requirement of robot also conform to the principle of simplicity the multiple mechanical action of substance be promoted to anthropomorphic question and answer, independence and with The intelligent robot that other robot is interacted, man-machine interaction also just turns into the key factor for determining intelligent robot development, Therefore, the interactive capability of intelligent robot is improved, intelligent, the weight as current urgent need to resolve of intelligent robot is lifted Want problem.
The content of the invention
Above mentioned problem it is an object of the invention to solve prior art, there is provided a kind of for the multi-modal of intelligent robot Exchange method.The described method comprises the following steps:
Obtain the multi-modal input information of user;
Parse involved object information or topic information in multi-modal input information;
The brain wave information that the user produces to the object or the topic is obtained, multi-modal man-machine interaction strategy is determined, it is defeated Go out corresponding multi-modal output data, the brain wave information carries user emotion information.
Multi-modal exchange method for intelligent robot of the invention, it is preferred that also including by following step It is rapid to preset multi-modal man-machine interaction strategy,
The advance visual information for gathering user in training environment, auditory information and brain wave information;
The brain wave information is parsed, the mood of user is determined;
Object or topic to being related in visual information and auditory information carry out the mark based on mood, with generate on Object or the policy information of the multi-modal output of topic.
Multi-modal exchange method for intelligent robot of the invention, it is preferred that also include:
After multi-modal output data is exported, brain wave letter of the collection user to the multi-modal output data feedback Breath, to obtain the feedback mood of the user;
The mood of user feedback is redefined into multi-modal man-machine interaction strategy, the multi-modal output number that output matches According to.
Multi-modal exchange method for intelligent robot of the invention, it is preferred that also include:
Obtain user's current emotional states and the default multi-modal man-machine interaction strategy is carried out according to emotional state Adjustment;
The emotional state is obtained by gathering brain wave information mode.
Multi-modal exchange method for intelligent robot of the invention, it is preferred that described brain wave information Gathered using EEG equipment.
According to another aspect of the present invention, a kind of multi-modal interactive device for intelligent robot is additionally provided, it is described Device is included with lower module:
Multi-modal information input module, its multi-modal input information for being used to obtain user;
Multi-modal information parsing module, it is used to parse involved object information or topic letter in multi-modal input information Breath;
Multi-modal information output module, it is used to obtain the brain wave information that the user produces the object or the topic, Determine multi-modal man-machine interaction strategy, export corresponding multi-modal output data, the brain wave information carries user emotion letter Breath.
Multi-modal interactive device for intelligent robot of the invention, it is preferred that also including the default mould of strategy Block, it presets multi-modal man-machine interaction strategy, and the module is further included such as lower unit:
Information acquisition unit, it is used to gather in advance visual information of the user in training environment, auditory information and brain electricity Ripple information;
Mood determining unit, it is used to parse the brain wave information, determines the mood of user;
Policy information generation unit, its object or topic for being used to being related in visual information and auditory information carry out base In the mark of mood, to generate the policy information on object or the multi-modal output of topic.
Multi-modal interactive device for intelligent robot of the invention, it is preferred that the multi-modal information is defeated Go out module, it further after multi-modal output data is exported, gathers brain of the user to the multi-modal output data feedback Wave information, to obtain the feedback mood of the user;The mood of user feedback is redefined into multi-modal man-machine interaction strategy, it is defeated Go out the multi-modal output data for matching.
Multi-modal interactive device for intelligent robot of the invention, it is preferred that also include:
Policy information adjusting module, it is used to obtain user's current emotional states and according to emotional state to described default Multi-modal man-machine interaction strategy is adjusted, and the emotional state is obtained by gathering brain wave information mode.
Multi-modal interactive device for intelligent robot of the invention, it is preferred that described information collecting unit Including EEG equipment, its described brain wave information of collection.
The present invention is brought to be beneficial in that, method according to embodiments of the present invention, and robot is obtaining user's After multi-modal input information, involved object or topic in multi-modal input information are parsed, then, obtain the user to the object Or the brain wave information that the topic is produced, determine multi-modal man-machine interaction strategy, corresponding multi-modal output data is exported, it is described Brain wave information carries user emotion information.Man-machine interaction strategy by being determined based on brain wave of the invention, with reference to the strategy Interact, interactive topic can be increased, improve user's unhealthy emotion, improve the intelligent and class human nature of intelligent robot, increase Strong user's viscosity.
Other features and advantages of the present invention will be illustrated in the following description, also, the partly change from specification Obtain it is clear that or being understood by implementing the present invention.The purpose of the present invention and other advantages can be by specification, rights Specifically noted structure is realized and obtained in claim and accompanying drawing.
Brief description of the drawings
Accompanying drawing is used for providing a further understanding of the present invention, and constitutes a part for specification, with reality of the invention Apply example to be provided commonly for explaining the present invention, be not construed as limiting the invention.In the accompanying drawings:
Fig. 1 is that the flow for being related to the multi-modal exchange method for intelligent robot of first embodiment of the invention is illustrated Figure.
Fig. 2 is the idiographic flow schematic diagram for being related to the default multi-modal man-machine interaction strategy of first embodiment of the invention.
Fig. 3 is that the flow for being related to the multi-modal exchange method for intelligent robot of second embodiment of the invention is illustrated Figure.
Fig. 4 is that the flow for being related to the multi-modal exchange method for intelligent robot of third embodiment of the invention is illustrated Figure.
Fig. 5 is the structural frames for being related to the multi-modal interactive device 400 for intelligent robot of fourth embodiment of the invention Figure.
Specific embodiment
Describe embodiments of the present invention in detail below with reference to drawings and Examples, how the present invention is applied whereby Technological means solves technical problem, and reaches the implementation process of relevant art effect and can fully understand and implement according to this.This Shen Each feature that please be in embodiment and embodiment, can be combined with each other under the premise of not colliding, the technical scheme for being formed Within protection scope of the present invention.
In addition, the flow of accompanying drawing can be in the such as one group computer system of computer executable instructions the step of illustrating Middle execution.And, although show logical order in flow charts, but in some cases, can be with different from herein Order performs shown or described step.
In the prior art, when user talks with robot, language, facial expression that robot can be sent by user With action etc. it is multi-modal enter data to determine active user heartbeat conditions, then according to heartbeat conditions come to user output compared with It is suitable multi-modal information.But, due to the multi-modal informations such as the language of user, facial expression and action have it is uncertain Property, such as user can cover up the true emotional of oneself by way of pretending facial expression, therefore, robot is to these When data are analyzed, parsing precision is relatively low, and the reliability of Emotion identification is low, so that the multi-modal data of correspondence output is not yet Match somebody with somebody.Therefore, the intelligent and personification of existing robot is poor, it is impossible to analyze the real heartbeat conditions of user, drop exactly Low user experience.Therefore, being accomplished by one kind can exactly analyze the real heartbeat conditions of user, increase interaction topic, Improve user's unhealthy emotion, improve the intelligent and class human nature of intelligent robot, the solution of enhancing user's viscosity.
The multi-modal exchange method for intelligent robot of the embodiment of the present invention can make robot with user mutual During, the real heartbeat conditions of user are analyzed exactly, increase interaction topic, improve intelligent robot and class human nature.Tool For body, intelligent robot obtains multi-modal input information, parses object or topic, Yi Jiyong involved by multi-modal input information To the object or the brain wave information of the topic, being then based on default multi-modal man-machine interaction strategy (can referred to as " man-machine friendship at family Mutual strategy ") the corresponding multi-modal output data of output, i.e. according to the user plane to the object or the topic when brain wave that produces The emotional information that information is shown finds corresponding multi-modal output data and is exported from man-machine interaction strategy.It is default Multi-modal man-machine interaction strategy can be directed to the Personalized Policies of the individual customization of a certain user, the user plane pair can be based on Brain wave information that different objects or theme are produced generates policy information, for example, in robot according to the corresponding brain of object When wave information obtains user's mood now for negative feeling, by combining multi-modal man-machine interaction strategy, exported to user One can make user's easily multi-modal data, song that such as user is liked etc., can so improve the bad of user Mood, enhancing user uses the viscosity of robot.
In addition, the preset procedures on multi-modal man-machine interaction strategy, robot collect in advance user plane to different things, Brain wave delta data during topic scene, and the thing observed and hear according to brain wave change, corresponding emotional state pair Thing, content, scene are classified, and user is liked according to fancy grade, disagreeable degree etc./nuisance thing, topic etc. set Determine priority, then deep learning is carried out based on above- mentioned information, generate man-machine interaction strategy.Because the man-machine interaction strategy is to be based on Brain wave information carries out mood parsing, therefore the degree of accuracy is higher, and accurate multimode can be provided when being easy to be used below State data are to user.
It should be noted that because default multi-modal interactive strategy is directed to hobby of the user within certain time Set policy information, with the change of user's subjective factor, the policy information in default policy library can exist certain inclined Difference, it is therefore desirable to which adjustment carrys out optimisation strategy information.The embodiment of the present invention obtains user to multi-modal defeated during interacting Go out the emotional state of data response, man-machine interaction strategy is adjusted according to emotional state, so as in next time interactive, can be defeated Go out and give user more accurate multi-modal information.
In addition, after multi-modal output data is exported to user, it is possible that user is to multi-modal output data institute Emotional information and unmatched phenomenon in the mood and man-machine interaction strategy of feedback, therefore, it is also desirable to further gather user For the brain wave information that a certain multi-modal output data is fed back, and then feedback mood is obtained, then again by feedback data weight It is new to combine man-machine interaction strategy, the content more matched to user is exported again, and it is possible to reference to feedback data, to man-machine Interactive strategy is adjusted, so that man-machine interaction strategy is closer to user's request.So can further adjust the bad of user Mood.Certainly, it is active mood in the mood for monitoring the brain wave message reflection of user feedback in order to increase topic content When, it is also possible to recombine man-machine interaction strategy, output and more abundant, detailed topic content.
When brain wave information is gathered, because EEG equipment costs are relatively low, information gathering is simple and easy to do and ensure that collection During user comfortableness, it is preferred to use EEG equipment.
First embodiment
Fig. 1 is the schematic flow sheet of the example one for being related to the multi-modal exchange method for intelligent robot of the invention, The method of the embodiment is mainly included the following steps that.
In step s 110, robot obtains the multi-modal input information of user.
In this step, the multi-modal input information of acquisition mainly include coming from the text data of user, voice data, Video data, view data and/or the program for allowing the robot to export certain action or execution software or hardware The information of instruction.In specific example, during robot and user mutual, robot is called and starts text input and sets Standby (such as touch-screen), sound collection equipment (such as microphone) and image capture device (such as camera), are set by these It is standby come persistently monitor and catch text message, voice messaging and user's face expression information and/or limb action information.
In the step s 120, involved object information or topic information in multi-modal input information are parsed.
Then, the multi-modal input information for receiving is parsed, obtains the thing included in these multi-modal datas Body information or topic information, specifically, object information is mainly the title of object, and subject information is mainly subject name and general Summary/particular content, such as film, music, dancing etc.
In this example, robot is mainly and parses the relevant thing of extraction according to keyword from text message or acoustic information The object information or subject information of body.By taking acoustic information as an example, robot obtain multi-modal data after, by the multi-modal data It is submitted to ASR and VPR (vocal prints that the ASR or local and cloud server of local ASR or cloud server mix Identification, Voiceprint Recognition) engine.These engines convert voice data to text message using ASR technology. The specific pretreatment that such as denoising etc is first carried out to multi-modal input data, then carries out language by pretreated voice messaging The comprehensive analysis of sound identification, generates text message corresponding with voice messaging.Furthermore, it is understood that according to voice in identification process The model of identification, the sound template that will be prestored is compared with the feature of the voice signal of input, according to certain search And matching strategy, find out a series of optimal templates with input voice match.Then according to the definition of this template, by tabling look-up Recognition result can be just given.Then, the keyword of noun part-of-speech is extracted from recognition result, then confirm object information or Subject information.If not parsing object information or subject information from text message or voice messaging, robot is further View data is parsed, the object information in image is identified by object recognition technique.
In step s 130, the brain wave information that the user produces to the object or the topic is obtained, multi-modal people is determined Machine interactive strategy, exports corresponding multi-modal output data, wherein, brain wave information carries user emotion information.
Specifically, the brain wave information that user produces to the object or topic is obtained, according to brain wave information correspondence Emotional information search corresponding multi-modal output data, be then output to user.
In default multi-modal man-machine interaction strategy is set up, be may include steps of for overview:Design induces each The stimulus of mood (active mood and negative feeling) is planted, is then obtained with the EEG signal for inducing mood using EEG equipment, EEG signals are obtained into temporal signatures, frequency domain character or time-frequency characteristics by mathematic(al) manipulation, is ground according to related physiology, psychology Study carefully achievement and extract EEG signal feature, by Data Dimensionality Reduction and feature selecting, so as to obtain being adapted to the EEG feature sets of mood classification, Feature based collection carries out mood classification.Fig. 2 is the tool for being related to the default multi-modal man-machine interaction strategy of first embodiment of the invention Body schematic flow sheet.Illustrate how to preset multi-modal man-machine interaction strategy below with reference to Fig. 2.
In sub-step S1010, the advance visual information for gathering user in training environment, auditory information and brain wave letter Breath.
During emotional training is carried out to a certain user, the stimulus material for inducing user's difference mood is first designed. The stimulus material of conventional induction mood has visual stimulus material, acoustic stimuli material and multimedia stimulus material etc. at present.Depending on Feel that stimulus material mainly has words and phrases sentence and picture etc., acoustic stimuli material mainly has the non-karst areas sound with different emotions Section, word, words and phrases or phrase etc., multimedia stimulus material often combine the dynamic materials such as vision, the sense of hearing.Mood induces Refer to directly to manipulate emotional state, then record and observe positive or negative feeling.Because viewing picture or film can be quantified Ground induces tested related emotional, and it turns into mood and induces a kind of the most frequently used method.Using picture as visual stimulus source, in order to Ensure that stimulus material can preferably induce specific mood, when selected picture is showed to user, every width picture is in The existing time may remain in five seconds.In addition, having various objects or subject information in these stimulus materials, mood is being carried out Before training, these information can be in advance obtained.
During being trained, user can produce different eeg signals, brain electric information collection according to stimulus material Equipment records the brain electric information of subject according to the sampling period of setting.At the same time, in order to preferably accurately identify user State, user can also select corresponding emotional state according to the impression of oneself, and the emotional state is synchronous with brain electric information Record.
In sub-step S1020, brain wave information is parsed, determine the mood of user.
Because EEG signals are very faint, therefore the interference of other noise signals is highly susceptible in gatherer process, needed The EEG signals for collecting are pre-processed to remove the artefact of doping.Spy then is carried out to the EEG signals for removing artefact Extraction is levied, these features can be temporal signatures, frequency domain character or time-frequency characteristics.According to these features and before according to training Sample obtained by different moods (such as tranquil, glad, sad, frightened) corresponding brain electrical feature determine the feelings of user Thread.In order to ensure the accuracy of mood classification, it is also possible to reference to the emotional state that user selects in training.In addition, except upper Outside the common attribute of three kinds of face, many other features can also be extracted from EEG signals, such as entropy, fractal dimension and self-defined Feature etc..
It should be noted that, although coming etc. to realize using in this example using single physiological signal, i.e. brain wave mode Trained for the Emotion identification of different objects/topic at family.It is also possible to be based on facial expression, voice, posture, text and brain telecommunications Number etc. the mode of Multi-source Information Fusion realize that this example is not specifically limited.
In sub-step S1030, object or topic to being related in visual information and auditory information are carried out based on mood Mark, to generate the policy information on object or the multi-modal output of topic.
Because different brain wave information is that the different object of correspondence or topic are produced, therefore according to the user after determination Mood is classified to different objects, topic, at the same time it can also excellent to object or topic setting according to hobby/disagreeable degree First level, being then based on above- mentioned information carries out deep learning (such as algorithm of support vector machine) to generate man-machine interactive strategy.
In this example, man-machine interaction strategy is also wrapped both comprising brain electric information, the emotional state for different objects/topic Multi-modal output data containing the different emotional states of correspondence.Represented with three base tables:Eeg data table, storage user is not for The brain electric information of same object/topic;Emotional state table, for the emotional state of different brain electric informations;Multi-modal output policy table, For the exportable different other multi-modal output datas of priority of different emotional states.
After robot obtains the multi-modal data from user, can be by searching eeg data table, emotional state table With multi-modal output policy table, the multi-modal data for most matching is exported according to priority.For example, speaking of one in user makes it During offending theme, the emotional state that robot passes through the current brain electricity display of user should according to man-machine interaction policy selection The favorite music of user or film are exported to the user, and then improve the unhappy mood of user.
To sum up, according to the present embodiment, robot can be according to using brain wave information during being interacted with user The multi-modal output data of the man-machine interaction strategy output matching of generation, more can accurately provide the multi-modal output number of matching According to, and the topic that interacts with user can be increased, and improve user's unhealthy emotion, improve the intelligent and class people of intelligent robot Property, enhancing user's viscosity.
Second embodiment
Fig. 3 is the schematic flow sheet of the example two for being related to the multi-modal exchange method for intelligent robot of the invention, The method of the embodiment mainly includes the following steps that, wherein, the step similar to first embodiment is marked with identical label, And its particular content is repeated no more, only difference step is specifically described.
In step s 110, robot obtains the multi-modal input information of user.
In the step s 120, involved object information or topic information in multi-modal input information are parsed.
In step s 130, the brain wave information that the user produces to the object or the topic is obtained, multi-modal people is determined Machine interactive strategy, exports corresponding multi-modal output data, and the brain wave information carries user emotion information.
In step S210, after multi-modal output data is exported, collection user is anti-to the multi-modal output data The brain wave information of feedback, to obtain the feedback mood of the user.
In other words, the step is the validity of the multi-modal output policy before being judged by brain wave information.In user After obtaining multi-modal output data, the data can be responded and feedback output is carried out, the brain wave using EEG equipment collection user is believed Breath, by being pre-processed to brain wave information and the conversion process such as time domain, frequency domain obtains brain electrical characteristic data, further according to these Different moods (such as tranquil, glad, sad, frightened) obtained by characteristic and the before sample according to training are corresponding Brain electrical feature determines the mood of user.
It should be noted that the brain wave information except gathering user, can also simultaneously gather facial expression, the language of user The data such as sound, further the analysis result according to these multi-modal datas judge the emotional state of user.
In step S220, the mood of user feedback is redefined into multi-modal man-machine interaction strategy, what output matched Multi-modal output data.
When strategy is reselected, on the one hand, in the case that the emotional state before user is negative feeling, if with The emotional state of family current feedback is active mood, then according to man-machine interaction strategy, select and the multi-modal output for sending before The more detailed content of object information or the subject information correlation in strategy further improves the interactive entertaining of user to user Property;If the emotional state of user's current feedback is negative feeling, according to man-machine interaction strategy, selection is more with what is sent before The content of object information or the low one-level of subject information priority level in mode output policy helps relax use to user, further The negative feeling at family.On the other hand, in the case that the emotional state before user is active mood, if user's current feedback Emotional state be active mood, then according to man-machine interaction strategy, can select with the multi-modal output policy for sending before Object information or the related more detailed content of subject information be shared with user;If the emotional state of user's current feedback It is negative feeling, then according to man-machine interaction strategy, highest-ranking object information or theme on fancy grade can be selected to believe Breath issues user as multi-modal output data, and help relaxes the negative feeling of user.
3rd embodiment
Fig. 4 is the schematic flow sheet of the example three for being related to the multi-modal exchange method for intelligent robot of the invention, The method of the embodiment mainly includes the following steps that, wherein, the step similar to second embodiment is marked with identical label, And its particular content is repeated no more, only difference step is specifically described.
In step s 110, robot obtains the multi-modal input information of user.
In the step s 120, involved object information or topic information in multi-modal input information are parsed.
In step s 130, the brain wave information that the user produces to the object or the topic is obtained, multi-modal people is determined Machine interactive strategy, exports corresponding multi-modal output data, and the brain wave information carries user emotion information.
In step S210, after multi-modal output data is exported, collection user is anti-to the multi-modal output data The brain wave information of feedback, to obtain the feedback mood of the user.
In step S220, the mood of user feedback is redefined into multi-modal man-machine interaction strategy, what output matched Multi-modal output data.
In step S310, user's current emotional states are obtained and according to emotional state to default multi-modal man-machine interaction Strategy is adjusted, and emotional state is obtained by gathering brain wave information mode.
After the multi-modal output data newly matched to user is exported again, the brain wave of user is gathered by EEG equipment Information, the analysis result according to brain wave information determines the emotional state of user, and the specific method for obtaining emotional state can be adopted Operated with step S210 identicals, man-machine interaction strategy is then adjusted according to emotional state.If for example, preceding once export multimode After state output data, the emotional state of user feedback is negative feeling, and after this exports new multi-modal output data again, The emotional state of user feedback is active mood, then be adjusted the priority level of this multi-modal output data for exporting. By so adjustment, can export and give user more accurate multi-modal information in next time interactive.
Fourth embodiment
Fig. 5 is the structured flowchart of the multi-modal interactive device 400 for intelligent robot of the embodiment of the present application.Such as Fig. 5 Shown, the multi-modal interactive device 400 of the embodiment of the present application mainly includes:Multi-modal information input module 410, multi-modal information Parsing module 420 and multi-modal information output module 430.
Wherein, multi-modal information input module 410, its multi-modal input information for being used to obtain user.
Multi-modal information parsing module 420, it is used to parse involved object or topic in multi-modal input information;
Multi-modal information output module 430, it is used to obtain the brain wave letter that the user produces the object or the topic Breath, determines multi-modal man-machine interaction strategy, exports corresponding multi-modal output data, and the brain wave information carries user emotion Information.Multi-modal information output module 430, further after multi-modal output data is exported, user is to described more for collection for it The brain wave information of mode output data feedback, to obtain the feedback mood of the user;The mood of user feedback is redefined Multi-modal man-machine interaction strategy, the multi-modal output data that output matches.
In addition, the device 400 also includes tactful presetting module 440, it presets multi-modal man-machine interaction strategy, the module 440 further include such as lower unit:Information acquisition unit 4410, mood determining unit 4420 and policy information generation unit 4430。
Information acquisition unit 4410, its be used to gather in advance visual information of the user in training environment, auditory information and Brain wave information.Information acquisition unit 4410 includes EGG equipment, its described brain wave information of collection.Mood determining unit 4420, It is used to parse the brain wave information, determines the mood of user.Policy information generation unit 4430, it is used to regarding The object or topic being related in feel information and auditory information carry out the mark based on mood, many on object or topic to generate The policy information of mode output.
In addition, the device 400 also includes policy information adjusting module 450, it is used to obtain user's current emotional states simultaneously The default multi-modal man-machine interaction strategy is adjusted according to emotional state, the emotional state is by gathering brain wave Information mode is obtained.
By rationally setting, the multi-modal interactive device 400 of the present embodiment can perform first embodiment, second embodiment With each step of 3rd embodiment, here is omitted.
Because the method for the present invention describes what is realized in computer systems.The computer system can for example be set In the control core processor of robot.For example, method described herein can be implemented as what can be performed with control logic Software, it is performed by the CPU in robot operating system.Function as herein described can be implemented as storage to be had in non-transitory Programmed instruction set in shape computer-readable medium.When implemented in this fashion, the computer program includes one group of instruction, When group instruction is run by computer, it promotes computer to perform the method that can implement above-mentioned functions.FPGA can be temporary When or be permanently mounted in non-transitory tangible computer computer-readable recording medium, for example ROM chip, computer storage, Disk or other storage mediums.In addition to being realized with software, logic as herein described can utilize discrete parts, integrated electricity What road and programmable logic device (such as, field programmable gate array (FPGA) or microprocessor) were used in combination programmable patrols Volume, or embodied including any other equipment that they are combined.All such embodiments are intended to fall under model of the invention Within enclosing.
It should be understood that disclosed embodiment of this invention is not limited to ad hoc structure disclosed herein, process step Or material, and the equivalent substitute of these features that those of ordinary skill in the related art are understood should be extended to.Should also manage Solution, term as used herein is only used for describing the purpose of specific embodiment, and is not intended to limit.
" one embodiment " or " embodiment " mentioned in specification means special characteristic, the structure for describing in conjunction with the embodiments Or characteristic is included at least one embodiment of the present invention.Therefore, the phrase " reality that specification various places throughout occurs Apply example " or " embodiment " same embodiment might not be referred both to.
While it is disclosed that implementation method as above, but described content is only to facilitate understanding the present invention and adopting Implementation method, is not limited to the present invention.Any those skilled in the art to which this invention pertains, are not departing from this On the premise of the disclosed spirit and scope of invention, any modification and change can be made in the formal and details implemented, But scope of patent protection of the invention, must be still defined by the scope of which is defined in the appended claims.

Claims (10)

1. a kind of multi-modal exchange method for intelligent robot, it is characterised in that the described method comprises the following steps:
Obtain the multi-modal input information of user;
Parse involved object information or topic information in multi-modal input information;
The brain wave information that the user produces to the object or the topic is obtained, multi-modal man-machine interaction strategy is determined, it is right to export The multi-modal output data answered, the brain wave information carries user emotion information.
2. the multi-modal exchange method of intelligent robot is used for as claimed in claim 1, it is characterised in that also including by such as Lower step presets multi-modal man-machine interaction strategy,
The advance visual information for gathering user in training environment, auditory information and brain wave information;
The brain wave information is parsed, the mood of user is determined;
Object or topic to being related in visual information and auditory information carry out the mark based on mood, to generate on object Or the policy information of the multi-modal output of topic.
3. the multi-modal exchange method of intelligent robot is used for as claimed in claim 1 or 2, it is characterised in that also included:
After multi-modal output data is exported, the brain wave information that collection user feeds back to the multi-modal output data, with Obtain the feedback mood of the user;
The mood of user feedback is redefined into multi-modal man-machine interaction strategy, the multi-modal output data that output matches.
4. the multi-modal exchange method of intelligent robot is used for as claimed in claim 3, it is characterised in that also included:
Obtain user's current emotional states and the default multi-modal man-machine interaction strategy is adjusted according to emotional state;
The emotional state is obtained by gathering brain wave information mode.
5. the multi-modal exchange method for intelligent robot as described in claim any one of 1-4, it is characterised in that described Brain wave information using EEG equipment gather.
6. a kind of multi-modal interactive device for intelligent robot, it is characterised in that described device is included with lower module:
Multi-modal information input module, its multi-modal input information for being used to obtain user;
Multi-modal information parsing module, it is used to parse involved object information or topic information in multi-modal input information;
Multi-modal information output module, it is used to obtain the brain wave information that the user produces the object or the topic, it is determined that Multi-modal man-machine interaction strategy, exports corresponding multi-modal output data, and the brain wave information carries user emotion information.
7. the multi-modal interactive device of intelligent robot is used for as claimed in claim 6, it is characterised in that also including tactful pre- If module, it presets multi-modal man-machine interaction strategy, and the module is further included such as lower unit:
Information acquisition unit, it is used to gather in advance visual information of the user in training environment, auditory information and brain wave letter Breath;
Mood determining unit, it is used to parse the brain wave information, determines the mood of user;
Policy information generation unit, its object or topic for being used to being related in visual information and auditory information are carried out based on feelings The mark of thread, to generate the policy information on object or the multi-modal output of topic.
8. the multi-modal interactive device for intelligent robot as claimed in claims 6 or 7, it is characterised in that
The multi-modal information output module, further after multi-modal output data is exported, user is to described more for collection for it The brain wave information of mode output data feedback, to obtain the feedback mood of the user;The mood of user feedback is redefined Multi-modal man-machine interaction strategy, the multi-modal output data that output matches.
9. the multi-modal interactive device of intelligent robot is used for as claimed in claim 8, it is characterised in that also included:
Policy information adjusting module, it is used to obtain user's current emotional states and according to emotional state to the default multimode State man-machine interaction strategy is adjusted, and the emotional state is obtained by gathering brain wave information mode.
10. the multi-modal interactive device for intelligent robot as described in claim any one of 7-9, it is characterised in that institute Stating information acquisition unit includes EEG equipment, its described brain wave information of collection.
CN201710034038.4A 2017-01-18 2017-01-18 Multi-modal interaction method and device for intelligent robot Active CN106933345B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710034038.4A CN106933345B (en) 2017-01-18 2017-01-18 Multi-modal interaction method and device for intelligent robot

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710034038.4A CN106933345B (en) 2017-01-18 2017-01-18 Multi-modal interaction method and device for intelligent robot

Publications (2)

Publication Number Publication Date
CN106933345A true CN106933345A (en) 2017-07-07
CN106933345B CN106933345B (en) 2020-02-07

Family

ID=59422826

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710034038.4A Active CN106933345B (en) 2017-01-18 2017-01-18 Multi-modal interaction method and device for intelligent robot

Country Status (1)

Country Link
CN (1) CN106933345B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108037905A (en) * 2017-11-21 2018-05-15 北京光年无限科技有限公司 A kind of interaction output method and intelligent robot for intelligent robot
CN108279777A (en) * 2018-02-11 2018-07-13 广东欧珀移动通信有限公司 Brain wave control method and relevant device
CN108392201A (en) * 2018-02-26 2018-08-14 广东欧珀移动通信有限公司 Brain training method and relevant device
CN108494952A (en) * 2018-03-05 2018-09-04 广东欧珀移动通信有限公司 Voice communication processing method and relevant device
CN108594989A (en) * 2018-03-26 2018-09-28 广东欧珀移动通信有限公司 Acquiring brain waves method and relevant device
CN109278051A (en) * 2018-08-09 2019-01-29 北京光年无限科技有限公司 Exchange method and system based on intelligent robot
CN110109541A (en) * 2019-04-25 2019-08-09 广州智伴人工智能科技有限公司 A kind of method of multi-modal interaction
CN110147729A (en) * 2019-04-16 2019-08-20 深圳壹账通智能科技有限公司 User emotion recognition methods, device, computer equipment and storage medium
CN110287305A (en) * 2019-07-03 2019-09-27 浪潮云信息技术有限公司 A kind of intelligent answer management system based on natural language processing
CN110742559A (en) * 2019-10-24 2020-02-04 佛山市云米电器科技有限公司 Floor sweeping robot control method and system based on brain wave detection
CN111046148A (en) * 2018-10-11 2020-04-21 上海智臻智能网络科技股份有限公司 Intelligent interaction system and intelligent customer service robot
US20200201435A1 (en) * 2018-12-20 2020-06-25 Massachusetts Institute Of Technology End-To-End Deep Neural Network For Auditory Attention Decoding
CN112106381A (en) * 2018-05-17 2020-12-18 高通股份有限公司 User experience assessment
CN112518743A (en) * 2020-11-20 2021-03-19 南京航空航天大学 Multi-mode neural decoding control system and method for on-orbit operation of space manipulator
CN116872222A (en) * 2023-06-21 2023-10-13 广州市恒缘健康信息科技有限公司 Psychological consultation robot, psychological consultation system, psychological consultation control method and psychological consultation storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140114588A (en) * 2013-03-19 2014-09-29 세종대학교산학협력단 Method and Apparatus for Emotion recognition based on Complex physiological signals
CN104407699A (en) * 2014-11-24 2015-03-11 深圳信息职业技术学院 Human-computer interaction method, device and system
CN105867633A (en) * 2016-04-26 2016-08-17 北京光年无限科技有限公司 Intelligent robot oriented information processing method and system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140114588A (en) * 2013-03-19 2014-09-29 세종대학교산학협력단 Method and Apparatus for Emotion recognition based on Complex physiological signals
CN104407699A (en) * 2014-11-24 2015-03-11 深圳信息职业技术学院 Human-computer interaction method, device and system
CN105867633A (en) * 2016-04-26 2016-08-17 北京光年无限科技有限公司 Intelligent robot oriented information processing method and system

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108037905A (en) * 2017-11-21 2018-05-15 北京光年无限科技有限公司 A kind of interaction output method and intelligent robot for intelligent robot
CN108279777A (en) * 2018-02-11 2018-07-13 广东欧珀移动通信有限公司 Brain wave control method and relevant device
CN108392201A (en) * 2018-02-26 2018-08-14 广东欧珀移动通信有限公司 Brain training method and relevant device
CN108392201B (en) * 2018-02-26 2022-03-22 Oppo广东移动通信有限公司 Brain training method and related equipment
CN108494952A (en) * 2018-03-05 2018-09-04 广东欧珀移动通信有限公司 Voice communication processing method and relevant device
CN108494952B (en) * 2018-03-05 2021-07-09 Oppo广东移动通信有限公司 Voice call processing method and related equipment
CN108594989A (en) * 2018-03-26 2018-09-28 广东欧珀移动通信有限公司 Acquiring brain waves method and relevant device
CN108594989B (en) * 2018-03-26 2022-03-22 Oppo广东移动通信有限公司 Brain wave acquisition method and related equipment
CN112106381A (en) * 2018-05-17 2020-12-18 高通股份有限公司 User experience assessment
CN109278051A (en) * 2018-08-09 2019-01-29 北京光年无限科技有限公司 Exchange method and system based on intelligent robot
CN111046148A (en) * 2018-10-11 2020-04-21 上海智臻智能网络科技股份有限公司 Intelligent interaction system and intelligent customer service robot
US20200201435A1 (en) * 2018-12-20 2020-06-25 Massachusetts Institute Of Technology End-To-End Deep Neural Network For Auditory Attention Decoding
US11630513B2 (en) * 2018-12-20 2023-04-18 Massachusetts Institute Of Technology End-to-end deep neural network for auditory attention decoding
CN110147729A (en) * 2019-04-16 2019-08-20 深圳壹账通智能科技有限公司 User emotion recognition methods, device, computer equipment and storage medium
CN110109541A (en) * 2019-04-25 2019-08-09 广州智伴人工智能科技有限公司 A kind of method of multi-modal interaction
CN110109541B (en) * 2019-04-25 2022-04-05 广州智伴人工智能科技有限公司 Multi-modal interaction method
CN110287305A (en) * 2019-07-03 2019-09-27 浪潮云信息技术有限公司 A kind of intelligent answer management system based on natural language processing
CN110742559A (en) * 2019-10-24 2020-02-04 佛山市云米电器科技有限公司 Floor sweeping robot control method and system based on brain wave detection
CN110742559B (en) * 2019-10-24 2021-09-14 佛山市云米电器科技有限公司 Floor sweeping robot control method and system based on brain wave detection
CN112518743A (en) * 2020-11-20 2021-03-19 南京航空航天大学 Multi-mode neural decoding control system and method for on-orbit operation of space manipulator
CN116872222A (en) * 2023-06-21 2023-10-13 广州市恒缘健康信息科技有限公司 Psychological consultation robot, psychological consultation system, psychological consultation control method and psychological consultation storage medium
CN116872222B (en) * 2023-06-21 2024-02-06 广州市恒缘健康信息科技有限公司 Psychological consultation robot, psychological consultation system, psychological consultation control method and psychological consultation storage medium

Also Published As

Publication number Publication date
CN106933345B (en) 2020-02-07

Similar Documents

Publication Publication Date Title
CN106933345A (en) For the multi-modal exchange method and device of intelligent robot
Cavallo et al. Emotion modelling for social robotics applications: a review
Feine et al. A taxonomy of social cues for conversational agents
Hu et al. Attention recognition in EEG-based affective learning research using CFS+ KNN algorithm
Zhang et al. Intelligent facial emotion recognition and semantic-based topic detection for a humanoid robot
US20210191506A1 (en) Affective interaction systems, devices, and methods based on affective computing user interface
Pell et al. Emotional speech processing: Disentangling the effects of prosody and semantic cues
Ma et al. ElderReact: a multimodal dataset for recognizing emotional response in aging adults
Wang et al. Multi-modal emotion recognition using EEG and speech signals
CN105760362B (en) A kind of question and answer evaluation method and device towards intelligent robot
CN103412646A (en) Emotional music recommendation method based on brain-computer interaction
CN105912530A (en) Intelligent robot-oriented information processing method and system
Guediche et al. Temporal cortex reflects effects of sentence context on phonetic processing
CN107016046A (en) The intelligent robot dialogue method and system of view-based access control model displaying
CN108186030A (en) A kind of stimulus information provides the cognition index analysis method of device and latent energy value test
CN106502382A (en) Active exchange method and system for intelligent robot
Heraz et al. Recognition of emotions conveyed by touch through force-sensitive screens: Observational study of humans and machine learning techniques
Ramzan et al. Emotion recognition by physiological signals
CN113749656B (en) Emotion recognition method and device based on multidimensional physiological signals
Le Maitre et al. Self-talk discrimination in human–robot interaction situations for supporting social awareness
CN113053492B (en) Self-adaptive virtual reality intervention system and method based on user background and emotion
Marcos et al. Emotional AI in Healthcare: a pilot architecture proposal to merge emotion recognition tools
CN112466435B (en) Psychological coaching scheme determination method and device, storage medium and electronic device
Esposito et al. On the recognition of emotional vocal expressions: motivations for a holistic approach
Jelec Are Abstract Concepts Like Dinosaur Feathers?: Conceptual Metaphor Theory and the conceptualisation strategies in gesture of blind and visually impaired children

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant