CN105843118B - A kind of robot interactive method and robot system - Google Patents

A kind of robot interactive method and robot system Download PDF

Info

Publication number
CN105843118B
CN105843118B CN201610179223.8A CN201610179223A CN105843118B CN 105843118 B CN105843118 B CN 105843118B CN 201610179223 A CN201610179223 A CN 201610179223A CN 105843118 B CN105843118 B CN 105843118B
Authority
CN
China
Prior art keywords
information
interactive object
interactive
image
robot
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610179223.8A
Other languages
Chinese (zh)
Other versions
CN105843118A (en
Inventor
郭家
石琰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Guangnian Wuxian Technology Co Ltd
Original Assignee
Beijing Guangnian Wuxian Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Guangnian Wuxian Technology Co Ltd filed Critical Beijing Guangnian Wuxian Technology Co Ltd
Priority to CN201610179223.8A priority Critical patent/CN105843118B/en
Publication of CN105843118A publication Critical patent/CN105843118A/en
Application granted granted Critical
Publication of CN105843118B publication Critical patent/CN105843118B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B19/00Programme-control systems
    • G05B19/02Programme-control systems electric
    • G05B19/04Programme control other than numerical control, i.e. in sequence controllers or logic controllers
    • G05B19/042Programme control other than numerical control, i.e. in sequence controllers or logic controllers using digital processors
    • G05B19/0423Input/output
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B2219/00Program-control systems
    • G05B2219/20Pc systems
    • G05B2219/23Pc programming
    • G05B2219/23067Control, human or man machine interface, interactive, HMI, MMI

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Automation & Control Theory (AREA)
  • Manipulator (AREA)

Abstract

The invention discloses a kind of robot interactive method and robot systems.The method of the present invention includes:Multi-modal extraneous input information is acquired, external world's input information includes text information, image information, acoustic information, robot self-test information and induction information;It analyzes the extraneous input information and interactively enters information, interactive object characteristic information and interactive environment characteristic information to obtain;The interactive object characteristic information and the interactive environment characteristic information are analyzed and limited with obtaining matched exchange scenario;Semantic parsing is carried out to obtain the interaction intention of interactive object to the information that interactively enters;Under exchange scenario restriction, it is intended to carry out multi-modal interbehavior output according to the interaction.Compared with prior art, method and system of the invention can preferably simulate the analysis generating process of human interaction behavior in person to person's interactive process, to obtain more naturally lively interaction output, substantially increase the application experience of robot.

Description

A kind of robot interactive method and robot system
Technical field
The present invention relates to robot field, a kind of robot interactive method and robot system are in particulard relate to.
Background technology
With the continuous development of computer technology and artificial intelligence technology is constantly progressive.In domestic environments small intelligent The application of robot is also more and more extensive, and the small intelligent robot towards household is being grown rapidly.
The existing small scale robot towards household, majority are the interactive mode of single voice or text, certain robots It is interacted with user using limb action.Although this enriches interactive form to a certain extent, due to alternate acknowledge mechanism It is cured, the alternate acknowledge of robot is stereotyped, and in multiple interactive process, robot is often returned using single response Using the interaction request under the different conditions of family.This is easy to that user's generation is allowed to be weary of, greatly reduces user experience.
Therefore, in order to make the response of robot more natural lively, the user experience of robot is improved, is needed a kind of new Robot interactive method.
Invention content
In order to make the response of robot more natural lively, the user experience of robot is improved, the present invention provides a kind of machines Device people's exchange method, the described method comprises the following steps:
It includes text information, image information, sound letter to acquire extraneous input information described in multi-modal extraneous input information Breath, robot self-test information and induction information;
It analyzes the extraneous input information and interactively enters information, interactive object characteristic information and interactive environment spy to obtain Reference ceases;
The interactive object characteristic information and the interactive environment characteristic information are analyzed to obtain matched friendship Mutual scene limits;
Semantic parsing is carried out to obtain the interaction intention of interactive object to the information that interactively enters;
Under exchange scenario restriction, it is intended to carry out multi-modal interbehavior output according to the interaction.
In one embodiment, the extraneous input information is analyzed to determine interactive object characteristic information, including:
In monitoring described image information whether comprising humanoid to determine whether to have the object that can interact.
In one embodiment, the extraneous input information is analyzed to determine interactive object characteristic information, including:
Interactive object face-image is parsed from described image information when including humanoid in described image information;
Position the interactive object face-image.
In one embodiment, the extraneous input information is analyzed to determine interactive object characteristic information, including:
Interactive object face-image is parsed from described image information when including humanoid in described image information;
It is extracted from the interactive object face-image and analyzes face feature information;
Determine the interactive object mood or interactive object identity that the face feature information is characterized.
In one embodiment, the extraneous input information is analyzed to determine interactive object characteristic information, including:
Monitor in the acoustic information whether comprising interactive object voice to determine whether to have pair that can be interacted As.
In one embodiment, the extraneous input information is analyzed to determine interactive object characteristic information, including:
The interactive object voice is detached when in the acoustic information including interactive object voice;
Parse the interactive object mood or user that the interactive object voice is characterized with the determination interactive object voice Identity.
The invention also provides a kind of robot system, the system comprises:
Acquisition module is configured to acquire multi-modal extraneous input information, and the acquisition module is adopted comprising text information Acquisition means, image information collecting device, acoustic information harvester, robot self-test information collecting device and induction information acquisition Device;
Analysis module is inputted, is configured to analyze the extraneous input information and interactively enters information, interactive object to obtain Characteristic information and interactive environment characteristic information;
Exchange scenario generation module is configured to believe the interactive object characteristic information and the interactive environment feature Breath is analyzed to be limited with obtaining matched exchange scenario;
Semantic meaning analysis module is configured to carry out semantic parsing to the information that interactively enters to obtain the friendship of interactive object Mutually it is intended to;
Interaction output module, is configured in the case where the exchange scenario limits, and is intended to carry out according to the interaction multi-modal Interbehavior output.
In one embodiment, the input analysis module includes humanoid confirmation device, and the humanoid confirmation device is configured to It whether monitors in described image information comprising humanoid to determine whether that there are interactive objects.
In one embodiment, the input analysis module also includes face-image positioning device, the face-image positioning Device is configured to:
Interactive object face-image is parsed from described image information when including humanoid in described image information;
Position the interactive object face-image.
In one embodiment, the input analysis module also includes face-image resolver, the face-image parsing Device is configured to:
It is extracted from the interactive object face-image and analyzes face feature information;
Determine the interactive object mood or interactive object identity that the face feature information is characterized.
Compared with prior art, method and system of the invention can preferably simulate the mankind in person to person's interactive process and hand over The analysis generating process of mutual behavior, to obtain more naturally lively interaction output, substantially increase robot applies body It tests.
The other feature or advantage of the present invention will illustrate in the following description.Also, the present invention Partial Feature or Advantage will be become apparent by specification, or be appreciated that by implementing the present invention.The purpose of the present invention and part Advantage can be realized or be obtained by specifically noted step in specification, claims and attached drawing.
Description of the drawings
Attached drawing is used to provide further understanding of the present invention, and a part for constitution instruction, the reality with the present invention It applies example and is used together to explain the present invention, be not construed as limiting the invention.In the accompanying drawings:
Fig. 1 is method flow diagram according to an embodiment of the invention;
Fig. 2, Fig. 4 and Fig. 6 are the flow that different embodiment according to the subject invention obtains interactive object characteristic information respectively Figure;
Fig. 3 and Fig. 5 is the flow chart that different embodiment according to the subject invention obtains interactive environment characteristic information respectively;
Fig. 7 is system structure schematic diagram according to an embodiment of the invention.
Specific implementation mode
Hereinafter, embodiments of the present invention will be described in detail with reference to the accompanying drawings and examples, whereby implementation personnel of the invention Can fully understand that how the invention applies technical means to solve technical problems, and reach technique effect realization process and according to The present invention is embodied according to above-mentioned realization process.If it should be noted that do not constitute conflict, each embodiment in the present invention And each feature in each embodiment can be combined with each other, be formed by technical solution protection scope of the present invention it It is interior.
Involved robot is by executing agency, driving device, control system and perception in this specification description System is constituted.Include mainly head, upper limb portion, trunk and lower limb portion in executing agency, in driving device, including electric drive dress It sets, fluid pressure drive device and actuating device of atmospheric pressure.Core of the control system as robot is similar to the brain of people, Include mainly processor and joint servo control device.
Sensory perceptual system includes internal sensor and external sensor.External sensor includes camera, microphone, ultrasonic wave (or laser radar, infrared) device, to perceive extraneous much information.Camera can be arranged on head, be similar to human eye.It is super Device can be arranged on any part of trunk or other positions for sound wave (or laser radar, infrared), to second camera The presence of head sense object or external environment.Robot has the sense of hearing, vision collecting ability.
What needs to be explained here is that the concrete structure of robot according to the present invention is not limited to foregoing description.According to Other arbitrary hardware configurations may be used on the basis of method of the present invention can be achieved in actual needs, robot.
Further, method of the invention describes to realize in computer systems.The computer system for example may be used To be arranged in the control core processor of robot.For example, method described herein can be implemented as with control logic to come The software of execution is executed by the CPU in robot control system.Function as described herein can be implemented as being stored in non-temporary Program instruction set in when property visible computer readable medium.When implemented in this fashion, which includes one Group instruction, when group instruction is run by computer, it promotes computer to execute the method that can implement above-mentioned function.It is programmable to patrol Collecting temporarily or permanently to be mounted in non-transitory visible computer readable medium, such as ROM chip, calculating Machine memory, disk or other storage mediums.In addition to software come other than realizing, logic as described herein can utilize discrete portion What part, integrated circuit and programmable logic device (such as, field programmable gate array (FPGA) or microprocessor) were used in combination Programmable logic, or embody including any other equipment that they are arbitrarily combined.All such embodiments are intended to fall within this Within the scope of invention.
The existing small scale robot towards household, majority are the interactive modes of single voice or text, are easy to make us detesting It is tired.In order to improve user experience, current certain robots are interacted using limb action with user.Although this is rich to a certain extent Rich interactive form, but since alternate acknowledge mechanism is cured, the alternate acknowledge of robot is stereotyped, is repeatedly interacting In the process, robot often responds the interaction request under user's different conditions using single response.This is easy to that user is allowed to produce It begins to detest tired, greatly reduces user experience.
In order to make the response of robot more natural lively, the user experience of robot is improved, the present invention proposes a kind of machine Device people's exchange method.The interbehavior of person to person is analysed in depth first according to the method for the present invention.In the friendship of person to person During mutually, most directly simplest is exactly to be talked by language, is then further by word, image and limbs Language is talked.And the above-mentioned interactive mode of simulation is corresponded to, robot in the prior art uses voice, text, image or limb The interactive mode of body action.
The complex situations for further considering Health For All, during being talked, interaction participant can be first to coming The semantic understanding carried out from the interactive information (language, word, image and body language) of interactive object, understands other side and thinks on earth Any meaning is expressed, corresponding answer is then made.In above process, interaction participant is not to be directed to interactive object Semanteme simply directly makes simple response, but according to the concrete condition of the interactive environment and interactive object that are presently in Targetedly modification adjustment is made to response.
For example, for " recently how is everything going " this problem, (it is not if interactive object is friend in work It is very close), it is simple " extremely busy, to hurry in processing company affair everyday " suitable answer to be calculated.And if interactive object is Close friend or household then need further specific " doing, be engage in business everyday, start dizziness, today ... recently ", It can just seem in this way and get close to.
Another example, again for " recently how is everything going " this problem, if it is on the way meeting by chance and hand over Mutual object is obviously had something to do, simple " to hurry in processing company affair, waits and leisure look for a chance we eat together everyday Meal " can be completed to talk with.And if be currently at, park is taken a walk and interactive object is apparent also very boring, can be detailed Talk " busy, to be engage in business everyday, all start dizziness, today ... recently ".
That is, in the interaction of person to person, interaction participant can analyze current environment and obtain for assisting interacting Exchange scenario limit (such as interactive object identity, interactive object state and be presently in environment), limited according to exchange scenario It sets the tone the whole interactive mode and interaction content of itself.
Based on above-mentioned analysis, in order to make the response of robot more natural lively, robot according to the method for the present invention exists Current exchange scenario can be analyzed in interactive process to limit, and is limited according to exchange scenario and generated matched interbehavior output.
Next the specific implementation step based on flow chart detailed description according to the method for the embodiment of the present invention.The stream of attached drawing Step shown in journey figure can execute in the computer system comprising such as a group of computer-executable instructions.Although flowing The logical order of each step is shown in journey figure, but in some cases, it can be with different from shown in sequence execution herein The step of going out or describing.
As shown in Figure 1, in an embodiment according to the method for the present invention, step S100 is first carried out, acquires extraneous input Information.In order to obtain accurately and effectively scene restriction, in the step s 100, the extraneous input information of acquisition includes not only user The interactively entering of (robot interactive object) (voice input, action instruction, word input etc.), and include and user and work as The preceding relevant other information of interactive environment.Specifically, in the present embodiment, robot acquires multi-modal extraneous input information, Extraneous input information includes text information, image information (action message for wherein including user), acoustic information, robot self-test Information (such as posture information of robot itself) and induction information (such as infrared induction ranging information).
Next step S110 is executed, extraneous input information is analyzed.In step s 110, mainly analysis is extraneous The concrete meaning (with the relevant concrete meaning of interbehavior) that input information is included, and next based on extraneous input information Analysis result executes step S111 (acquisition interactively enters information), step S112 (obtaining interactive object characteristic information) and step S113 (obtains interactive environment characteristic information).
In the step 111 of the present embodiment, in 112,113:
It refers to interaction content (voice input, action instruction, the word input that user sends out to robot to interactively enter information Deng), it is the basis that robot makes that corresponding interaction is responded to interactively enter information;
Interactive object characteristic information mainly describes characteristic attribute (the current body of user identity, user emotion, user of user Body state etc.);
The environment that interactive environment characteristic information mainly describes robot and user is presently in.
(step 112,113) can execute step after getting interactive object characteristic information and interactive environment characteristic information Rapid S130, obtain exchange scenario limit step, to interactive object characteristic information and interactive environment characteristic information analyzed with Matched exchange scenario is obtained to limit.When get interactively enter information after (step S111) step S120, language can be executed Adopted analyzing step carries out semantic parsing to obtain the interaction intention of interactive object to interactively entering information.Finally execute step S140, interbehavior export step, under exchange scenario restriction, are intended to the multi-modal interbehavior of progress according to interaction and export.
It should be noted that in the ideal case, based on extraneous input information analysis result can determine it is whole upper State information.But in some cases, since the missing or robot of extraneous input information are in specific state/environment, base Can only determining a part of above- mentioned information in the analysis result of extraneous input information, (for example, current, there is no users, then are not present and hand over Mutual input information, and interactive object characteristic information only includes the label information there is no user).In this case, then it is Step S120,130 are executed according to the information of acquisition, and corresponding interbehavior output (S140) is carried out based on implementing result.
For example, there is no interactively entering information, and interactive object characteristic information only includes the label there is no user In the case of information, into standby mode, the interbehavior of corresponding standby mode is exported.
To sum up, method according to the invention it is possible to preferably simulate point of human interaction behavior in person to person's interactive process Generating process is analysed, to obtain more naturally lively interaction output, substantially increases the application experience of robot.Following base The specific implementation details of the method for the present invention are further described in specific embodiment.
In the extraneous input information that robot is got, image information is a kind of critically important information.In the present invention one In embodiment, acquisition user can be not only analyzed based on image information interactively enters information (such as gesture instruction of user), And image analysis can be based on and obtain interactive object characteristic information and interactive environment characteristic information.
As shown in Fig. 2, acquiring image information in step s 200 and next analyzing image information.In this implementation In example, it can be monitored first based on image information currently with the presence or absence of the object (user) that can be interacted.Step S210 is executed, Humanoid detecting step detects in the image information got with the presence or absence of humanoid.Further, in order to avoid picture, model etc. Influence (be detected as portrait, manikin in picture humanoid) of the similar object to humanoid detection, is also wrapped in step S210 Whether step containing In vivo detection, humanoid in detection image information are live body.
If there is no humanoid, then illustrate to hold at this time there is no the object that can be interacted in current robot visual range Row step S240 exports interactive object characteristic information.In this case, it is labeled with and works as in the interactive object characteristic information of output It is preceding that interactive object is not present.In this way in interbehavior output step later, robot can be according to interactive object feature Information (label that interactive object is not present) output preset interbehavior for no interactive object.
Then the interactive object feature of interactive object is further analyzed when there are interactive object, step is first carried out S220, separated facial image step parse interactive object face-image from the humanoid image confirmed in step S210.It connects down It executes step S231, positions facial image step, interactive object face-image is positioned (namely to currently interaction pair Face/head of elephant is positioned).Step S240 can be executed after the completion of step S240, output includes interactive object face The interactive object characteristic information of portion's framing information (interactive object face/head positioning information).In this way in interaction later Behavior exports in step, and robot can export corresponding interbehavior (example according to interactive object face-image location information As rotary machine head part makes face/head of robot face/eyes face interactive object).
In the present embodiment, interactive object face-image is also further analyzed.After step S220, it can also hold Row step S232, parsing face-image determine interactive object identity, specifically:
Interactive object face-image is parsed from image information when including humanoid in image information;
It is extracted from interactive object face-image and analyzes face feature information;
Determine the interactive object identity that face feature information is characterized.
Step S240 can be executed after step S232, interactive object of the output comprising interactive object identity information is special Reference ceases.In this way in interbehavior output step later, robot can be according to interactive object identity information the output phase The interbehavior (such as using different interactive strategies for robot owner and stranger) answered.
Further, after step S220, step S233 can also be performed, parsing face-image determines interactive object feelings Thread, specifically:
Interactive object face-image is parsed from image information when including humanoid in image information;
It is extracted from interactive object face-image and analyzes face feature information;
Determine the interactive object mood that face feature information is characterized.
Step S240 can be executed after step S233, interactive object of the output comprising interactive object emotional information is special Reference ceases.In this way in interbehavior output step later, robot can be according to interactive object emotional information the output phase The interbehavior (such as different interactive strategies is respectively adopted when interactive object is angry or sadness) answered.
Further, it in an embodiment according to the method for the present invention, is also based on the analysis to image information and obtains Interactive environment characteristic information.As shown in figure 3, step S300 acquisition image informations are first carried out.Then step S310 is executed, from figure As isolating the background image except interactive object image in information.Then parsing background image information with confirm interactive object/ Residing for robot interactive environment (whether in the room, current weather condition, lamplight status, around whether have other objects/ It is humanoid etc.) (step S320).Finally output includes the interactive environment characteristic information (step S330) of analysis result.In this way at it In interbehavior output step afterwards, robot can be exported according to residing interactive environment corresponding interbehavior (such as Sun-proof measure is taken when prompt interactive object leaves when sunlight is too strong outside room).
In the extraneous input information that robot is got, acoustic information is also a kind of critically important information.In the present invention In one embodiment, based on acoustic information not only can analyze obtain user interactively enter information (such as the interactive voice of user, Sound instruction), and can also audiovideo analysis acquisition interactive object characteristic information and interactive environment characteristic information.
As shown in figure 4, acquiring acoustic information in step S400 and next analyzing acoustic information.In this implementation In example, it can be monitored first based on acoustic information currently with the presence or absence of the user's (interactive object) for having interaction demand.Execute step Rapid S410, whether interactive object speech detection step, it includes interactive object voice to detect in the acoustic information got.
If not including interactive object voice, illustrate not send out interactive voice within the scope of current robot sound collecting User, at this time execute step S440, export interactive object characteristic information.In this case, the interactive object feature of output It is labeled with that there is currently no interactive objects in information.In this way in interbehavior output step later, robot can root Preset interbehavior is exported for no interactive object according to interactive object characteristic information (label that interactive object is not present).
Then the interactive object feature of interactive object is done further when there are interactive object voice (there are interactive objects) Step S420 is first carried out in analysis, detaches interactive object voice step, and interactive object voice is parsed from acoustic information.It connects down Execute step S431, positioning interaction object carries out source of sound analyzing and positioning interactive object voice to interactive object voice and send out Position (interactive object position).Step S440 can be executed after the completion of step S440, output includes interactive object positioning letter The interactive object characteristic information of breath.In this way in interbehavior output step later, robot can be according to interactive object The corresponding interbehavior of location information output (such as rotary machine head part makes robot face/eyes face interaction pair As).
In the present embodiment, interactive object voice is also further analyzed.After step S420, step can also be performed Rapid S432, parsing interactive object voice determine interactive object identity, specifically, carrying out voiceprint analysis to interactive object voice with true Determine the user identity corresponding to interactive object voice.
Step S440 can be executed after step S432, interactive object of the output comprising interactive object identity information is special Reference ceases.In this way in interbehavior output step later, robot can be according to interactive object identity information the output phase The interbehavior (such as using different interactive strategies for robot owner and stranger) answered.
Further, after step S420, step S433 can also be performed, parsing interactive object voice determines interaction pair As mood, specifically, carrying out voiceprint analysis to interactive object voice to determine interactive object feelings that interactive object voice is characterized Thread.
Step S440 can be executed after step S433, interactive object of the output comprising interactive object emotional information is special Reference ceases.In this way in interbehavior output step later, robot can be according to interactive object emotional information the output phase The interbehavior (such as different interactive strategies is respectively adopted when interactive object is angry or sadness) answered.
Further, it in an embodiment according to the method for the present invention, is also based on the analysis to acoustic information and obtains Interactive environment characteristic information.As shown in figure 5, step S500 acquisition acoustic informations are first carried out.Then step S510 is executed, from sound The background sound except interactive object voice is isolated in message breath.Then parsing background sound information with confirm interactive object/ Interactive environment residing for robot (is in quiet interior is in the big near roads of vehicle flowrate, whether surrounding has it Other people etc.) (step S520).Finally output includes the interactive environment characteristic information (step S530) of analysis result.In this way at it In interbehavior output step afterwards, robot can be exported according to residing interactive environment corresponding interbehavior (such as In the big near roads of vehicle flowrate (in background sound information there are a large amount of vehicles traveling and whistle sound) when remind user from Traffic safety is paid attention to when opening).
It should be noted that in Fig. 3 or embodiment illustrated in fig. 4, the interactive object characteristic information of final output includes to be It is no that there are the expression information of interactive object, interactive object Face location information, interactive object identity information and interactive object feelings Thread information.Of course, in practical implementation, according to specific image information, one or more of above- mentioned information can be with For sky.In addition, in other embodiments of the present invention, can also simplify step according to specific interaction demand, cancel said one Or the generation step of multiple information (such as in the occasion that need not be directed to user identity and take distinct interaction strategy, can be cancelled Step S232 or step S432).
In addition, in Fig. 2-embodiment illustrated in fig. 5, it is based respectively on the analysis to image information and acoustic information and obtains in detail Interactive object characteristic information and interactive environment characteristic information.But it is to be herein pointed out in practical implementation, Fig. 2- Embodiment illustrated in fig. 5 is an overall execution flow part respectively.Simple only relies only in Fig. 2-embodiment illustrated in fig. 5 One embodiment can not obtain enough accurately and reliably information.
For example, step S210 is humanoid to determine whether in the presence of pair that can be interacted by whether there is in detection image information As, but when in image information there is no it is humanoid when, it is possible to user is other than robot sight, and user still can be at this time Based on voice and robot interactive.Equally, whether step S410 is sentenced comprising interactive object voice by detecting in acoustic information It is disconnected to whether there is the object that interacted, but when interactive object voice is not present in acoustic information, it is possible to user does not have Any sound is sent out, but there are users in face of robot.
Therefore, the judging result of step S210 or S410 be not one really determine as a result, can only be for assisting Robot is determined further.Based on the above situation, in an embodiment of the present invention, uses and be performed simultaneously step S210 And the mode of step S410, the testing result of humanoid detection and interactive object speech detection is integrated to judge currently whether deposit In interactive object.
As shown in fig. 6, step S600 is first carried out, image information and acoustic information are acquired.Then step S610 is executed, Humanoid detecting step (including In vivo detection step), detects in the image information got with the presence or absence of humanoid.
If there is humanoid, S611 is thened follow the steps, separated facial image step, what is confirmed from step S610 is humanoid Interactive object face-image is parsed in image.Next step S612 is executed, parsing face-image obtains the parsing of face-image As a result.Meanwhile step S613 is executed, facial image step is positioned, interactive object face-image is positioned.
While handling interactive object face-image (or before/after), robot is carried out for acoustic information Processing.Step S630 is executed, whether interactive object speech detection step, it includes interactive object to detect in the acoustic information got Voice.
If including interactive object voice in acoustic information, S631 is thened follow the steps, detaches interactive object voice, and then Step S632 is executed, interactive object voice is parsed.
(step S632 and step S612 have been executed after interactive object voice and interactive object face-image parse Finish) step S640, comprehensive analysis image analysis result and speech analysis result are executed to determine that interactive object characteristic information (is handed over Identity, the mood etc. of mutual object).Particularly, when it is in step S630 the result is that in acoustic information do not include interactive object voice, Image analysis result is only analyzed in step S640.
Step S650 can be executed after the completion of step S613 and/or step S640, export interactive object characteristic information (including interactive object Face location information, interactive object identity and/or interactive object mood).
Particularly, in above process, if the testing result of step S610 is thened follow the steps there is no humanoid S620, whether interactive object speech detection step, it includes interactive object voice to detect in the acoustic information got.If sound Interactive object voice (being yet not present in image information humanoid) is not included in information, then explanation is currently without interactive object, at this time Step S650 is executed, interactive object characteristic information is exported.In this case, it is labeled in the interactive object characteristic information of output There is currently no interactive objects.
Then interactive object voice is done into one when there are interactive object voice (but there is no humanoid in image information) Step analysis, is first carried out step S621, detaches interactive object voice step, and interactive object voice is parsed from acoustic information.It connects Get off to execute step S622, positioning interaction object carries out interactive object voice the hair of source of sound analyzing and positioning interactive object voice Out position (interactive object position).
The interactive object characteristic information comprising interactive object location information can be exported after the completion of step S622.Machine People can rotate head according to interactive object location information and make robot eyes (camera of acquisition image information) face Interactive object (executes step S623).In this way, carrying out humanoid detection (executing step S624) again, robot eyes (are adopted at this time Collect image information camera) obtain image information in just comprising humanoid (user images).
Next step S661, separated facial image (step S611) are executed;Further execute step S662 positioning faces Image (step S613) and step S663 parsing face-images (step S612) are performed simultaneously step S625 parsing interactive objects Voice (step S632);Finally execute step S664 comprehensive analysis image/speech analysis result (step S640).Finally, in step Interactive object characteristic information is exported in rapid S650.
Particularly, if carry out humanoid detection (step S624) again, in the image information that robot obtains still not Including humanoid (user images) (due to being blocked etc. sight), then step S625 is directly executed, current user's language is parsed Sound (step 632) is simultaneously finally exported the result of speech analysis by step S650.
Further, in an embodiment of the present invention, synthesis is also used when analysis obtains interactive environment characteristic information The analysis result of background image (embodiment illustrated in fig. 3) and background sound (embodiment illustrated in fig. 5) analyzes acquisition interactive environment Characteristic information.
The invention also provides a kind of robot systems for exchange method based on the present invention.As shown in fig. 7, in the present invention one In embodiment, robot system includes acquisition module 700, input analysis module 710, exchange scenario generation module 730, semantic solution Analyse module 720 and interaction output module 740.
Acquisition module 700 is configured to acquire multi-modal extraneous input information.In the present embodiment, acquisition module 700 wraps Harvester containing text information 701, image information collecting device 702, acoustic information harvester 703 and induction information acquisition dress Set 704.Further, acquisition module 700 further includes robot self-test information collecting device.Robot self-test information collecting device It can be realized by the self-test component in robot hardware's component, or be realized in a manner of software and hardware combining, do not limited to.
It should be pointed out that according to specific requirements, in other embodiments of the present invention, can be constructed in acquisition module 700 It is one or several in above-mentioned apparatus, or construct the device with other acquisition functions.
Input analysis module 710 is configured to analyze extraneous input information and interactively enters information, interactive object feature with determination Information and interactive environment characteristic information.Exchange scenario generation module 730 is configured to interactive object characteristic information and interaction Environmental characteristic information is analyzed to be limited with obtaining matched exchange scenario.Semantic meaning analysis module 720 is configured to interactively entering Information is carried out semantic parsing and is intended to the interaction for obtaining interactive object.Interaction output module 740 is configured to limit in exchange scenario Under, it is intended to carry out multi-modal interbehavior output according to interaction.
Compared with prior art, system of the invention can preferably simulate human interaction behavior in person to person's interactive process Analysis generating process, to obtain more naturally lively interaction output, the application for substantially increasing robot is experienced.
Specifically, in the present embodiment, input analysis module 710 include for image information humanoid confirmation device 711, Face-image positioning device 712 and face-image resolver 713.Wherein, humanoid confirmation device 711 is configured to monitoring image Whether comprising humanoid to determine whether that there are interactive objects in information.Face-image positioning device 712 is configured to:Work as image Interactive object face-image is parsed from image information when including humanoid in information;Positioning interaction subjects face image.Face figure As resolver 713 is configured to:It is extracted from interactive object face-image and analyzes face feature information;Determine that facial characteristics is believed Breath characterized interactive object mood or interactive object identity.
It includes the interactive object voice confirmation device 714 for being directed to acoustic information, voice positioning dress to input analysis module 710 also Set 715 and interactive object speech analysis device 716.Wherein, interactive object voice confirms that device 714 is configured to monitoring sound letter In breath whether comprising interactive object voice to determine whether that there are interactive objects.Voice positioning device 715 is configured to:Work as sound Source of sound positioning is carried out to determine interactive object position to interactive object voice when including interactive object voice in message breath.Interaction pair As speech analysis device 716 is configured to carry out vocal print parsing to interactive object voice to determine friendship that interactive object voice is characterized Mutual subjects' mood or interactive object identity.
What needs to be explained here is that according to specific requirements, in other embodiments of the present invention, input in analysis module 710 It can construct one or several in above-mentioned apparatus, or construct the device with other analytic functions.
While it is disclosed that embodiment content as above but described only to facilitate understanding the present invention and adopting Embodiment is not limited to the present invention.Method of the present invention can also have other various embodiments.Without departing substantially from In the case of essence of the present invention, those skilled in the art make various corresponding changes or change in accordance with the present invention Shape, but these corresponding changes or deformation should all belong to the scope of the claims of the present invention.

Claims (10)

1. a kind of robot interactive method, which is characterized in that the described method comprises the following steps:
Acquire multi-modal extraneous input information, external world's input information include text information, image information, acoustic information, Robot self-test information and induction information;
It analyzes the extraneous input information and interactively enters information, interactive object characteristic information and interactive environment feature letter to obtain Breath;
The interactive object characteristic information and the interactive environment characteristic information are analyzed to obtain matched interactive feelings Scape limits;
Semantic parsing is carried out to obtain the interaction intention of interactive object to the information that interactively enters;
Under exchange scenario restriction, it is intended to carry out multi-modal interbehavior output according to the interaction.
2. according to the method described in claim 1, it is characterized in that, the analysis extraneous input information is to determine interactive object spy Reference ceases, including:
In monitoring described image information whether comprising humanoid to determine whether to have the object that can interact.
3. according to the method described in claim 2, it is characterized in that, the analysis extraneous input information is to determine interactive object spy Reference ceases, including:
Interactive object face-image is parsed from described image information when including humanoid in described image information;
Position the interactive object face-image.
4. according to the method described in claim 3, it is characterized in that, the analysis extraneous input information is to determine interactive object spy Reference ceases, including:
Interactive object face-image is parsed from described image information when including humanoid in described image information;
It is extracted from the interactive object face-image and analyzes face feature information;
Determine the interactive object mood or interactive object identity that the face feature information is characterized.
5. according to the method described in any one of claim 1-4, which is characterized in that the analysis extraneous input information is with true Determine interactive object characteristic information, including:
Monitor in the acoustic information whether comprising interactive object voice to determine whether to have the object that can be interacted.
6. according to the method described in claim 5, which is characterized in that the analysis extraneous input information is to determine interactive object Characteristic information, including:
The interactive object voice is detached when in the acoustic information including interactive object voice;
Parse the interactive object mood or user identity that the interactive object voice is characterized with the determination interactive object voice.
7. a kind of robot system, which is characterized in that the system comprises:
Acquisition module is configured to acquire multi-modal extraneous input information, and the acquisition module includes text information acquisition dress It sets, image information collecting device, acoustic information harvester, robot self-test information collecting device and induction information acquisition dress It sets;
Analysis module is inputted, is configured to analyze the extraneous input information and interactively enters information, interactive object feature to obtain Information and interactive environment characteristic information;
Exchange scenario generation module, be configured to the interactive object characteristic information and the interactive environment characteristic information into Row analysis is limited with obtaining matched exchange scenario;
Semantic meaning analysis module is configured to carry out semantic parsing to the information that interactively enters to obtain the interaction meaning of interactive object Figure;
Interaction output module, is configured in the case where the exchange scenario limits, and is intended to carry out multi-modal friendship according to the interaction Mutual behavior output.
8. system according to claim 7, which is characterized in that the input analysis module includes humanoid confirmation device, institute Humanoid confirmation device is stated whether to be configured in monitoring described image information comprising humanoid to determine whether that there are interactive objects.
9. system according to claim 8, which is characterized in that the input analysis module also includes face-image positioning dress It sets, the face-image positioning device is configured to:
Interactive object face-image is parsed from described image information when including humanoid in described image information;
Position the interactive object face-image.
10. system according to claim 9, which is characterized in that the input analysis module also includes that face-image parses Device, the face-image resolver are configured to:
It is extracted from the interactive object face-image and analyzes face feature information;
Determine the interactive object mood or interactive object identity that the face feature information is characterized.
CN201610179223.8A 2016-03-25 2016-03-25 A kind of robot interactive method and robot system Active CN105843118B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610179223.8A CN105843118B (en) 2016-03-25 2016-03-25 A kind of robot interactive method and robot system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610179223.8A CN105843118B (en) 2016-03-25 2016-03-25 A kind of robot interactive method and robot system

Publications (2)

Publication Number Publication Date
CN105843118A CN105843118A (en) 2016-08-10
CN105843118B true CN105843118B (en) 2018-07-27

Family

ID=56583915

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610179223.8A Active CN105843118B (en) 2016-03-25 2016-03-25 A kind of robot interactive method and robot system

Country Status (1)

Country Link
CN (1) CN105843118B (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106773820B (en) * 2016-12-02 2019-07-19 北京奇虎科技有限公司 Robot interactive approach, device and robot
CN108614987A (en) * 2016-12-13 2018-10-02 深圳光启合众科技有限公司 The method, apparatus and robot of data processing
CN108227906B (en) * 2016-12-22 2021-04-23 深圳大森智能科技有限公司 Man-machine interaction method and device
CN107053191B (en) 2016-12-31 2020-05-08 华为技术有限公司 Robot, server and man-machine interaction method
CN107016402A (en) * 2017-02-20 2017-08-04 北京光年无限科技有限公司 A kind of man-machine interaction method and device for intelligent robot
CN106959839A (en) * 2017-03-22 2017-07-18 北京光年无限科技有限公司 A kind of human-computer interaction device and method
CN107221332A (en) * 2017-06-28 2017-09-29 上海与德通讯技术有限公司 The exchange method and system of robot
CN107180115A (en) * 2017-06-28 2017-09-19 上海与德通讯技术有限公司 The exchange method and system of robot
CN107704169B (en) * 2017-09-26 2020-11-17 北京光年无限科技有限公司 Virtual human state management method and system
CN108052506B (en) * 2017-12-28 2021-06-29 Oppo广东移动通信有限公司 Natural language processing method, device, storage medium and electronic equipment
CN108334583B (en) * 2018-01-26 2021-07-09 上海智臻智能网络科技股份有限公司 Emotion interaction method and device, computer readable storage medium and computer equipment
WO2019144542A1 (en) 2018-01-26 2019-08-01 Institute Of Software Chinese Academy Of Sciences Affective interaction systems, devices, and methods based on affective computing user interface
CN108780361A (en) * 2018-02-05 2018-11-09 深圳前海达闼云端智能科技有限公司 Human-computer interaction method and device, robot and computer readable storage medium
CN108733208A (en) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 The I-goal of smart machine determines method and apparatus
CN110969053B (en) * 2018-09-29 2023-12-22 深圳市神州云海智能科技有限公司 Method and device for classifying players and lottery robot

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1956528A2 (en) * 2007-02-08 2008-08-13 Samsung Electronics Co., Ltd. Apparatus and Method for Expressing Behavior of Software Robot
CN101604204A (en) * 2009-07-09 2009-12-16 北京科技大学 Distributed cognitive technology for intelligent emotional robot
CN101618280A (en) * 2009-06-30 2010-01-06 哈尔滨工业大学 Humanoid-head robot device with human-computer interaction function and behavior control method thereof
CN101661569A (en) * 2009-09-18 2010-03-03 北京科技大学 Intelligent emotional robot multi-modal behavioral associative expression system
CN103413113A (en) * 2013-01-15 2013-11-27 上海大学 Intelligent emotional interaction method for service robot
CN103679203A (en) * 2013-12-18 2014-03-26 江苏久祥汽车电器集团有限公司 Robot system and method for detecting human face and recognizing emotion
CN104023063A (en) * 2014-06-11 2014-09-03 合肥工业大学 Robot cloud system
US8965576B2 (en) * 2012-06-21 2015-02-24 Rethink Robotics, Inc. User interfaces for robot training
CN104965552A (en) * 2015-07-03 2015-10-07 北京科技大学 Intelligent home environment cooperative control method and system based on emotion robot
CN105058389A (en) * 2015-07-15 2015-11-18 深圳乐行天下科技有限公司 Robot system, robot control method, and robot
CN105082150A (en) * 2015-08-25 2015-11-25 国家康复辅具研究中心 Robot man-machine interaction method based on user mood and intension recognition

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8892256B2 (en) * 2008-01-28 2014-11-18 Seegrid Corporation Methods for real-time and near real-time interactions with robots that service a facility
US8818556B2 (en) * 2011-01-13 2014-08-26 Microsoft Corporation Multi-state model for robot and user interaction

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1956528A2 (en) * 2007-02-08 2008-08-13 Samsung Electronics Co., Ltd. Apparatus and Method for Expressing Behavior of Software Robot
CN101618280A (en) * 2009-06-30 2010-01-06 哈尔滨工业大学 Humanoid-head robot device with human-computer interaction function and behavior control method thereof
CN101604204A (en) * 2009-07-09 2009-12-16 北京科技大学 Distributed cognitive technology for intelligent emotional robot
CN101661569A (en) * 2009-09-18 2010-03-03 北京科技大学 Intelligent emotional robot multi-modal behavioral associative expression system
US8965576B2 (en) * 2012-06-21 2015-02-24 Rethink Robotics, Inc. User interfaces for robot training
CN103413113A (en) * 2013-01-15 2013-11-27 上海大学 Intelligent emotional interaction method for service robot
CN103679203A (en) * 2013-12-18 2014-03-26 江苏久祥汽车电器集团有限公司 Robot system and method for detecting human face and recognizing emotion
CN104023063A (en) * 2014-06-11 2014-09-03 合肥工业大学 Robot cloud system
CN104965552A (en) * 2015-07-03 2015-10-07 北京科技大学 Intelligent home environment cooperative control method and system based on emotion robot
CN105058389A (en) * 2015-07-15 2015-11-18 深圳乐行天下科技有限公司 Robot system, robot control method, and robot
CN105082150A (en) * 2015-08-25 2015-11-25 国家康复辅具研究中心 Robot man-machine interaction method based on user mood and intension recognition

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于头眼行为的交互意图检测方法;赵其杰 等;;《仪器仪表学报》;20141015;第35卷(第10期);全文 *

Also Published As

Publication number Publication date
CN105843118A (en) 2016-08-10

Similar Documents

Publication Publication Date Title
CN105843118B (en) A kind of robot interactive method and robot system
CN105868827B (en) A kind of multi-modal exchange method of intelligent robot and intelligent robot
CN107030691B (en) Data processing method and device for nursing robot
McColl et al. A survey of autonomous human affect detection methods for social robots engaged in natural HRI
CN107894833B (en) Multi-modal interaction processing method and system based on virtual human
CN106956271B (en) Predict the method and robot of affective state
CN112034977B (en) Method for MR intelligent glasses content interaction, information input and recommendation technology application
Zhang et al. Intelligent facial emotion recognition and semantic-based topic detection for a humanoid robot
CN106997243B (en) Speech scene monitoring method and device based on intelligent robot
CN108942919B (en) Interaction method and system based on virtual human
CN106361356A (en) Emotion monitoring and early warning method and system
CN109086860B (en) Interaction method and system based on virtual human
CN106933345A (en) For the multi-modal exchange method and device of intelligent robot
CN106502382B (en) Active interaction method and system for intelligent robot
CN105912530A (en) Intelligent robot-oriented information processing method and system
JP2021146214A (en) Techniques for separating driving emotion from media induced emotion in driver monitoring system
Qi-rong Research on intelligent tutoring system based on affective model
CN108960191B (en) Multi-mode fusion emotion calculation method and system for robot
WO2020175969A1 (en) Emotion recognition apparatus and emotion recognition method
US20230129746A1 (en) Cognitive load predictor and decision aid
JP7469467B2 (en) Digital human-based vehicle interior interaction method, device, and vehicle
CA3206212A1 (en) Methods and systems enabling natural language processing, understanding and generation
Zhang et al. ECMER: Edge-Cloud Collaborative Personalized Multimodal Emotion Recognition Framework in the Internet of Vehicles
US20220009082A1 (en) Method for controlling a plurality of robot effectors
Mansouri Benssassi et al. Wearable assistive technologies for autism: opportunities and challenges

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant