WO2019169854A1 - Human-computer interaction method, and interactive robot - Google Patents

Human-computer interaction method, and interactive robot Download PDF

Info

Publication number
WO2019169854A1
WO2019169854A1 PCT/CN2018/106780 CN2018106780W WO2019169854A1 WO 2019169854 A1 WO2019169854 A1 WO 2019169854A1 CN 2018106780 W CN2018106780 W CN 2018106780W WO 2019169854 A1 WO2019169854 A1 WO 2019169854A1
Authority
WO
WIPO (PCT)
Prior art keywords
interaction
user
information
robot
content
Prior art date
Application number
PCT/CN2018/106780
Other languages
French (fr)
Chinese (zh)
Inventor
乔倚松
吴海周
Original Assignee
南京阿凡达机器人科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 南京阿凡达机器人科技有限公司 filed Critical 南京阿凡达机器人科技有限公司
Publication of WO2019169854A1 publication Critical patent/WO2019169854A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/0005Manipulators having means for high-level communication with users, e.g. speech generator, face recognition means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models

Definitions

  • the invention relates to the field of human-computer interaction, in particular to a human-computer interaction method and an interactive robot.
  • Robot is an emerging comprehensive discipline developed in recent decades. It concentrates on the latest research results in various disciplines such as mechanical engineering, electronic engineering, information science, automatic control and artificial intelligence. It is one of the most active research fields in science and technology. . With the development of science and technology, service robots have been widely used.
  • a good human-computer interaction experience is the key to the service performance of service-oriented robots, and it is also the most basic user demand for robots.
  • mainstream service robots generally have a certain degree of human-computer interaction.
  • the usual human-computer interaction includes PC-type mouse-and-key human-computer interaction, touch-sliding human-computer interaction in tablet mobile phones, and voice human-computer interaction.
  • the human-computer interaction mode of the voice interaction class has become one of the most important human-computer interaction modes of service-oriented robots due to the convenience and nature of the interaction mode and the low cost of interactive learning.
  • the mainstream voice human-computer interaction mode mainly uses passive voice activation mode to activate the whole human-computer interaction process.
  • the robot continuously listens to the user's voice command and starts to perform voice recognition after receiving a specific voice command. Identify the specific content and give the user the corresponding answers and feedback.
  • human-computer interaction is relatively passive, the robot does not actively communicate with the user, and the attraction to the user is not strong enough; and the interactive content of the interaction is relatively rigid, not flexible enough, and the answer is different for different people.
  • the content is the same, not personalized enough. Will reduce the user's experience.
  • the present invention provides an active personalized human-computer interaction.
  • the object of the present invention is to provide a human-computer interaction method and an interactive robot.
  • a personalized interactive interaction can be realized according to different user information of different users and combined with information of the current environment.
  • the invention provides a method for human-computer interaction, comprising the steps of: S1 acquiring user comprehensive information when the robot detects a user who needs active interaction, the user comprehensive information including the personal information of the user, and the current robot system Environment information; S2 generates active interaction content that matches the user comprehensive information; S3 actively interacts with the user according to the active interaction content.
  • the step includes: S0 setting an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and an interaction used in the interaction.
  • Interactive resources
  • step S2 specifically includes: S20, according to the interaction rule, acquiring an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
  • the interaction resource includes: voice content, action content, or multimedia content;
  • the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
  • the personal information of the current user includes: gender, age, expression, face angle, face space location, number of face appearances, user name, number of faces currently detected, and voice information;
  • the environmental information of the robot system includes: time, location, temperature, weather, network connection status, system language.
  • the present invention also provides an interactive robot, comprising: an information acquisition module, configured to acquire user comprehensive information when the robot detects a user who needs active interaction, the user comprehensive information including personal information of the user And the current environment information of the robot system; the processing module is electrically connected to the information acquisition module, and configured to generate active interaction content that matches the user comprehensive information; and the interaction module is configured to use the active interaction content and the user according to the Make active interactions.
  • an information acquisition module configured to acquire user comprehensive information when the robot detects a user who needs active interaction, the user comprehensive information including personal information of the user And the current environment information of the robot system
  • the processing module is electrically connected to the information acquisition module, and configured to generate active interaction content that matches the user comprehensive information
  • the interaction module is configured to use the active interaction content and the user according to the Make active interactions.
  • the scene setting module is further configured to set an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
  • the processing module is further configured to acquire, according to the interaction rule, an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
  • the interaction resource includes: voice content, action content, or multimedia content;
  • the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
  • the personal information of the current user includes: gender, age, expression, face angle, face space location, number of face appearances, user name, number of faces currently detected, and voice information;
  • the environmental information of the robot system includes: time, location, temperature, weather, network connection status, system language.
  • the interaction mode of the robot is different from the current passive interaction.
  • the robot can actively interact with the user, thereby attracting the customer to participate in the interaction and improving the interaction experience.
  • the robot acquires the personal information of the current user and the environmental information of the current robot system, and comprehensively forms the active interactive content, so that the active interactive content can better conform to the current environment and the personal characteristics of the user, and the user is more integrated into the human.
  • the machine interaction improve the experience of human-computer interaction.
  • the robot After the robot recognizes the current user, the user can obtain the personal information of the user, such as age, gender, expression, etc., and obtain the current facial expression of the user. If the user speaks, the voice information will also be obtained. This information is combined to form personalized user information.
  • the robot can obtain the current date, time, location of the robot, weather and other environmental information through the network system or the current system of the robot. The robot will generate corresponding active interactive content based on the comprehensive information of the user with the user's personality, so that the interaction is closer to the user and the intelligence of the interaction is improved.
  • FIG. 1 is a flow chart of an embodiment of a method for human-computer interaction of the present invention
  • FIG. 2 is a flow chart of another embodiment of a method for human-computer interaction of the present invention.
  • FIG. 3 is a flow chart of still another embodiment of a method for human-computer interaction of the present invention.
  • FIG. 4 is a schematic structural view of an embodiment of an interactive robot according to the present invention.
  • 1-Scenario setting module 2-information acquiring module, 3-user integrated information module, 4-processing module, 41-matching sub-module, 42-interactive content generating sub-module, 5-interaction module.
  • the present invention provides an embodiment of a human-computer interaction method, as shown in FIG. 1 , including:
  • S2 generates active interactive content that matches the user comprehensive information
  • the step S1 includes the step of: S0 setting an application scenario of the robot that matches the environment information according to the environment information of the current robot, the application scenario includes an interaction rule of the robot interaction and an interaction used during the interaction. Resources.
  • the application scenario includes interaction rules and interaction resources required for interaction when the robot interacts in the scenario.
  • the user customizes the customized user scene by selecting or based on the webpage, application and other platforms, and deploys it to the robot in real time, so as to avoid the change of the system level, the robot can satisfy the rapid application in different environments.
  • the step S2 generating the active interaction content that matches the user comprehensive information specifically includes: S20 acquiring, according to the interaction rule, an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
  • the user can set an application scenario that is more in line with the current use environment according to the scenario of the robot application, and can better conform to the current environment in the process of human-computer interaction, so that the user is more integrated into the human-computer interaction, thereby improving The experience of human-computer interaction.
  • Application scenarios include interaction rules and interaction resources. If the robot is used in a mall, it will increase the shopping-related interaction rules and interaction resources. For example, the interaction rule is to ask the user which products to purchase when the robot recognizes the user, or to ask the user mall which services are not enough, etc.; the interactive resources can set cheerful songs, or clapping hands, squatting actions, in the user When responding to purchased goods, you can make a clap action and encourage users to continue spending.
  • the robot If the robot is used in a hospital, it will increase the interaction rules and interaction resources related to medical treatment, drugs, and epidemic prevention. For example, when the robot recognizes the user's sadness, the user will be asked why the user is not happy, and comfort the user, while making the action of patting the user's shoulder and playing some cheerful music; some cheerful music can be set in the interactive resource. Shoulder, refueling and other actions.
  • the robot When the robot recognizes the user, it will obtain the current user information through face recognition, and obtain the environment information of the robot through the internal system of the robot, and finally form the user comprehensive information with the user's personalization. Then the robot will get the current interactions according to the interaction rules in the application scenario, retrieve the corresponding interaction resources in the interaction resources, form active interaction content, and actively interact with the user. Different from the current interaction mode, many current interaction methods are passive interactions. The robot continuously listens to the user's voice commands, and after receiving a specific voice command, starts voice recognition, and performs for the user according to the specific content of the recognition. Give the corresponding answers and feedback. Such human-computer interaction is relatively passive, and the robot does not actively communicate with the user, which is not attractive to the user. In the present invention, after detecting the user, the robot can form the active interactive content with the user personalized through the obtained comprehensive information of the user, and actively interact with the user, thereby attracting the user to participate in the human-computer interaction.
  • the user can add shopping-related interaction rules and interaction resources.
  • the interaction rule is: after the robot recognizes the user, first identifies the user, obtains the current user information and the environment information, and if it is an adult, it will ask the user which products to purchase and respond, and then ask the user mall which services are not enough. In place, respond; finally say goodbye to the user. If it is a child, it will first greet the children and dance for the children.
  • the user can set a cheerful song in the interactive resource, or an action of clap, slap, wave, squat, etc., when the user responds to the purchased product, can make a clap action, and encourage the user to continue to consume.
  • the robot When the user identified by the robot is a lady, the current time is 10:30 am, the robot will say: “Mrs. Good morning! What products did you buy today?" At the same time, make a wave of gestures to say hello to the lady. The lady responded: “Buy a piece of clothing.” The robot will respond: “Thank you for your support of the mall!” At the same time, he made awkward moves. Then continue to say: “There are not enough services in the mall,” and then record what the user said. Finally, say “goodbye” to the user.
  • the interaction mode of the robot is also very different due to the different user comprehensive information embodied by the user.
  • the active interaction content of the robot for each user is the same, after the user is identified. It will say: "Hello, what can I help you?" Unlike the present invention, it is possible to output personalized active interactive content according to different characteristics carried by different users.
  • the above interaction rules and interaction resources can be set by themselves.
  • the robot will find corresponding responses in the interaction rules according to the user comprehensive information, and generate active interaction content in combination with the interaction resources to perform active interaction. It can be seen that the comprehensive information of users of different users is different.
  • the robot generates different active interactive content according to different genders, different expressions and different ages of different users, so that the active interactive content is personalized.
  • the present invention also provides an embodiment of a method for human-computer interaction, including:
  • S0 sets an application scenario of the robot that matches the environment information according to the environment information of the current robot, and the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
  • S202 Generate active interactive content according to the interaction resources corresponding to the plurality of candidate rule nodes; the active interaction content includes voice interaction content, action interaction content, and multimedia interaction content.
  • the interaction resource includes: voice content, action content, or multimedia content; the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
  • the interaction rule includes: whether the robot responds to certain specific detection results when interacting, for example, whether to query the user name, whether to detect the system time, etc.; and when the detection result of the response is selected, the specific content of the corresponding feedback is selected. For example, different names for different genders, different actions for different users, and so on.
  • the interaction resources include: resources required under corresponding interaction rules, such as all voice text content, all optional action interaction content, all music and video content, and the like.
  • the current user's personal information includes: gender, age, expression, face angle, face space location, face appearance times, user name, current detected face number, and voice information;
  • the current robot system Environmental information includes: time, location, temperature, weather, network connection status, system language.
  • the system performs various identification and detection by human body recognition, face recognition, environment system detection, etc., and comprehensively outputs a set of user information descriptions for the user to describe the specific information of the current interactive user.
  • the user's comprehensive information form is expressed as follows:
  • the string identifier of the string identifies the specific content of the string to distinguish it from other types of strings.
  • the key is a feature in the user comprehensive information, that is, a user feature keyword and an environmental feature keyword, and each key represents a feature used to describe the current user, and the features may include: face number, face name, gender, Age, time, expression, face angle, face position, face size, weather, location, temperature, type of exercise, network connection, etc.
  • the value value is a specific parameter value corresponding to the current key value, that is, the user characteristic parameter value and the environmental characteristic parameter value described in the present invention.
  • the ⁇ key:value> pair can modify the number and content of the user description features according to different human body recognition, face recognition, system detection, motion detection and other detection and recognition tool output changes.
  • the key, value of each pair represents an information feature of the user.
  • the specific can be interpreted as: the user's face information is continuous; user name: avatar; gender: male; age: 53 years old; record generation time: 13:04; user face expression: no; face angle roll value :-13.61209 degrees; face angle pitch value: 23.196611 degrees; face angle yaw value: 23.330135 degrees; face record number: 1; the number of faces in the current picture; 1; the user's face is in the total person
  • the face is the first one; the face position X value: 646px; the face position Y value: 189px; the face width: 352px; the face length: 352px;
  • the method generates a set of active interactive content corresponding to the input user comprehensive information.
  • the active interactive content of the feedback includes the following three types: voice interactive content, action interactive content, and multimedia interactive content.
  • the voice interactive content is an active voice prompt played by the robot;
  • the action interactive content is a set of sports content of the active part of the head, the limbs, etc.;
  • the multimedia interactive content includes pictures, music, video, application, etc., and passes through the chest of the robot.
  • the display platform is played.
  • the content of the multimedia can be played simultaneously with the voice prompt, or can be played after the voice prompt ends, to meet the needs of different scenes.
  • the rule node includes a Node node, and the interaction rule stores the voice, action, and multimedia interaction content corresponding to each recognition result in a tree-shaped data structure.
  • the interaction rule tree includes a plurality of Node nodes, each of which includes a plurality of preset feature keywords and corresponding preset feature parameter values, and also includes multiple interaction resources such as voice, action, and multimedia. .
  • the Key value in the Node node describes the necessary conditions for the set of statements, actions, and multimedia to be selected.
  • each group of Node nodes will match the currently integrated user information. If the current user comprehensive information is satisfied with the necessary conditions of the Node node, the Node node will become a candidate. Node, wait for later selection. If the current user comprehensive information and the necessary conditions of the Node node are not completely satisfied, the Node node will not become a candidate rule node.
  • the present invention further provides an embodiment of a method for human-computer interaction, including:
  • S0 sets an application scenario of the robot that matches the environment information according to the environment information of the current robot, and the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
  • S2011 determines whether each rule node meets a preset condition
  • S2021 analyzes a priority value of each of the candidate rule nodes, and sorts the candidate rule nodes according to the priority value; and randomly selects one candidate rule node randomly or weighted for multiple candidate rule nodes of the same priority value. Participate in sorting;
  • S2022 combines the interaction resources corresponding to the sorted candidate rule nodes in sequence to generate the active interaction content.
  • the preset condition is determining whether all preset feature keywords and corresponding preset feature parameter values in each rule node are related to partial feature keywords and corresponding features in the user comprehensive information.
  • the parameter values are the same.
  • the feature keyword includes a user feature keyword and an environment feature keyword
  • the feature parameter value includes a user feature parameter value and an environment feature parameter value.
  • the feature keyword is a key value
  • the feature parameter value is a value corresponding to the key value
  • the rule node is a Node node, that is, a specific step of filtering out a plurality of candidate rule nodes that match the user comprehensive information according to a preset condition: If the Value value of all the features in the Node node is equal to the Value value of the same feature in the user comprehensive information, the Node node is used as the candidate rule node. If the value of the feature in the Node node is All, it means that the feature is considered to be satisfied for the Value result of the corresponding feature in all user information.
  • the features in the user's comprehensive information are usually more than the feature values required by the Node node. For the extra feature values, the system does not judge and filter based on the results.
  • a complete voice prompt will be combined according to the value of its priority value Priority.
  • the method decomposes a complete speech interaction content into different sentence segments, and each sentence segment is a segment of a complete voice prompt.
  • the priority value Priority in each Node node indicates the position of the statement segment in the complete statement.
  • a complete statement into multiple statement segments, such as a title segment, a time segment, a content segment, and a problem segment.
  • the invention does not limit the number of segments of the statement, and the user can segment the segment according to the completeness of the statement. There is a free combination between each paragraph and the next paragraph. Therefore, the final combination of the complete statement will become very flexible.
  • the method selects a Node node that satisfies the condition by random selection.
  • the voice interaction content to be generated includes a calling segment, a greeting segment, and a content segment.
  • the calling segment has two Node nodes with the same priority value, namely “old man” and “old man”; the greeting segment has three Node nodes with the same priority value. "Hello! "Good morning! "Good morning!”
  • the content segment has a Node node, which is "You are in good health.”
  • a plurality of Item options (ie, content executed by the robot) may be set in advance, and for a plurality of Item options existing in the Node node, a final result that satisfies the condition is selected according to the difference of the Key values thereof. If the same Key value result has multiple corresponding Item options, the method selects a result as a final result by sequential selection or random selection, and outputs the result.
  • the candidate rule nodes of the same priority value only randomly select one participation order, for each candidate rule.
  • the content of the interaction is different for the same scenario, and the fixed mode interaction is not very rigid.
  • the action corresponding to the key value may be three item options: handshake, wave, and salute.
  • one of the items may be selected for output.
  • the greeting language can be set to multiple, assigned to different Item options.
  • the language of the output will be different, and the content will be diversified when interacting.
  • the action interaction content and the multimedia interaction content exist by being appended to each specific voice sentence.
  • additional actions and multimedia interactions after each specific statement when the active feedback content combination is completed, the corresponding actions and multimedia interaction content are also generated.
  • the action and multimedia interaction content is determined by the additional content of the last statement component. Therefore, different content, different lengths of statements can also make different actions and prompts on multimedia interactive content.
  • Flexible use of voice scripts and resource scripts, as the user's comprehensive information input, the feedback information generated by this method can also be changed accordingly.
  • the collected user comprehensive information may include: male, 8 years old, name, facial expression crying, the current time is 10 am, the current weather is 2 degrees Celsius, the current location is In the hospital, the air quality is good, and after the value is assigned, a string is formed for the system to retrieve.
  • the robot finds the candidate rule nodes that satisfy the matching condition in the rule node according to the user comprehensive information, sorts the rule nodes according to the priority values, and sequentially combines the interaction resources corresponding to the candidate rule nodes. After the robot recognizes the little boy, it will greet the little boy actively.
  • the present invention provides an embodiment of an interactive robot, including:
  • the information obtaining module 2 is configured to acquire current user information and environment information
  • the user integrated information module 3 is electrically connected to the information acquiring module 2, and is configured to generate user comprehensive information according to the current user information and environment information;
  • the processing module 4 is electrically connected to the user integrated information module 3, and configured to generate active interactive content according to the user comprehensive information and the application scenario;
  • the interaction module 5 is configured to perform active human-computer interaction according to the active interaction content.
  • the interactive robot further includes: a scene setting module 1 configured to set an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and The interactive resources used in the interaction.
  • a scene setting module 1 configured to set an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and The interactive resources used in the interaction.
  • the scene setting module 1 is further configured to store a preset interaction rule of the robot, and an interaction resource required when the robot interacts under the interaction rule, and use the interaction rule and the application resource as An application scenario;
  • the interaction rule includes multiple rule nodes, and each rule node includes multiple preset feature keywords and corresponding preset feature parameter values.
  • the user integrated information module 3 is further configured to assign a corresponding user feature keyword and a user feature parameter value to each user feature in the current user information, and assign a corresponding to each environment feature in the environment information.
  • the environment feature keyword and the environment feature parameter value; the user integrated information module 3 is further configured to combine the user feature keyword and the corresponding user feature parameter value, the environment feature keyword, and the corresponding environment feature parameter value into a string And assigning the string to the corresponding string identifier; using the string with the string identifier as the current user comprehensive information.
  • the processing module 4 specifically includes: a matching sub-module 41, configured to filter, in the application scenario, a plurality of candidate rule nodes that match the user comprehensive information; and an active interaction content generation sub-module 42 according to The interaction resources corresponding to the plurality of candidate rule nodes generate active interaction content; the active interaction content includes voice interaction content, action interaction content, and multimedia interaction content.
  • a matching sub-module 41 configured to filter, in the application scenario, a plurality of candidate rule nodes that match the user comprehensive information
  • an active interaction content generation sub-module 42 according to The interaction resources corresponding to the plurality of candidate rule nodes generate active interaction content; the active interaction content includes voice interaction content, action interaction content, and multimedia interaction content.
  • the matching sub-module 41 is further configured to determine, by one by one, whether all preset feature keywords and corresponding preset feature parameter values in each rule node are corresponding to some feature keywords in the user comprehensive information.
  • the feature parameter values are the same; the feature keyword includes a user feature keyword and an environment feature keyword, and the feature parameter value includes a user feature parameter value and an environment feature parameter value; if yes, the rule node that satisfies the condition is used as a candidate rule node.
  • the active interaction content generation sub-module 42 is further configured to analyze respective priority values of the plurality of candidate rule nodes, and sort the candidate rule nodes according to the priority value; for the same priority value
  • the candidate rule nodes select one of the candidate rule nodes to participate in the sorting; and the interactive resources corresponding to the sorted candidate rule nodes are sequentially combined to generate the active interactive content.
  • the information obtaining module 2 has two major functions, one is to identify the current user, and obtain the user resource; for example, when it is detected that there is a user within the preset range, the user is initially identified, and the user mainly recognizes the user's face. feature. Through the face recognition technology, the current user's expression can be identified, and combined with the Internet big data, some basic data information of the user is obtained. The second is to obtain environmental information such as the current location, time, weather, etc. of the current robot according to the current robot system.
  • the processing module 4 may be composed of a processor of a robot, and the interaction module 5 includes a voice control system, a display system, a drive system, and the like used in the robot interaction process.
  • the interaction module 5 includes a voice control system, a display system, a drive system, and the like used in the robot interaction process.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Mechanical Engineering (AREA)
  • Robotics (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Manipulator (AREA)

Abstract

A human-computer interaction method, and an interactive robot. The method comprises: when a robot detects a user needing active interaction, acquiring user comprehensive information (S1), wherein the user comprehensive information comprises personal information of a current user, and environment information of a current robot system; generating active interaction content matching the user comprehensive information (S2); and performing active interaction with the user according to the active interaction content (S3). During the process of a robot performing active interaction, active personalized human-computer interaction can be achieved according to different user information of different users and current environment information.

Description

一种人机交互方法及交互机器人Human-computer interaction method and interactive robot
本申请要求2018年03月09日提交的申请号为:201810193098.5、发明名称为“一种人机交互方法及交互机器人”的中国专利申请的优先权,其全部内容合并在此。The present application claims priority to Chinese Patent Application No. 201101193098.5, entitled "A Human-Machine Interaction Method and Interactive Robot", the entire contents of which are incorporated herein by reference.
技术领域Technical field
本发明涉及人机交互领域,尤指一种人机交互方法及交互机器人。The invention relates to the field of human-computer interaction, in particular to a human-computer interaction method and an interactive robot.
背景技术Background technique
机器人是近几十年来发展起来的新兴综合学科,它集中了机械工程、电子工程、信息科学、自动控制以及人工智能等多种学科的最新研究成果,是目前科技发展最活跃的研究领域之一。随着科学技术的发展,服务型机器人已经得到了广泛的应用。Robot is an emerging comprehensive discipline developed in recent decades. It concentrates on the latest research results in various disciplines such as mechanical engineering, electronic engineering, information science, automatic control and artificial intelligence. It is one of the most active research fields in science and technology. . With the development of science and technology, service robots have been widely used.
对于服务型机器人,良好的人机交互体验是服务型机器人服务性能的关键所在,也是用户对机器人最为基本的需求。目前主流的服务性机器人一般都具备一定程度的人机交互功能。通常的人机交互包括了PC类的鼠标键盘式人机交互,平板手机类的触摸滑动式人机交互,以及语音人机交互等等。其中,语音交互类的人机交互方式由于交互模式的便利、自然,以及交互学习成本低等优势,日渐成为了服务型机器人最为重要的人机交互方式之一。For service robots, a good human-computer interaction experience is the key to the service performance of service-oriented robots, and it is also the most basic user demand for robots. At present, mainstream service robots generally have a certain degree of human-computer interaction. The usual human-computer interaction includes PC-type mouse-and-key human-computer interaction, touch-sliding human-computer interaction in tablet mobile phones, and voice human-computer interaction. Among them, the human-computer interaction mode of the voice interaction class has become one of the most important human-computer interaction modes of service-oriented robots due to the convenience and nature of the interaction mode and the low cost of interactive learning.
主流的语音人机交互模式,主要使用的是被动的语音激活方式来激活整个人机交互的流程,机器人通过不断监听用户的语音指令,在接受到特定的语音指令之后,开始进行语音识别,根据识别的具体内容,为用户做出相应的回答与反馈。但是这样的人机交互方式比较被动,机器人不会主动与用户进行交流,对用户的吸引力不够强;并且这种交互的回答内容比较死板,不够灵活多变,并且针对不同的人,其回答的内容都一样,不够个性化。会降低用户的体验感。The mainstream voice human-computer interaction mode mainly uses passive voice activation mode to activate the whole human-computer interaction process. The robot continuously listens to the user's voice command and starts to perform voice recognition after receiving a specific voice command. Identify the specific content and give the user the corresponding answers and feedback. However, such human-computer interaction is relatively passive, the robot does not actively communicate with the user, and the attraction to the user is not strong enough; and the interactive content of the interaction is relatively rigid, not flexible enough, and the answer is different for different people. The content is the same, not personalized enough. Will reduce the user's experience.
因此,为了解决上述弊端,本发明提供了一种主动式的个性化的人机交互。Therefore, in order to solve the above drawbacks, the present invention provides an active personalized human-computer interaction.
发明内容Summary of the invention
本发明的目的是提供一种人机交互的方法及交互机器人,在机器人进行主动交互的过程中,能够根据不同用户的不同用户信息,并结合当前环境的信息,实现个性化的主动交互。The object of the present invention is to provide a human-computer interaction method and an interactive robot. In the process of actively interacting with a robot, a personalized interactive interaction can be realized according to different user information of different users and combined with information of the current environment.
本发明提供的技术方案如下:The technical solution provided by the present invention is as follows:
本发明提供了一种人机交互的方法,包括步骤:S1当机器人检测到需要主动交互的用户时,获取用户综合信息,所述用户综合信息包括所述用户的个人信息,以及当前机器人系统的环境信息;S2生成与所述用户综合信息相匹配的主动交互内容;S3根据所述主动交互内容与用户进行主动交互。The invention provides a method for human-computer interaction, comprising the steps of: S1 acquiring user comprehensive information when the robot detects a user who needs active interaction, the user comprehensive information including the personal information of the user, and the current robot system Environment information; S2 generates active interaction content that matches the user comprehensive information; S3 actively interacts with the user according to the active interaction content.
优选的,在所述步骤S1之前包括步骤:S0根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。Preferably, before the step S1, the step includes: S0 setting an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and an interaction used in the interaction. Interactive resources.
优选的,步骤S2具体包括:S20根据所述交互规则,获取与所述用户综合信息相匹配的交互资源,从而生成主动交互内容。Preferably, step S2 specifically includes: S20, according to the interaction rule, acquiring an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
优选的,所述交互资源包括:语音内容、动作内容、或者多媒体内容;所述交互规则包括多个规则节点,每个规则节点体现了不同用户综合信息与不同交互资源的映射关系。Preferably, the interaction resource includes: voice content, action content, or multimedia content; the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
优选的,所述当前用户的个人信息包括:性别、年纪、表情、人脸角度、人脸空间位置、人脸出现次数、用户名称、当前被检测到人脸数量,以及语音信息;所述当前机器人系统的环境信息包括:时间、地点、温度、天气、网络连接状态、系统语言。Preferably, the personal information of the current user includes: gender, age, expression, face angle, face space location, number of face appearances, user name, number of faces currently detected, and voice information; The environmental information of the robot system includes: time, location, temperature, weather, network connection status, system language.
本发明还提供了一种交互机器人,其特征在于,包括:信息获取模块,用于当机器人检测到需要主动交互的用户时,获取用户综合信息,所述用户综合信息包括所述用户的个人信息,以及当前机器人系统的环境信息;处理模块,与所述信息获取模块电连接,用于生成与所述用户综合信息相匹配的主动交互内容;交互模块,用于根据所述主动交互内容与用户进行主动交互。The present invention also provides an interactive robot, comprising: an information acquisition module, configured to acquire user comprehensive information when the robot detects a user who needs active interaction, the user comprehensive information including personal information of the user And the current environment information of the robot system; the processing module is electrically connected to the information acquisition module, and configured to generate active interaction content that matches the user comprehensive information; and the interaction module is configured to use the active interaction content and the user according to the Make active interactions.
优选的,场景设置模块,还用于根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。Preferably, the scene setting module is further configured to set an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
优选的所述处理模块,还用于根据所述交互规则,获取与所述用户综合信息相匹配的交互资源,从而生成主动交互内容。Preferably, the processing module is further configured to acquire, according to the interaction rule, an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
优选的,所述交互资源包括:语音内容、动作内容、或者多媒体内容;所述交互规则包括多个规则节点,每个规则节点体现了不同用户综合信息与不同交互资源的映射关系。Preferably, the interaction resource includes: voice content, action content, or multimedia content; the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
优选的,所述当前用户的个人信息包括:性别、年纪、表情、人脸角度、人脸空间位置、人脸出现次数、用户名称、当前被检测到人脸数量,以及语音信息;所述当前机器人系统的环境信息包括:时间、地点、温度、天气、网络连接状态、系统语言。Preferably, the personal information of the current user includes: gender, age, expression, face angle, face space location, number of face appearances, user name, number of faces currently detected, and voice information; The environmental information of the robot system includes: time, location, temperature, weather, network connection status, system language.
通过本发明提供的一种人机交互的方法及交互机器人,能够带来以下至少一种有益效果:The human-computer interaction method and the interactive robot provided by the invention can bring at least one of the following beneficial effects:
1、本发明中,机器人的交互方式不同于目前的被动式交互,机器人在识别到用户之后,能够主动与用户进行交互,更加吸引客户参与到交互中来,提高交互体验。同时,机器人在交互中,会获取当前用户的个人信息,以及当前机器人系统的环境信息,综合形成主动交互内容,使得主动交互内容能够更加契合当前的环境和用户的个人特征,让用户更加融入人机交互中,提高人机交互的体验感。1. In the present invention, the interaction mode of the robot is different from the current passive interaction. After the robot recognizes the user, the robot can actively interact with the user, thereby attracting the customer to participate in the interaction and improving the interaction experience. At the same time, in the interaction, the robot acquires the personal information of the current user and the environmental information of the current robot system, and comprehensively forms the active interactive content, so that the active interactive content can better conform to the current environment and the personal characteristics of the user, and the user is more integrated into the human. In the machine interaction, improve the experience of human-computer interaction.
2、机器人会识别当前用户后,通过人脸识别获取到用户的个人信息,如年龄、性别、表情等等,并且获取到用户当前的面部表情,若用户说话,则还会获取语音信息,将这些信息综合起来形成具有个性化的当前用户信息。机器人会通过网络系统或者机器人的当前系统,可以获取到当前的日期、时间、机器人所在地点、天气等等环境信息。机器人会根据这些带有用户个性的用户综合信息,生成对应的主动交互内容,使得交互更加贴近用户,提高交互的智能化。2. After the robot recognizes the current user, the user can obtain the personal information of the user, such as age, gender, expression, etc., and obtain the current facial expression of the user. If the user speaks, the voice information will also be obtained. This information is combined to form personalized user information. The robot can obtain the current date, time, location of the robot, weather and other environmental information through the network system or the current system of the robot. The robot will generate corresponding active interactive content based on the comprehensive information of the user with the user's personality, so that the interaction is closer to the user and the intelligence of the interaction is improved.
附图说明DRAWINGS
下面将以明确易懂的方式,结合附图说明优选实施方式,对一种人机交互的方法及交互机器人的上述特性、技术特征、优点及其实现方式予以进一步说明。The preferred embodiment will be described in a clear and easy-to-understand manner with reference to the accompanying drawings, and a method for human-computer interaction and the above-mentioned characteristics, technical features, advantages and implementation manners of the interactive robot will be further described.
图1是本发明一种人机交互的方法的一个实施例的流程图;1 is a flow chart of an embodiment of a method for human-computer interaction of the present invention;
图2是本发明一种人机交互的方法的另一个实施例的流程图;2 is a flow chart of another embodiment of a method for human-computer interaction of the present invention;
图3是本发明一种人机交互的方法的再一个实施例的流程图;3 is a flow chart of still another embodiment of a method for human-computer interaction of the present invention;
图4是本发明一种交互机器人的一个实施例结构示意图;4 is a schematic structural view of an embodiment of an interactive robot according to the present invention;
附图标号说明:Description of the reference numerals:
1-场景设置模块、2-信息获取模块、3-用户综合信息模块、4-处理模块、41-匹配子模块、42-交互内容生成子模块、5-交互模块。1-Scenario setting module, 2-information acquiring module, 3-user integrated information module, 4-processing module, 41-matching sub-module, 42-interactive content generating sub-module, 5-interaction module.
具体实施方式Detailed ways
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对照附图说明本发明的具体实施方式。显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图,并获得其他的实施方式。In order to more clearly illustrate the embodiments of the present invention or the technical solutions of the prior art, the specific embodiments of the present invention will be described below with reference to the accompanying drawings. Obviously, the drawings in the following description are only some embodiments of the present invention, and those skilled in the art can obtain other drawings according to the drawings without obtaining creative labor, and obtain Other embodiments.
为使图面简洁,各图中只示意性地表示出了与本发明相关的部分,它们并不代表其作为产品的实际结构。另外,以使图面简洁便于理解,在有些图中具有相同结构或功能的部件,仅示意性地绘示了其中的一个,或仅标出了其中的一个。在本文中,“一个”不仅表示“仅此一个”,也可以表示“多于一个”的情形。In order to simplify the drawings, only the parts related to the present invention are schematically shown in the drawings, and they do not represent the actual structure of the product. In addition, in order to make the drawings simple and easy to understand, components having the same structure or function in some of the figures are only schematically illustrated, or only one of them is marked. In the present context, "a" means not only "only one" but also "more than one".
本发明提供了一种人机交互的方法一个实施例,如图1所示,包括:The present invention provides an embodiment of a human-computer interaction method, as shown in FIG. 1 , including:
S1当机器人检测到需要主动交互的用户时,获取用户综合信息,所述用户综合信息包括当前用户的个人信息,以及当前机器人系统的环境信息;S1, when the robot detects a user who needs active interaction, acquiring user comprehensive information, where the user comprehensive information includes personal information of the current user, and environment information of the current robot system;
S2生成与所述用户综合信息相匹配的主动交互内容;S2 generates active interactive content that matches the user comprehensive information;
S3根据所述主动交互内容与用户进行主动交互。S3 actively interacts with the user according to the active interaction content.
优选的,所述步骤S1之前包括步骤:S0根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。Preferably, the step S1 includes the step of: S0 setting an application scenario of the robot that matches the environment information according to the environment information of the current robot, the application scenario includes an interaction rule of the robot interaction and an interaction used during the interaction. Resources.
本实施例中,应用场景包括了在该场景下,机器人交互时,所遵守的交互规则和交互时所需的交互资源。用户通过选择或者基于网页,应用等平台,定制自定义的用户场景,实时部署到机器人上,以达到在避免系统层级的改变的同时,使得机器人满足了在不同环境下的快速应用。In this embodiment, the application scenario includes interaction rules and interaction resources required for interaction when the robot interacts in the scenario. The user customizes the customized user scene by selecting or based on the webpage, application and other platforms, and deploys it to the robot in real time, so as to avoid the change of the system level, the robot can satisfy the rapid application in different environments.
优选的,所述步骤S2生成与所述用户综合信息相匹配的主动交互内容具体包括:S20根据所述交互规则,获取与所述用户综合信息相匹配的交互资源,从而生成主动交互内容。Preferably, the step S2 generating the active interaction content that matches the user comprehensive information specifically includes: S20 acquiring, according to the interaction rule, an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
具体的,在机器人使用前,用户可以预先根据机器人应用的场景,设置更加符合当前使用环境的应用场景,在人机交互过程中能够更加契合当前的环境,让用户更加融入人机交互中,提高人机交互的体验感。应用场景包括交互规则以及交互资源。机器人如果在商场中使用,则会增加和购物相关的交互规则以及交互资源。例如交互规则为机器人识别到用户时,询问用户购买了哪些商品,或是询问用户商场还有哪些服务不够到位等等;交互资源中可以设置欢快的歌曲,或是拍手、鞠躬的动作,在用户回应购买的商品时,可以做出拍手的动作,鼓励用户继续消费。Specifically, before the robot is used, the user can set an application scenario that is more in line with the current use environment according to the scenario of the robot application, and can better conform to the current environment in the process of human-computer interaction, so that the user is more integrated into the human-computer interaction, thereby improving The experience of human-computer interaction. Application scenarios include interaction rules and interaction resources. If the robot is used in a mall, it will increase the shopping-related interaction rules and interaction resources. For example, the interaction rule is to ask the user which products to purchase when the robot recognizes the user, or to ask the user mall which services are not enough, etc.; the interactive resources can set cheerful songs, or clapping hands, squatting actions, in the user When responding to purchased goods, you can make a clap action and encourage users to continue spending.
机器人若是应用在医院中,则会增加医疗、药品、防疫等相关的的交互规则以及交互资源。例如交互规则为机器人识别到用户伤心时,会询问用户为什么不开心,并安慰用户,同时做出拍拍用户肩膀的动作,并播放一些欢快的音乐;交互资源中可以设置一些欢快的音乐,拍肩,加油等动作。If the robot is used in a hospital, it will increase the interaction rules and interaction resources related to medical treatment, drugs, and epidemic prevention. For example, when the robot recognizes the user's sadness, the user will be asked why the user is not happy, and comfort the user, while making the action of patting the user's shoulder and playing some cheerful music; some cheerful music can be set in the interactive resource. Shoulder, refueling and other actions.
当机器人识别到用户后,会通过人脸识别,获取当前用户信息,并通过机器人内部系统获取机器人的环境信息,最后形成带有用户个性化的用户综合信息。然后机器人会根据应用场景中的交互规则,得到当前应该完成的交互是什么,调取交互资源中的相应的交互资源,形成主动交互内容,主动与用户进行交互。不同于当前交互方式的是,当前很多交互方式都是被动式的交互,机器人通过不断监听用户的语音指令,在接受到特定的语音指令之后,开始进行语音识别,根据识别的具体内容,为用户做出相应的回答与反馈。这样的人机交互方式比较被动,机器人不会主动与用户进行交流,对用户的吸引力不够强。而本发明中,机器人在检测到用户之后,能够通过获取到的用户综合信息形成带有用户个性化的主动交互内容,主动与用户进行交互,更加吸引用户参与到人机交互当中。When the robot recognizes the user, it will obtain the current user information through face recognition, and obtain the environment information of the robot through the internal system of the robot, and finally form the user comprehensive information with the user's personalization. Then the robot will get the current interactions according to the interaction rules in the application scenario, retrieve the corresponding interaction resources in the interaction resources, form active interaction content, and actively interact with the user. Different from the current interaction mode, many current interaction methods are passive interactions. The robot continuously listens to the user's voice commands, and after receiving a specific voice command, starts voice recognition, and performs for the user according to the specific content of the recognition. Give the corresponding answers and feedback. Such human-computer interaction is relatively passive, and the robot does not actively communicate with the user, which is not attractive to the user. In the present invention, after detecting the user, the robot can form the active interactive content with the user personalized through the obtained comprehensive information of the user, and actively interact with the user, thereby attracting the user to participate in the human-computer interaction.
例如,机器人如果在商场中使用,用户可以增加和购物相关的交互规则以及交互资源。交互规则是:机器人识别到用户后,首先识别用户,获得当前用户信息以及环境信息,若识别出来是成人,后会询问用户购买了哪些商品,做 出回应;其次询问用户商场还有哪些服务不够到位,做出回应;最后和用户说再见。若是小朋友,则会先同小朋友打招呼,再为小孩儿跳一支舞。For example, if the robot is used in a mall, the user can add shopping-related interaction rules and interaction resources. The interaction rule is: after the robot recognizes the user, first identifies the user, obtains the current user information and the environment information, and if it is an adult, it will ask the user which products to purchase and respond, and then ask the user mall which services are not enough. In place, respond; finally say goodbye to the user. If it is a child, it will first greet the children and dance for the children.
用户在交互资源中可以设置欢快的歌曲,或是拍手、鞠躬、挥手、蹲下等等的动作,在用户回应购买的商品时,可以做出拍手的动作,鼓励用户继续消费。The user can set a cheerful song in the interactive resource, or an action of clap, slap, wave, squat, etc., when the user responds to the purchased product, can make a clap action, and encourage the user to continue to consume.
当机器人识别到的用户是一位女士,当前时间为上午10点半,机器人会说:“女士上午好!请问您今天买了哪些商品呢?”同时做出挥手的动作向女士打招呼。女士回应:“买了一件衣服。”机器人会回应:“感谢对我们商场的支持!”同时做出鞠躬的动作。然后继续说:“商场还有哪些服务不够到位”然后将用户说的内容记录下来。最后和用户说“再见”。When the user identified by the robot is a lady, the current time is 10:30 am, the robot will say: "Mrs. Good morning! What products did you buy today?" At the same time, make a wave of gestures to say hello to the lady. The lady responded: “Buy a piece of clothing.” The robot will respond: “Thank you for your support of the mall!” At the same time, he made awkward moves. Then continue to say: "There are not enough services in the mall," and then record what the user said. Finally, say "goodbye" to the user.
当机器人识别到的用户是以为小女孩儿,当前的时间是下午5点,当前小女孩的表情为笑脸,机器人会说:“小朋友下午好!怎么这么开心啊?”同时做出蹲下的动作,与小女孩保持同一高度。小女孩回应:“今天妈妈给我买了件新衣服。”机器人会回应:“那我为您跳一支舞吧”同时播放舞蹈音乐。When the user recognized by the robot is a little girl, the current time is 5 pm, the current little girl's expression is a smiley face, the robot will say: "Good afternoon, children! How are you so happy?" At the same time, make a gesture of squatting, Keep the same height as the little girl. The little girl responded: "Mom bought me a new dress today." The robot will respond: "I will dance for you" and play dance music.
可以看出,由于用户体现出的用户综合信息不同,机器人的交互方式也有很大的不同,而现有技术中,机器人对每个用户的主动交互内容都是一样的,在识别到用户后只会说:“您好,请问有什么可以帮您?”而不同于本发明中能够根据不同用户携带的不同特征,输出个性化的主动交互内容。It can be seen that the interaction mode of the robot is also very different due to the different user comprehensive information embodied by the user. In the prior art, the active interaction content of the robot for each user is the same, after the user is identified. It will say: "Hello, what can I help you?" Unlike the present invention, it is possible to output personalized active interactive content according to different characteristics carried by different users.
上述的交互规则和交互资源都可以自行设定,机器人会根据用户综合信息,在交互规则中找到相应的回应,结合交互资源生成主动交互内容,进行主动交互。可以看到,不同用户的用户综合信息是不一样的,机器人会根据不同用户的不同性别,不同表情,不同年龄,生成不同的主动交互内容,让主动交互内容具有个性化。The above interaction rules and interaction resources can be set by themselves. The robot will find corresponding responses in the interaction rules according to the user comprehensive information, and generate active interaction content in combination with the interaction resources to perform active interaction. It can be seen that the comprehensive information of users of different users is different. The robot generates different active interactive content according to different genders, different expressions and different ages of different users, so that the active interactive content is personalized.
如图2所示,本发明还提供了一种人机交互的方法的一个实施例,包括:As shown in FIG. 2, the present invention also provides an embodiment of a method for human-computer interaction, including:
S0根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。S0 sets an application scenario of the robot that matches the environment information according to the environment information of the current robot, and the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
S1当机器人检测到需要主动交互的用户时,获取用户综合信息,所述用户综合信息包括当前用户的个人信息,以及当前机器人系统的环境信息;S1, when the robot detects a user who needs active interaction, acquiring user comprehensive information, where the user comprehensive information includes personal information of the current user, and environment information of the current robot system;
S201在所述交互规则中筛选出与所述用户综合信息相匹配的若干个候选 规则节点;S201: filtering, in the interaction rule, a plurality of candidate rule nodes that match the user comprehensive information;
S202根据所述若干个候选规则节点对应的交互资源,生成主动交互内容;所述主动交互内容包括语音交互内容、动作交互内容、多媒体交互内容。S202: Generate active interactive content according to the interaction resources corresponding to the plurality of candidate rule nodes; the active interaction content includes voice interaction content, action interaction content, and multimedia interaction content.
S3根据所述主动交互内容与用户进行主动交互。S3 actively interacts with the user according to the active interaction content.
所述交互资源包括:语音内容、动作内容、或者多媒体内容;所述交互规则包括多个规则节点,每个规则节点体现了不同用户综合信息与不同交互资源的映射关系。The interaction resource includes: voice content, action content, or multimedia content; the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
所述交互规则包括:机器人在交互时,对某些特定的检测结果是否响应,例如是否询问用户姓名,是否检测系统时间等等;以及在对响应的检测结果时,选择相应的反馈的具体内容,例如,对不同性别的不同称呼,不同用户的不同动作等等。The interaction rule includes: whether the robot responds to certain specific detection results when interacting, for example, whether to query the user name, whether to detect the system time, etc.; and when the detection result of the response is selected, the specific content of the corresponding feedback is selected. For example, different names for different genders, different actions for different users, and so on.
所述交互资源包括:对应的交互规则下所需的资源,例如所有的语音文本内容、所有可选的动作交互内容、所有音乐以及视频内容等等。The interaction resources include: resources required under corresponding interaction rules, such as all voice text content, all optional action interaction content, all music and video content, and the like.
所述当前用户的个人信息包括:性别、年纪、表情、人脸角度、人脸空间位置、人脸出现次数、用户名称、当前被检测到人脸数量,以及语音信息;所述当前机器人系统的环境信息包括:时间、地点、温度、天气、网络连接状态、系统语言。The current user's personal information includes: gender, age, expression, face angle, face space location, face appearance times, user name, current detected face number, and voice information; the current robot system Environmental information includes: time, location, temperature, weather, network connection status, system language.
具体的,本实施例中,系统通过人体识别,人脸识别,环境系统检测等各种识别与检测,为用户综合输出一组用户信息描述,用来描述当前的交互用户的具体信息。用户综合信息形式表示如下:Specifically, in this embodiment, the system performs various identification and detection by human body recognition, face recognition, environment system detection, etc., and comprehensively outputs a set of user information descriptions for the user to describe the specific information of the current interactive user. The user's comprehensive information form is expressed as follows:
head;key1:value1;key2:value2;key3:value3;...;Head;key1:value1;key2:value2;key3:value3;...;
其中head是固定的,为该字串的字串标识符,标识了该字串的具体内容,以便与其他类型的字串做出区别。key是用户综合信息中的各个特征,即用户特征关键词及环境特征关键词,每一个key表示了用来描述当前用户的一个特征,这些特征可以包括:人脸编号、人脸姓名、性别、年纪、时间、表情、人脸角度、人脸位置、人脸大小、天气、地点、温度、运动类型、网络连接状况等等。value值是与当前key值相对应的具体参数值,即本发明中所述的用户特征参数值及环境特征参数值。<key:value>对可以根据不同的人体识别、人脸识别、系统检测、运动检测等各种检测与识别工具的输出变化,修改用户描述特 征的个数与内容。Where head is fixed, the string identifier of the string identifies the specific content of the string to distinguish it from other types of strings. The key is a feature in the user comprehensive information, that is, a user feature keyword and an environmental feature keyword, and each key represents a feature used to describe the current user, and the features may include: face number, face name, gender, Age, time, expression, face angle, face position, face size, weather, location, temperature, type of exercise, network connection, etc. The value value is a specific parameter value corresponding to the current key value, that is, the user characteristic parameter value and the environmental characteristic parameter value described in the present invention. The <key:value> pair can modify the number and content of the user description features according to different human body recognition, face recognition, system detection, motion detection and other detection and recognition tool output changes.
简单的用户综合信息如下:Simple user comprehensive information is as follows:
例子1:Example 1:
rvn;ultrasound;event;payload;type:passby;Rvn;ultrasound;event;payload;type:passby;
其中rvn;ultrasound;event;payload;作为字串头,表示了该字串是包含了超声波传感器的用户信息描述。该条描述较为简单,仅仅表示了机器人通过超声波传感器感知到了有人过从其面前走过。Where rvn;ultrasound;event;payload; as a string header, indicating that the string is a description of the user information containing the ultrasonic sensor. The description of the article is relatively simple and only indicates that the robot has sensed that someone has passed through it through the ultrasonic sensor.
例子2:Example 2:
rvf;vison;event;face;payload;type:face;stop:yes;name:avatar;gender:masculine;age:53;time:13.04;emotion:none;roll:-13.61209;pitch:23.196611;yaw:23.330135;faceID:1;number:1;sequence:1;px:646;py:189;pw:352;ph:352;Rvf;vison;event;face;payload;type:face;stop:yes;name:avatar;gender:masculine;age:53;time:13.04;emotion:none;roll:-13.61209;pitch:23.196611;yaw:23.330135 ;faceID:1;number:1;sequence:1;px:646;py:189;pw:352;ph:352;
其中rvf;vison;event;face;payload;作为字串头,表示了该字串是包含了视觉传感器信息的用户综合信息描述。每一对的key,value表示了这一用户的一个信息特征。具体可以解读为:用户的人脸信息是连续的;用户姓名:avatar;性别:男性;年纪:53岁;本条记录产生时间:13点04分;用户人脸表情:无;人脸角度roll值:-13.61209度;人脸角度pitch值:23.196611度;人脸角度yaw值:23.330135度;人脸记录编号:1号;当前画面中人脸个数;1个;该用户的人脸在总人脸中为第一个;人脸位置X值:646px;人脸位置Y值:189px;人脸宽度:352px;人脸长度:352px;Where rvf;vison;event;face;payload; as a string header, indicating that the string is a comprehensive information description of the user containing the visual sensor information. The key, value of each pair represents an information feature of the user. The specific can be interpreted as: the user's face information is continuous; user name: avatar; gender: male; age: 53 years old; record generation time: 13:04; user face expression: no; face angle roll value :-13.61209 degrees; face angle pitch value: 23.196611 degrees; face angle yaw value: 23.330135 degrees; face record number: 1; the number of faces in the current picture; 1; the user's face is in the total person The face is the first one; the face position X value: 646px; the face position Y value: 189px; the face width: 352px; the face length: 352px;
不同的用户,不同的环境,不同的交互熟悉程度,不同的检测与识别工具,会生成不同的用户描述结果。这些用户的综合描述,是个性化的用户描述,通过对这些不同描述的解析,系统通过当前场景的规则与资源,生成与之相对应的主动交互内容。Different users, different environments, different levels of familiarity, and different detection and identification tools generate different user description results. The comprehensive description of these users is a personalized user description. Through the analysis of these different descriptions, the system generates active interactive content corresponding to the rules and resources of the current scene.
本方法根据输入的用户综合信息,生成一组与之相对应的主动交互内容。反馈的主动交互内容包含了以下三种:语音交互内容、动作交互内容与多媒体交互内容。语音交互内容为机器人播放的主动语音提示;动作交互内容为其头部、四肢等活动部位的一组运动内容;多媒体交互内容包括了图片、音乐、视频、应用等等,并通过机器人胸前的展示平台予以播放。多媒体的内容可以随语音提示同时播放,也可以在语音提示结束后播放,以满足不同场景的需求。The method generates a set of active interactive content corresponding to the input user comprehensive information. The active interactive content of the feedback includes the following three types: voice interactive content, action interactive content, and multimedia interactive content. The voice interactive content is an active voice prompt played by the robot; the action interactive content is a set of sports content of the active part of the head, the limbs, etc.; the multimedia interactive content includes pictures, music, video, application, etc., and passes through the chest of the robot. The display platform is played. The content of the multimedia can be played simultaneously with the voice prompt, or can be played after the voice prompt ends, to meet the needs of different scenes.
所述规则节点包括Node节点,所述交互规则以树形的数据结构存储了每条识别结果所对应的语音、动作、多媒体交互内容。交互规则树中包含了多个Node节点,每一个Node节点中包含了预先设置的若干个预设特征关键词及对应的预设特征参数值,还包含了多条语音、动作、多媒体等交互资源。The rule node includes a Node node, and the interaction rule stores the voice, action, and multimedia interaction content corresponding to each recognition result in a tree-shaped data structure. The interaction rule tree includes a plurality of Node nodes, each of which includes a plurality of preset feature keywords and corresponding preset feature parameter values, and also includes multiple interaction resources such as voice, action, and multimedia. .
Node节点中的Key值描述了该组语句、动作、多媒体需要被选中的必要条件。首先,当存在一条用户的综合信息时,每一组Node节点都会与当前输入的用户综合信息进行匹配,如果当前的用户综合信息与Node节点的必要条件相满足,该Node节点将会成为一个候选节点,待之后选择。如果当前的用户综合信息与Node节点的必要条件不完全满足,则该Node节点便不会成为候选规则节点。The Key value in the Node node describes the necessary conditions for the set of statements, actions, and multimedia to be selected. First, when there is a comprehensive information of a user, each group of Node nodes will match the currently integrated user information. If the current user comprehensive information is satisfied with the necessary conditions of the Node node, the Node node will become a candidate. Node, wait for later selection. If the current user comprehensive information and the necessary conditions of the Node node are not completely satisfied, the Node node will not become a candidate rule node.
如图3所示,本发明还提供了一种人机交互的方法的一个实施例,包括:As shown in FIG. 3, the present invention further provides an embodiment of a method for human-computer interaction, including:
S0根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。S0 sets an application scenario of the robot that matches the environment information according to the environment information of the current robot, and the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
S10当机器人检测到需要主动交互的用户时,获取用户综合信息,并给所述当前用户的个人信息中的每个用户特征赋予对应的用户特征关键词和用户特征参数值,给所述当前机器人系统的环境信息中的每个环境特征赋予对应的环境特征关键词和环境特征参数值。S10: when the robot detects a user who needs active interaction, acquires user comprehensive information, and assigns a corresponding user feature keyword and a user feature parameter value to each user feature in the current user's personal information, and gives the current robot Each environmental feature in the environmental information of the system is assigned a corresponding environmental feature keyword and an environmental feature parameter value.
S2011判断每个规则节点是否满足预设条件;S2011 determines whether each rule node meets a preset condition;
S2012若是,则将满足条件的规则节点作为候选规则节点;If S2012 is, the rule node that satisfies the condition is used as the candidate rule node;
S2021分析所述若干个候选规则节点各自的优先值,并将所述候选规则节点按照所述优先值进行排序;对于同一优先值的多个候选规则节点,随机或加权随机选取其中一个候选规则节点参与排序;S2021 analyzes a priority value of each of the candidate rule nodes, and sorts the candidate rule nodes according to the priority value; and randomly selects one candidate rule node randomly or weighted for multiple candidate rule nodes of the same priority value. Participate in sorting;
S2022将排序后的候选规则节点对应的交互资源依次进行组合,生成所述主动交互内容。S2022 combines the interaction resources corresponding to the sorted candidate rule nodes in sequence to generate the active interaction content.
S3根据所述主动交互内容与用户进行主动交互。S3 actively interacts with the user according to the active interaction content.
具体的,本实施例中,预设条件为判断每个规则节点中所有的预设特征关键词及对应的预设特征参数值是否与所述用户综合信息中的部分特征关键词及对应的特征参数值相同。所述特征关键词包括用户特征关键词和环境特征关键词,所述特征参数值包括用户特征参数值和环境特征参数值。假设特征关键 词为key值,特征参数值为key值对应的Value值,规则节点为Node节点,即根据预设条件筛选出与所述用户综合信息相匹配的若干个候选规则节点的具体步骤:如果Node节点中的所有特征中的Value值与用户综合信息中相同的特性的Value值都满足,则将此Node节点作为候选规则节点。如果Node节点中的特征的值为All,即表示,该特征对所有的用户信息中相应特征的Value结果都视为满足。用户的综合信息中的特征通常都会多于Node节点所需的特征值,对于多出的特征值,系统不会根据其结果进行判断与筛选。Specifically, in this embodiment, the preset condition is determining whether all preset feature keywords and corresponding preset feature parameter values in each rule node are related to partial feature keywords and corresponding features in the user comprehensive information. The parameter values are the same. The feature keyword includes a user feature keyword and an environment feature keyword, and the feature parameter value includes a user feature parameter value and an environment feature parameter value. It is assumed that the feature keyword is a key value, the feature parameter value is a value corresponding to the key value, and the rule node is a Node node, that is, a specific step of filtering out a plurality of candidate rule nodes that match the user comprehensive information according to a preset condition: If the Value value of all the features in the Node node is equal to the Value value of the same feature in the user comprehensive information, the Node node is used as the candidate rule node. If the value of the feature in the Node node is All, it means that the feature is considered to be satisfied for the Value result of the corresponding feature in all user information. The features in the user's comprehensive information are usually more than the feature values required by the Node node. For the extra feature values, the system does not judge and filter based on the results.
在匹配候选Node节点的同时,对于成为候选节点的Node节点,将会根据其优先值Priority的数值组合出一句完整的语音提示。本方法将一句完整的语音交互内容分解为不同的语句段,每一语句段为一句完整语音提示中的一个分段。每一个Node节点中的优先值Priority,表示了该语句段在完整语句中的位置。While matching the candidate Node nodes, for a Node node that becomes a candidate node, a complete voice prompt will be combined according to the value of its priority value Priority. The method decomposes a complete speech interaction content into different sentence segments, and each sentence segment is a segment of a complete voice prompt. The priority value Priority in each Node node indicates the position of the statement segment in the complete statement.
简单的,我们可以将一句完整的语句分成多个语句段,例如称呼段,时间段,内容段和问题段。本发明并不限制语句的分段数目,使用者可以根据语句完整度自行分段。每一段与下一段之间为自由组合。因此,最后组合出的完整语句将会变得十分灵活。对于处于同一个位置的多个候选节点,本方法通过随机选择,选择一个满足条件的Node节点。Simply, we can divide a complete statement into multiple statement segments, such as a title segment, a time segment, a content segment, and a problem segment. The invention does not limit the number of segments of the statement, and the user can segment the segment according to the completeness of the statement. There is a free combination between each paragraph and the next paragraph. Therefore, the final combination of the complete statement will become very flexible. For multiple candidate nodes in the same location, the method selects a Node node that satisfies the condition by random selection.
例如,即将生成的语音交互内容包括称呼段、问候段和内容段,称呼段有两个优先值相同的Node节点,分别是“老人家”、“老伯”;问候段有三个优先值相同的Node节点,分别是“您好!”“上午好!”“早上好!”内容段有一个Node节点,是“您身体真好。”对于称呼段和问候段,可以随机选取其中一个内容参与语音信息的排序。可以看到,语音信息的内容有6种组合方式,在相同条件下,机器人进行语音交互的内容会不断变化,不会非常死板,影响用户的体验。For example, the voice interaction content to be generated includes a calling segment, a greeting segment, and a content segment. The calling segment has two Node nodes with the same priority value, namely “old man” and “old man”; the greeting segment has three Node nodes with the same priority value. "Hello!" "Good morning!" "Good morning!" The content segment has a Node node, which is "You are in good health." For the title segment and the greeting segment, you can randomly select one of the content to participate in the voice message. Sort. It can be seen that there are six combinations of voice message content. Under the same conditions, the content of the robot's voice interaction will constantly change, and it will not be very rigid and affect the user experience.
在一些Node节点中,可以预先设置多个Item选项(即机器人执行的内容),对于Node节点中存在的多个Item选项,根据其Key值的不同,选择满足条件的最终结果。如果相同的Key值结果存在对应的多个Item选项,本方法通过顺序选择或是随机选择的方法,挑选一个结果作为最终结果,予以输出。In some Node nodes, a plurality of Item options (ie, content executed by the robot) may be set in advance, and for a plurality of Item options existing in the Node node, a final result that satisfies the condition is selected according to the difference of the Key values thereof. If the same Key value result has multiple corresponding Item options, the method selects a result as a final result by sequential selection or random selection, and outputs the result.
由于交互的内容要取决于候选规则节点的组合方式,以及每个候选规则节 点中Item选项的选取,在同一场景下,同一优先值的候选规则节点只随机选取一个参与排序,对于每个候选规则节点,其key值对应的多个Item选项可以随机、加权随机,或按照一定顺序选取一个,则组合的方式非常多。在交互过程中,对于同一种场景,交互的内容也有所不同,不会非常死板地进行固定模式的交互。Since the content of the interaction depends on the combination of the candidate rule nodes and the selection of the Item option in each candidate rule node, in the same scenario, the candidate rule nodes of the same priority value only randomly select one participation order, for each candidate rule. A node, whose multiple item options corresponding to the key value can be randomly, weighted randomly, or selected in a certain order, the combination is very much. In the interaction process, the content of the interaction is different for the same scenario, and the fixed mode interaction is not very rigid.
例如在打招呼时,可能key值对应的动作为握手、挥手、敬礼三个Item选项,在这种情况下,可以任选其中一个Item进行输出。在打招呼时,不会由于动作太过单一导致机器人的交互太过死板。同样,打招呼的语言也可以设置为多种,赋到不同的Item选项中。在随机选取一个Item时,输出的语言也会不同,交互时内容会呈现多样化。For example, when greeting, the action corresponding to the key value may be three item options: handshake, wave, and salute. In this case, one of the items may be selected for output. When you say hello, the interaction of the robot is not too rigid because the action is too single. Similarly, the greeting language can be set to multiple, assigned to different Item options. When an item is randomly selected, the language of the output will be different, and the content will be diversified when interacting.
动作交互内容与多媒体交互内容通过附加在每一个具体的语音语句之后所存在。通过定义每一个具体语句之后的附加动作与多媒体交互内容,在主动反馈内容组合完成时,其相应的动作与多媒体交互内容也随之生成。动作与多媒体交互内容由最后一个语句成分的附加内容所决定。因此,不同内容,不同长短的语句也可以做出不同的动作与多媒体交互内容上的提示。灵活使用语音脚本与资源脚本,随着用户综合信息输入的不同,本方法与之生成的反馈信息也可以与之相应的变化起来。The action interaction content and the multimedia interaction content exist by being appended to each specific voice sentence. By defining additional actions and multimedia interactions after each specific statement, when the active feedback content combination is completed, the corresponding actions and multimedia interaction content are also generated. The action and multimedia interaction content is determined by the additional content of the last statement component. Therefore, different content, different lengths of statements can also make different actions and prompts on multimedia interactive content. Flexible use of voice scripts and resource scripts, as the user's comprehensive information input, the feedback information generated by this method can also be changed accordingly.
通过人脸识别方法,结合互联网中的个人信息,可以获取到用户的某些个人信息,如姓名,年龄,性别等基础信息。本实施例将举一个例子说明本发明交互系统。Through the face recognition method, combined with the personal information in the Internet, some personal information of the user, such as basic information such as name, age, and gender, can be obtained. This embodiment will illustrate an interactive system of the present invention by way of an example.
例如当机器人识别到前方有一个正在哭泣的小男孩儿时,采集的用户综合信息可以包括:男性,8岁,姓名,面部表情哭泣,当前时间为上午10点,当前天气为2摄氏度,当前地点为医院,空气质量良好,并将其赋值之后形成字串,供系统调取。机器人根据用户综合信息,在规则节点中查找出满足匹配条件的候选规则节点,将这些规则节点按照优先值进行排序,并将候选规则节点对应的交互资源依次进行组合。机器人识别到小男孩儿之后,会主动和小男孩儿打招呼,在打招呼时,其Node节点中key值对应的Item值可能会有很多个(其语音交互内容分别为“小朋友你好!”、“你好,小帅哥!”“hi,小朋友”,其动作交互内容分别为挥手、握手、鞠躬、敬礼)机器人在只需在语音 交互内容或者动作交互内容中人选其一作为主动交互内容即可,因此,每次输出的主动交互内容都是不断变化的,不会太过死板。For example, when the robot recognizes that there is a crying boy in front, the collected user comprehensive information may include: male, 8 years old, name, facial expression crying, the current time is 10 am, the current weather is 2 degrees Celsius, the current location is In the hospital, the air quality is good, and after the value is assigned, a string is formed for the system to retrieve. The robot finds the candidate rule nodes that satisfy the matching condition in the rule node according to the user comprehensive information, sorts the rule nodes according to the priority values, and sequentially combines the interaction resources corresponding to the candidate rule nodes. After the robot recognizes the little boy, it will greet the little boy actively. When you say hello, there may be a lot of Item values corresponding to the key value in the Node node (the voice interaction content is “Hello Kids!”, “Hello. "Little, handsome boy!" "hi, children", the action interaction content is wave, handshake, swear, salute) the robot can only choose one of the interactive content in the voice interactive content or action interaction content, therefore, The active interactions of each output are constantly changing and not too rigid.
如图4所示,本发明提供了一种交互机器人的一个实施例,包括:As shown in FIG. 4, the present invention provides an embodiment of an interactive robot, including:
信息获取模块2,用于获取当前用户信息,以及环境信息;The information obtaining module 2 is configured to acquire current user information and environment information;
用户综合信息模块3,与所述信息获取模块2电连接,用于根据所述当前用户信息以及环境信息生成用户综合信息;The user integrated information module 3 is electrically connected to the information acquiring module 2, and is configured to generate user comprehensive information according to the current user information and environment information;
处理模块4,与所述用户综合信息模块3电连接,用于根据所述用户综合信息以及所述应用场景,生成主动交互内容;The processing module 4 is electrically connected to the user integrated information module 3, and configured to generate active interactive content according to the user comprehensive information and the application scenario;
交互模块5,用于根据所述主动交互内容进行主动人机交互。The interaction module 5 is configured to perform active human-computer interaction according to the active interaction content.
优选的,所述交互机器人还包括:场景设置模块1,还用于根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。Preferably, the interactive robot further includes: a scene setting module 1 configured to set an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and The interactive resources used in the interaction.
优选的,所述场景设置模块1,还用于储存预先设置的机器人的交互规则,以及机器人在所述交互规则下交互时所需的交互资源,并将所述交互规则和所述应用资源作为应用场景;所述交互规则包括多个规则节点,每个规则节点包括多个预设特征关键词及对应的预设特征参数值。Preferably, the scene setting module 1 is further configured to store a preset interaction rule of the robot, and an interaction resource required when the robot interacts under the interaction rule, and use the interaction rule and the application resource as An application scenario; the interaction rule includes multiple rule nodes, and each rule node includes multiple preset feature keywords and corresponding preset feature parameter values.
优选的,用户综合信息模块3,还用于给所述当前用户信息中的每个用户特征赋予对应的用户特征关键词和用户特征参数值,给所述环境信息中的每个环境特征赋予对应的环境特征关键词和环境特征参数值;用户综合信息模块3,还用于将所述用户特征关键词及对应的用户特征参数值、环境特征关键词及对应的环境特征参数值组合成字串,并给所述字串赋予相应的字串标识符;将带有所述字串标识符的字串作为当前用户综合信息。Preferably, the user integrated information module 3 is further configured to assign a corresponding user feature keyword and a user feature parameter value to each user feature in the current user information, and assign a corresponding to each environment feature in the environment information. The environment feature keyword and the environment feature parameter value; the user integrated information module 3 is further configured to combine the user feature keyword and the corresponding user feature parameter value, the environment feature keyword, and the corresponding environment feature parameter value into a string And assigning the string to the corresponding string identifier; using the string with the string identifier as the current user comprehensive information.
优选的,所述处理模块4具体包括:匹配子模块41,用于在所述应用场景中筛选出与所述用户综合信息相匹配的若干个候选规则节点;主动交互内容生成子模块42,根据所述若干个候选规则节点对应的交互资源,生成主动交互内容;所述主动交互内容包括语音交互内容、动作交互内容、多媒体交互内容。Preferably, the processing module 4 specifically includes: a matching sub-module 41, configured to filter, in the application scenario, a plurality of candidate rule nodes that match the user comprehensive information; and an active interaction content generation sub-module 42 according to The interaction resources corresponding to the plurality of candidate rule nodes generate active interaction content; the active interaction content includes voice interaction content, action interaction content, and multimedia interaction content.
优选的,所述匹配子模块41,还用于逐一判断每个规则节点中所有的预设特征关键词及对应的预设特征参数值是否与所述用户综合信息中的部分特征关键词及对应的特征参数值相同;所述特征关键词包括用户特征关键词和环境 特征关键词,所述特征参数值包括用户特征参数值和环境特征参数值;若是,则将满足条件的规则节点作为候选规则节点。Preferably, the matching sub-module 41 is further configured to determine, by one by one, whether all preset feature keywords and corresponding preset feature parameter values in each rule node are corresponding to some feature keywords in the user comprehensive information. The feature parameter values are the same; the feature keyword includes a user feature keyword and an environment feature keyword, and the feature parameter value includes a user feature parameter value and an environment feature parameter value; if yes, the rule node that satisfies the condition is used as a candidate rule node.
优选的,所述主动交互内容生成子模块42,还用于分析所述若干个候选规则节点各自的优先值,并将所述候选规则节点按照所述优先值进行排序;对于同一优先值的多个候选规则节点,选取其中一个候选规则节点参与排序;将排序后的候选规则节点对应的交互资源依次进行组合,生成所述主动交互内容。Preferably, the active interaction content generation sub-module 42 is further configured to analyze respective priority values of the plurality of candidate rule nodes, and sort the candidate rule nodes according to the priority value; for the same priority value The candidate rule nodes select one of the candidate rule nodes to participate in the sorting; and the interactive resources corresponding to the sorted candidate rule nodes are sequentially combined to generate the active interactive content.
具体的,信息获取模块2有两大作用,其一是用来识别当前用户,获取用户资源;例如当检测到在预设范围内有用户时,开始对用户进行识别,其主要识别用户的面部特征。通过人脸识别技术,可以识别出当前用户的表情,并结合互联网大数据,获取到用户的一些基本数据信息。其二是根据当前机器人系统,获取到当前机器人所在地点、时间、天气等等环境信息。Specifically, the information obtaining module 2 has two major functions, one is to identify the current user, and obtain the user resource; for example, when it is detected that there is a user within the preset range, the user is initially identified, and the user mainly recognizes the user's face. feature. Through the face recognition technology, the current user's expression can be identified, and combined with the Internet big data, some basic data information of the user is obtained. The second is to obtain environmental information such as the current location, time, weather, etc. of the current robot according to the current robot system.
所述处理模块4可以由机器人的处理器组成,所述交互模块5包括了机器人交互过程中所用到的声控系统、显示系统、驱动系统等等。机器人交互的过程可参照上述方法实施例,此处不再赘述。The processing module 4 may be composed of a processor of a robot, and the interaction module 5 includes a voice control system, a display system, a drive system, and the like used in the robot interaction process. For the process of the robot interaction, reference may be made to the foregoing method embodiment, and details are not described herein again.
应当说明的是,上述实施例均可根据需要自由组合。以上所述仅是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。It should be noted that the above embodiments can be freely combined as needed. The above description is only a preferred embodiment of the present invention, and it should be noted that those skilled in the art can also make several improvements and retouchings without departing from the principles of the present invention. It should be considered as the scope of protection of the present invention.

Claims (10)

  1. 一种人机交互的方法,其特征在于,包括步骤:A method for human-computer interaction, comprising the steps of:
    S1当机器人检测到需要主动交互的用户时,获取用户综合信息,所述用户综合信息包括所述用户的个人信息,以及当前机器人系统的环境信息;S1, when the robot detects a user who needs active interaction, acquiring user comprehensive information, where the user comprehensive information includes personal information of the user, and environment information of the current robot system;
    S2生成与所述用户综合信息相匹配的主动交互内容;S2 generates active interactive content that matches the user comprehensive information;
    S3根据所述主动交互内容与用户进行主动交互。S3 actively interacts with the user according to the active interaction content.
  2. 根据权利要求1所述的一种人机交互的方法,其特征在于,所述步骤S1之前包括步骤:The method of human-computer interaction according to claim 1, wherein the step S1 comprises the steps of:
    S0根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。S0 sets an application scenario of the robot that matches the environment information according to the environment information of the current robot, and the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
  3. 根据权利要求2所述的一种人机交互的方法,其特征在于,步骤S2具体包括:The method of human-computer interaction according to claim 2, wherein the step S2 specifically includes:
    S20根据所述交互规则,获取与所述用户综合信息相匹配的交互资源,从而生成主动交互内容。S20 acquires, according to the interaction rule, an interaction resource that matches the user comprehensive information, so as to generate active interaction content.
  4. 根据权利要求3所述的一种人机交互的方法,其特征在于:A method of human-computer interaction according to claim 3, wherein:
    所述交互资源包括:语音内容、动作内容、或者多媒体内容;所述交互规则包括多个规则节点,每个规则节点体现了不同用户综合信息与不同交互资源的映射关系。The interaction resource includes: voice content, action content, or multimedia content; the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
  5. 根据权利要求1-4中任一项所述的一种人机交互的方法,其特征在于:A method for human-computer interaction according to any one of claims 1 to 4, characterized in that:
    所述用户的个人信息包括:性别、年纪、表情、人脸角度、人脸空间位置、人脸出现次数、用户名称、当前被检测到人脸数量,以及语音信息;The personal information of the user includes: gender, age, expression, face angle, face space location, number of face appearances, user name, number of faces currently detected, and voice information;
    所述当前机器人系统的环境信息包括:时间、地点、温度、天气、网络连接状态、系统语言。The environmental information of the current robot system includes: time, location, temperature, weather, network connection status, system language.
  6. 一种交互机器人,其特征在于,包括:An interactive robot characterized by comprising:
    信息获取模块,用于当机器人检测到需要主动交互的用户时,获取用户综合信息,所述用户综合信息包括所述用户的个人信息,以及当前机器人系统的环境信息;An information obtaining module, configured to acquire user comprehensive information when the robot detects a user who needs active interaction, where the user comprehensive information includes personal information of the user, and environment information of the current robot system;
    处理模块,与所述信息获取模块电连接,用于生成与所述用户综合信息相匹配的主动交互内容;The processing module is electrically connected to the information acquiring module, and configured to generate active interactive content that matches the user comprehensive information;
    交互模块,用于根据所述主动交互内容与用户进行主动交互。The interaction module is configured to actively interact with the user according to the active interaction content.
  7. 根据权利要求6所述的一种交互机器人,其特征在于,还包括:The interactive robot according to claim 6, further comprising:
    场景设置模块,还用于根据当前机器人所处环境信息,设置与所述环境信息匹配的机器人的应用场景,所述应用场景包括机器人交互的交互规则和交互时用到的交互资源。The scenario setting module is further configured to set an application scenario of the robot that matches the environment information according to the environment information of the current robot, where the application scenario includes an interaction rule of the robot interaction and an interaction resource used in the interaction.
  8. 根据权利要求7所述的一种交互机器人,其特征在于:An interactive robot according to claim 7, wherein:
    所述处理模块,还用于根据所述交互规则,获取与所述用户综合信息相匹配的交互资源,从而生成主动交互内容。The processing module is further configured to acquire, according to the interaction rule, an interaction resource that matches the user comprehensive information, thereby generating active interaction content.
  9. 根据权利要求8所述的一种交互机器人,其特征在于:An interactive robot according to claim 8 wherein:
    所述交互资源包括:语音内容、动作内容、或者多媒体内容;所述交互规则包括多个规则节点,每个规则节点体现了不同用户综合信息与不同交互资源的映射关系。The interaction resource includes: voice content, action content, or multimedia content; the interaction rule includes multiple rule nodes, and each rule node embodies a mapping relationship between different user comprehensive information and different interaction resources.
  10. 根据权利要求6-9中任一项所述的一种交互机器人,其特征在于:An interactive robot according to any one of claims 6-9, characterized in that:
    所述当前用户的个人信息包括:性别、年纪、表情、人脸角度、人脸空间位置、人脸出现次数、用户名称、当前被检测到人脸数量,以及语音信息;The personal information of the current user includes: gender, age, expression, face angle, face space location, number of face appearances, user name, number of faces currently detected, and voice information;
    所述当前机器人系统的环境信息包括:时间、地点、温度、天气、网络连接状态、系统语言。The environmental information of the current robot system includes: time, location, temperature, weather, network connection status, system language.
PCT/CN2018/106780 2018-03-09 2018-09-20 Human-computer interaction method, and interactive robot WO2019169854A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810193098.5A CN108363492B (en) 2018-03-09 2018-03-09 Man-machine interaction method and interaction robot
CN201810193098.5 2018-03-09

Publications (1)

Publication Number Publication Date
WO2019169854A1 true WO2019169854A1 (en) 2019-09-12

Family

ID=63003702

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/106780 WO2019169854A1 (en) 2018-03-09 2018-09-20 Human-computer interaction method, and interactive robot

Country Status (2)

Country Link
CN (1) CN108363492B (en)
WO (1) WO2019169854A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111993438A (en) * 2020-08-26 2020-11-27 陕西工业职业技术学院 Intelligent robot
CN112527095A (en) * 2019-09-18 2021-03-19 奇酷互联网络科技(深圳)有限公司 Man-machine interaction method, electronic device and computer storage medium
CN114385000A (en) * 2021-11-30 2022-04-22 达闼机器人有限公司 Intelligent equipment control method, device, server and storage medium

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108363492B (en) * 2018-03-09 2021-06-25 南京阿凡达机器人科技有限公司 Man-machine interaction method and interaction robot
CN109409063A (en) * 2018-10-10 2019-03-01 北京小鱼在家科技有限公司 A kind of information interacting method, device, computer equipment and storage medium
CN109492074A (en) * 2018-11-22 2019-03-19 广州小鹏汽车科技有限公司 Intelligent greeting method, system, storage medium and automobile based on Weather information
CN110154048B (en) * 2019-02-21 2020-12-18 北京格元智博科技有限公司 Robot control method and device and robot
CN110097400A (en) * 2019-04-29 2019-08-06 贵州小爱机器人科技有限公司 Information recommendation method, apparatus and system, storage medium, intelligent interaction device
CN111949773A (en) * 2019-05-17 2020-11-17 华为技术有限公司 Reading equipment, server and data processing method
CN110716634A (en) * 2019-08-28 2020-01-21 北京市商汤科技开发有限公司 Interaction method, device, equipment and display equipment
CN111176503A (en) * 2019-12-16 2020-05-19 珠海格力电器股份有限公司 Interactive system setting method and device and storage medium
CN111327772B (en) * 2020-02-25 2021-09-17 广州腾讯科技有限公司 Method, device, equipment and storage medium for automatic voice response processing
CN111428637A (en) * 2020-03-24 2020-07-17 新石器慧通(北京)科技有限公司 Method for actively initiating human-computer interaction by unmanned vehicle and unmanned vehicle
CN113147771A (en) * 2021-05-10 2021-07-23 前海七剑科技(深圳)有限公司 Active interaction method and device based on vehicle-mounted virtual robot

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105701211A (en) * 2016-01-13 2016-06-22 北京光年无限科技有限公司 Question-answering system-oriented active interaction data processing method and system
CN106297789A (en) * 2016-08-19 2017-01-04 北京光年无限科技有限公司 The personalized interaction method of intelligent robot and interactive system
CN106774845A (en) * 2016-11-24 2017-05-31 北京智能管家科技有限公司 A kind of intelligent interactive method, device and terminal device
CN107045587A (en) * 2016-12-30 2017-08-15 北京光年无限科技有限公司 A kind of interaction output intent and robot for robot
CN108363492A (en) * 2018-03-09 2018-08-03 南京阿凡达机器人科技有限公司 A kind of man-machine interaction method and interactive robot

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105598972B (en) * 2016-02-04 2017-08-08 北京光年无限科技有限公司 A kind of robot system and exchange method
CN106537294A (en) * 2016-06-29 2017-03-22 深圳狗尾草智能科技有限公司 Method, system and robot for generating interactive content of robot
WO2018000258A1 (en) * 2016-06-29 2018-01-04 深圳狗尾草智能科技有限公司 Method and system for generating robot interaction content, and robot
CN106462254A (en) * 2016-06-29 2017-02-22 深圳狗尾草智能科技有限公司 Robot interaction content generation method, system and robot
CN106537293A (en) * 2016-06-29 2017-03-22 深圳狗尾草智能科技有限公司 Method and system for generating robot interactive content, and robot
WO2018000267A1 (en) * 2016-06-29 2018-01-04 深圳狗尾草智能科技有限公司 Method for generating robot interaction content, system, and robot
CN106843463B (en) * 2016-12-16 2020-07-28 北京光年无限科技有限公司 Interactive output method for robot
CN106625711A (en) * 2016-12-30 2017-05-10 华南智能机器人创新研究院 Method for positioning intelligent interaction of robot

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105701211A (en) * 2016-01-13 2016-06-22 北京光年无限科技有限公司 Question-answering system-oriented active interaction data processing method and system
CN106297789A (en) * 2016-08-19 2017-01-04 北京光年无限科技有限公司 The personalized interaction method of intelligent robot and interactive system
CN106774845A (en) * 2016-11-24 2017-05-31 北京智能管家科技有限公司 A kind of intelligent interactive method, device and terminal device
CN107045587A (en) * 2016-12-30 2017-08-15 北京光年无限科技有限公司 A kind of interaction output intent and robot for robot
CN108363492A (en) * 2018-03-09 2018-08-03 南京阿凡达机器人科技有限公司 A kind of man-machine interaction method and interactive robot

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112527095A (en) * 2019-09-18 2021-03-19 奇酷互联网络科技(深圳)有限公司 Man-machine interaction method, electronic device and computer storage medium
CN111993438A (en) * 2020-08-26 2020-11-27 陕西工业职业技术学院 Intelligent robot
CN114385000A (en) * 2021-11-30 2022-04-22 达闼机器人有限公司 Intelligent equipment control method, device, server and storage medium

Also Published As

Publication number Publication date
CN108363492B (en) 2021-06-25
CN108363492A (en) 2018-08-03

Similar Documents

Publication Publication Date Title
WO2019169854A1 (en) Human-computer interaction method, and interactive robot
CN109844741B (en) Generating responses in automated chat
CN106297789B (en) Personalized interaction method and system for intelligent robot
Lee et al. Can robots manifest personality?: An empirical test of personality recognition, social responses, and social presence in human–robot interaction
US11276398B1 (en) System and method for providing natural language recommendations
US20110144804A1 (en) Device and method for expressing robot autonomous emotions
JP2019523714A (en) Multi-interaction personality robot
KR20020071917A (en) User interface/entertainment device that simulates personal interaction and charges external database with relevant data
JP2018014094A (en) Virtual robot interaction method, system, and robot
KR20020067591A (en) Self-updating user interface/entertainment device that simulates personal interaction
KR20020067592A (en) User interface/entertainment device that simulates personal interaction and responds to user&#39;s mental state and/or personality
KR20020067590A (en) Environment-responsive user interface/entertainment device that simulates personal interaction
CN110413755A (en) A kind of extending method, device and server, the storage medium in question and answer library
JPWO2005093650A1 (en) Will expression model device, psychological effect program, will expression simulation method
US11074491B2 (en) Emotionally intelligent companion device
Wilks et al. A prototype for a conversational companion for reminiscing about images
CN112204563A (en) System and method for visual scene construction based on user communication
WO2021049254A1 (en) Information processing method, information processing device, and program
Khosla et al. Assistive robot enabled service architecture to support home-based dementia care
Elor et al. isam: Personalizing an artificial intelligence model for emotion with pleasure-arousal-dominance in immersive virtual reality
CN111949773A (en) Reading equipment, server and data processing method
JP2001249949A (en) Feeling generation method, feeling generator and recording medium
Irfan et al. Coffee with a hint of data: towards using data-driven approaches in personalised long-term interactions
Chi et al. Dialogue distillery: Crafting interpolable, interpretable, and introspectable dialogue from llms
JP6034459B1 (en) Interactive interface

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18908471

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18908471

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 10.03.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 18908471

Country of ref document: EP

Kind code of ref document: A1