CN105468145A - Robot man-machine interaction method and device based on gesture and voice recognition - Google Patents

Robot man-machine interaction method and device based on gesture and voice recognition Download PDF

Info

Publication number
CN105468145A
CN105468145A CN201510795938.1A CN201510795938A CN105468145A CN 105468145 A CN105468145 A CN 105468145A CN 201510795938 A CN201510795938 A CN 201510795938A CN 105468145 A CN105468145 A CN 105468145A
Authority
CN
China
Prior art keywords
module
information
gesture
robot
emergency
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510795938.1A
Other languages
Chinese (zh)
Other versions
CN105468145B (en
Inventor
丁希仑
齐静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN201510795938.1A priority Critical patent/CN105468145B/en
Publication of CN105468145A publication Critical patent/CN105468145A/en
Application granted granted Critical
Publication of CN105468145B publication Critical patent/CN105468145B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Manipulator (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention discloses a robot man-machine interaction method and device based on gesture and voice recognition belonging to the technical field of the man-machine interaction and robot. The device comprises a sensor module, a gesture recognition module, a voice recognition module, an information fusion module, an information confirmation module, a robot control module and an emergency help module; the method is: fusing a gesture recognition result and a voice recognition result by the information fusion module, sending the obtained final fusion result to the information confirmation module or the robot control module for realizing emergency help and action execution of the robot. The invention combines two channels of the gesture and voice, overcomes the defect of the single channel and can controls the robot better. The invention can be applied to remotely control the robot, enables the robot to work in fields such as space station, hazardous article treatment and public security instead of human beings, also can realize face to face direct interaction of the man and the robot and can be applied in fields such as medical treatment and domestic service.

Description

A kind of robot man-machine interaction method based on gesture and speech recognition and device
Technical field
The invention belongs to man-machine interaction and robotics, be specifically related to a kind of robot man-machine interaction method based on gesture identification and speech recognition and device.
Background technology
China just marches toward aging society, and the elderly needs to look after, and person between twenty and fifty need work to make a living, and have no time to look after old man.Robot can serve as part labour, as invalid pregnant etc. in the children that helps the elderly.In the process, good man-machine interaction plays an important role.
Man-machine interaction is according to different research fields, usually three kinds of different descriptions are had (see list of references [1]: ArvinAgah, Humaninteractionswithintelligentsystems:researchtaxonomy, ComputersandElectricalEngineering, 27,2001, pp71-107).From in broad terms, man-machine interaction refers to mutual (human-machineinteractions is called for short HMI) of the mankind and machinery compartment; From computer realm, man-machine interaction refers to mutual (human-computerinteraction is called for short HCI) between the mankind and computer system; From robot field, man-machine interaction refers to mutual (human-robotinteraction is called for short HRI) between the mankind and robot.Man-machine interaction, according to the difference of research object, can be divided into these three kinds different descriptions.Three is interrelated, and computing machine and robot are all a kind of special machines, so, can think that HMI comprises HCI and HRI; The normally specific computer system of core of robot, HRI can use the method for HCI, also can independent of HCI.
See list of references [2]: Gong Jiemin, Wang Xianqing, the progress of human-computer interaction technology and developing trend. Xian Electronics Science and Technology University's journal, 1998,25 (6): 782-786, list of references [3]: Liu Kejun, about man-machine interaction, the thinking of man-machine harmonious environment. computer utility, 2005,25 (10): 2226-2227. man-machine interactions experienced by elementary man-machine interaction stage, single channel man-machine interaction stage, the elementary man-machine interaction of future development that two or more passage of present forward combines uses the particular devices such as simple telepilot to realize man-machine conversation usually, be generally unidirectional, seldom there is the person-to-person feedback of machine, usually need people to coordinate robot just can complete particular task, single channel man-machine interaction uses single mode interaction technique, as voice, gesture, sense of touch, eye motion etc. are (see list of references [4]: Potamianos, G., Neti, C., LuettJ., etal., Audio-VisualAutomaticSpeechRecognition:AnOverview.Issues inVisualandAudio-VisualSpeechProcessing, 2004, 356-396. list of references [5]: Pavlovic, V.I., Sharma, R., Huang, T.S., VisualInterpretationofHandGesturesforHuman-ComputerInter action:AReview.IEEETransactionsonPatternAnalysisandMachi neIntelligence, 1997, 19 (7): 677-695. list of references [6]: Benali-Khoudja, M., Hafez, M., Alexandre, J.M., etal., TactileInterfaces:AState-of-the-ArtSurvey, Citeseer, 2004. lists of references [7]: Grauman, K., Betke, M., Lombardi, J., etal., CommunicationViaEyeBlinksandEyebrowRaises:Video-BasedHum an-ComputerInterfaces.UniversalAccessintheInformationSoc iety, 2003, 2 (4): 359-373.), this mode is conducive to the development of nature man-machine interaction, but each passage has oneself advantage and shortcoming, there is certain limitation.In order to learn from other's strong points to offset one's weaknesses, the more each channel information of good utilisation, two or more passage combines becomes the trend of man-machine interaction development.
As, voice and gesture are as daily mutual working gangway, but speech recognition, as the IBMViaVoice of IBM Corporation, the MicrosoftSpeechSDK of Microsoft, the CMUPocketSphinx etc. of Carnegie Mellon University, discrimination is to a certain extent by the impact of speaker's dialect, voice, intonation, surrounding environment etc.; The gesture identification of view-based access control model is without the need to specialized training user, do not need to wear specialized equipment yet, have directly perceived, contain much information, the feature of close friend naturally, meet the requirement of natural man-machine interaction, one of core technology of multimode man-machine interaction, but easily by the impact of illumination, complex background, partial occlusion etc.Static gesture is conventional a kind of information representation mode, and has multiple expression-form, but the photo needing resolution higher in Hand Gesture Segmentation and identifying, when staff is away from camera, the resolution of the staff region photo collected affects recognition effect.
Six sufficient legs/arms combined mobile robots can be used for the fields such as space station, nuclear power station, hypertoxic biochemical plant working, dangerous material process and public safety be anti-riot, also can be used for the field such as family, medical services.Six sufficient legs/arms combined mobile robots not only can be used for Long-distance Control, also can be used for the face-to-face direct interaction with robot, as, when six sufficient legs/arms combined mobile robots are rescued, not only can rescue by remote control robot, also can be on-the-spot after calamity, direct interaction face-to-face with robot, good man-machine interaction contributes to the better assisting users of robot and completes rescue task.
Six sufficient legs/arms combined mobile robots also can be used for the field such as medical treatment, home services, e.g., help old children invalid pregnant; Children are accompanied to play; Side old man, patient, physical disabilities, pregnant woman take medicine, pick up thing on the ground; When old man, patient, physical disabilities, pregnant woman fall down, or when there is emergency condition in healthy ordinary people, and as plundered, use gesture or voice help-asking, system just can notify household in the mode such as note, multimedia message in time, to take appropriate measure in time.
Summary of the invention
The object of the present invention is to provide a kind of robot man-machine interaction method based on gesture and speech recognition and device, gesture and voice two passages are combined, overcomes the deficiency of single channel, better control.The present invention can be used for remote control robot, allow robot replaces the mankind in space station, nuclear power station, hypertoxic biochemical plant working, dangerous material process, nuclear power station, the field operation such as hypertoxic biochemical plant working, also the face-to-face direct interaction of human and computer people can be carried out, for fields such as medical treatment, home services.
Installation environment detecting sensor on device provided by the invention, can monitoring of environmental, at breaking out of fire, CH 4in content overproof or CO content overproof situation, system notifies predefined particular person in time in modes such as voice messaging, note, multimedia messages.Wherein, multimedia message is the scene photograph that camera is taken when abnormal conditions occur.
The man-machine interaction of indication of the present invention is mutual (HRI) between people (user) and robot.
First the present invention provides a kind of robot man-machine interaction method based on gesture and speech recognition, and described method comprises the steps:
The first step, has judged whether interactive object, if there is interactive object, then opens interactive mode, turns second step; If there is no interactive object, robot open detection pattern;
Second step, information inputs: gesture recognition module and sound identification module carry out information acquisition in real time, if the information of collecting, then performs the 3rd step, otherwise performs the 5th step;
3rd step, gesture recognition module, by RGB-D camera collection deep image information and RGB image information, is carried out the identification of Pre-defined gesture, and gesture identification result is sent to information fusion module; Meanwhile, sound identification module gathers audio-frequency information by the built-in Mike of RGB-D camera, is converted to particular text information as voice identification result, and voice identification result is sent to information fusion module by speech recognition software; Voice identification result and gesture identification result are carried out information fusion from semantic layer by information fusion module, obtain final fusion results;
4th step, perform and feedback, the final fusion results according to information fusion module sends command adapted thereto;
If final fusion results is control information, then send corresponding steering order to robot control module, control moves; If final fusion results is emergency information, then send out emergency information to validation of information module, validation of information module is broadcast to user by speech form, whether inquiry user will perform, obtain positive reply or not response in setting-up time, then validation of information module sends emergency information to emergent emergency module, turns the 6th step.
5th step, if gesture recognition module and sound identification module do not have information to input in setting-up time, and after this situation certain time, information fusion module sends emergency information to validation of information module, allow validation of information module carry out indicative voice inquiry according to current task, turn the first step;
6th step, emergent emergency:
Emergent emergency module receives the emergency information that validation of information module sends over, or temperature, CH that sensor assembly collects 4when content or CO content exceed certain value, emergent emergency module sends voice, note and multimedia message emergency to the particular person of pre-registration.
The present invention also provides a kind of robot human-computer interaction device based on gesture and speech recognition, comprises sensor assembly, gesture recognition module, sound identification module, information fusion module, validation of information module, robot control module and emergent emergency module; Sensor assembly comprises RGB-D camera, Temperature Humidity Sensor, CH 4detecting sensor and CO detecting sensor, the RGB image information of described RGB-D camera collection and deep image information send to gesture recognition module or emergent emergency module; Described RGB-D camera has built-in Mike, and the voice messaging that Mike gathers sends to sound identification module; The form that described validation of information module is broadcasted by loudspeaker broadcasts emergency information to user, confirms to obtain user; Described Temperature Humidity Sensor, CH 4detecting sensor and CO detecting sensor are respectively used to gather CH in humiture, air 4with CO gas concentration, and the data of collection are sent to emergent emergency module; Gesture recognition module carries out gesture identification according to the RGB image information of RGB-D camera collection and deep image information, obtains gesture identification result and sends to information fusion module; The voice messaging that Mike gathers by sound identification module carries out speech recognition, obtains specific character information, as voice identification result, and recognition result is sent to information fusion module; Information fusion module is that gesture identification result and voice identification result are carried out information fusion at semantic layer, generate a final fusion results, when final fusion results is control information, control information is then sent to robot control module by information fusion module, and control completes particular task; When final fusion results is emergency information, information fusion module then sends emergency information to validation of information module, whether the form inquiry user that validation of information module is broadcasted by loudspeaker will perform, obtain positive reply or not response in setting-up time, validation of information module sends emergency information to emergent emergency module; After emergent emergency module receives the emergent message that validation of information module transmits, or temperature, CH that sensor assembly collects 4when content or CO content exceed certain value, then emergent emergency module sends voice, note and multimedia message emergency to the particular person of registration.
The invention has the advantages that:
(1) the robot man-machine interaction method based on gesture and speech recognition of the present invention, makes to use gesture and voice two passage carries out information fusion from semantic layer, overcomes the deficiency of single exchange channels, improves man-machine interaction effect.
(2) the robot man-machine interaction method based on gesture and speech recognition of the present invention, remote controlled robot, in hazardous location operation, also can be used for robot closely mutual.
(3) the robot man-machine interaction method based on gesture and speech recognition of the present invention, have detecting pattern and interactive mode, interactive mode is divided into again gesture mode, speech pattern and united mode.Wherein, detecting pattern can monitoring of environmental automatically.When situations such as breaking out of fire, gas leak, natural gas leakings, send note, voice messaging or multimedia message from the predefined particular person of trend.Interactive mode not only can with robot interactive, also can occur emergency condition time, for crying for help to particular person.With robot interactive process, user can select distinct interaction pattern according to different occasion, and can use specific instruction in speech pattern, speech pattern is converted to gesture mode or united mode.As user will talk with other people, think terminated speech control model, switch to single, can use voice command " gesture mode ", now, can only use gesture control.
(4) the robot man-machine interaction method based on gesture and speech recognition of the present invention, has certain versatility.The method of the invention not only can be used for teleoperation robot, also can be used for closely mutual robot.The method of the invention has certain portability, and user according to the specific function of specific robotic, can do suitable predefine in gesture recognition module and sound identification module, and voice and gesture just can be used to complete particular task.
(5) the robot man-machine interaction method based on gesture and speech recognition of the present invention, specific gesture is predefined and voice are distress signal in gesture recognition module and sound identification module, user is in some emergency case, as, house robbery, fall down after can not climb, the situation such as can only to call for help, by special sound or gesture emergency.
Accompanying drawing explanation
Fig. 1 is the structural framing figure of the robot human-computer interaction device based on gesture and speech recognition;
Fig. 2 is sensor module structure frame diagram;
Fig. 3 is the robot man-machine interaction method process flow diagram based on gesture and speech recognition.
Embodiment
Below in conjunction with drawings and Examples, the present invention is described in detail.
The invention provides a kind of robot human-computer interaction device based on gesture and speech recognition, as shown in Figure 1, described human-computer interaction device is based on the groovy version of the RobotOperatingSystem (ROS) that LinuxUbuntu12.04 installs, and human-computer interaction device of the present invention comprises: sensor assembly, gesture recognition module, sound identification module, information fusion module, validation of information module, robot control module and emergent emergency module.In the present invention, be different node by gesture recognition module, sound identification module, information fusion module, validation of information module, robot control module and emergent emergency module abstracts, namely the different node in ROS, then the communication of intermodule can be regarded as between different node and communicates, therefore intermodule is with topic form communication in ROS, the content of communication is the message in topic.
Sensor assembly comprises RGB-D camera, Temperature Humidity Sensor, CH 4detecting sensor and CO detecting sensor, as shown in Figure 2.RGB-D camera can not only collect common RGB photo, can also collect depth image simultaneously.In depth image, the value of each pixel represents the distance between object corresponding to this pixel and camera, depth image can be regarded as a kind of single pass gray-scale map, the distance of what only the gray-scale value of each pixel represented the is object distance camera that this pixel is corresponding.Described RGB-D camera is Asus XtionProLive.The RGB image information of described RGB-D camera collection and deep image information send to gesture recognition module or emergent emergency module; Described RGB-D camera has built-in Mike, and the voice messaging that Mike gathers sends to sound identification module; The form that described validation of information module is broadcasted by loudspeaker broadcasts emergency information to user, confirms to obtain user.Described Temperature Humidity Sensor, CH 4detecting sensor and CO detecting sensor are respectively used to gather CH in humiture, air 4with CO gas concentration, and the data of collection are sent to emergent emergency module.
Gesture recognition module, the first kind of Pre-defined gesture and implication, then carry out gesture identification according to the RGB image information of RGB-D camera collection and deep image information, obtain gesture identification result and send to information fusion module.When gesture recognition module carries out gesture identification, OpenNI storehouse can be called and identify, also can use oneself specific Gesture Recognition Algorithm.Described Pre-defined gesture can carry out concrete restriction as required, and citing below provides several Pre-defined gesture.
Following predefine is done in gesture recognition module:
For six sufficient legs/arms combined mobile robots, the gesture based on single hand of image is done to give a definition: a finger, and representative allows robot lift one leg; Two fingers, representative allows robot lift two legs; Three fingers, representative allows robot lift three legs, if allow robot ambulation, then means and allows robot walk with " 3+3 " gait; Four fingers, representative allows robot carry out the switching of wheel leg, mode that original state is defaulted as " leg is capable "; Five fingers, representative of consumer requires that robot sends distress signals; The gesture (namely just stretching out thumb and little finger of toe) of numeral 6, representative allows robot six legs land; OK gesture, representative " good " "Yes" etc. confirms operation, is the affirmative acknowledgement (ACK) to validation of information module; Clench fist, representing the negative implications such as "no", is the negative acknowledge to validation of information module.
Do to give a definition to the gesture based on video: hand moves upward, representative allows robot go ahead; Hand moves downward, and representative allows robot walk backward; Hand is to left movement, and representative allows robot turn left; Hand moves right, and representative allows robot turn right; Hand pushes away forward, and representative allows robot stop motion; Hand draws circle clockwise, and representative allows robot original place turn clockwise; Hand draws circle counterclockwise, and representative allows robot original place be rotated counterclockwise.
Sound identification module, carries out speech recognition by the voice messaging that Mike gathers, obtains specific character information, as voice identification result, and recognition result is sent to information fusion module.Described speech recognition can adopt the CMUPocketSphinx software simulating of Carnegie Mellon University.
Following predefine is done at sound identification module:
User's predefine voice command in sound identification module is as follows: voice command " speech pattern ", represents that robot enters Voice command pattern; Voice command " gesture mode ", represents that robot enters gesture control model; Voice command " united mode ", represents that robot enters gesture and voice jointly control pattern; Voice command " help ", represents and allows robot send distress signals; Voice command " advance ", control is goed ahead; Voice command " retrogressing ", control is walked back; Voice command " left-hand rotation ", control turns left; Voice command " right-hand rotation ", control is turned right; Voice command " stopping ", control stop motion; Voice command " acceleration ", control pulling speed; Voice command " deceleration ", control underspeeds; Phonetic order "Yes", represents the affirmative acknowledgement (ACK) to validation of information in validation of information module; Phonetic order "no", represents the negative acknowledge to validation of information in validation of information module.
Information fusion module is that gesture identification result and voice identification result are carried out information fusion at semantic layer, generates a final fusion results and sends to robot control module or validation of information module.Specifically, when final fusion results is control information, control information is then sent to robot control module by information fusion module, and control completes particular task; When final fusion results is emergency information, information fusion module then sends emergency information to validation of information module, whether the form inquiry user that validation of information module is broadcasted by loudspeaker will perform, obtain positive reply or not response in setting-up time T (as 3 minutes), validation of information module sends emergency information to emergent emergency module.
Information fusion module, if only have gesture information to input, using gesture identification result as final fusion results; If only have voice messaging to input, using voice identification result as final fusion results; If existing voice messaging, gesture information is had again to input, and gesture identification result and voice identification result do not conflict, then the final fusion results of gesture and voice is optionally issued robot control module or validation of information module by information fusion module: when final fusion results is control information, final fusion results is issued robot control module, and control moves; When final fusion results is emergency information, sends emergency information and carry out validation of information to validation of information module, avoid maloperation.When gesture identification result and voice identification result conflict, two kinds of results are directly sent to validation of information module, then confirms which kind of performs operates by validation of information module to user.
Information fusion module carries out the fusion that information fusion divides three levels: data Layer, characteristic layer and decision-making level.The difference of these three levels is that the data type merged is different: data Layer directly merges the raw data gathered; Characteristic layer be the raw data that gathers after feature extraction, then carry out comprehensive treatment and analysis; Decision-making level, from the demand of concrete decision-making, directly for decision objective, makes full use of the characteristic information of extraction, is three grades of net results merged, directly affects level of decision-making.
Information fusion module in the present invention adopts decision-making level's information fusion.This kind of information fusion mode is applied the most general in these three kinds of amalgamation modes, it is the information integration based on semantic layer, described decision-making level's information fusion mode first to single channel interactive mode from Semantic Layer Comprehend, then by two or more channel information merge get up.It is generally applicable to loose coupling.Compared with feature fusion, the advantage of decision-making level's information fusion is: reliability and antijamming capability are comparatively strong, and the computing cost in processing procedure is less.Shortcoming is the correlativity can not looking for each passage from bottom data.
The control information obtained from information fusion module is converted to the actual motion of robot by robot control module, realizes the specific action of robot.
Validation of information module by final fusion results in the form of speech (loudspeaker broadcast) feed back to user, if obtain user's positive reply, then perform prompting task; If obtain user's negative reply, then do not perform prompting task; If user's nonreply in preset time t (as 2 minutes), then validation of information module inputs gesture or voice messaging by voice (loudspeaker broadcast) prompting user, if user's still nonreply, does not then perform prompting task.The text message synthetic speech information that voice message in validation of information module adopts speech synthesis technique (Text-to-Speech) to export, then by loudspeaker by sound feedback to user.
Emergent emergency module, after emergent emergency module receives the emergent message that validation of information module transmits, or temperature, CH that sensor assembly collects 4when content or CO content exceed certain value, then emergent emergency module sends voice, note and multimedia message emergency to the particular person of pre-registration.Wherein, voice messaging is the voice messaging of the text message synthesis adopting speech synthesis technique (Text-to-Speech) to export.When multimedia message refers to that emergent emergency module receives emergency information, by the scene photograph of the RGB-D camera shooting in sensor assembly.
The situation that emergent emergency module obtains emergency information from validation of information module is: user is in some special circumstances, as, by house robbery, fall down after cannot get up, when the situations such as emergency of can not making a phone call occur, distress signals are sent by voice or Pre-defined gesture, net result after information fusion module merges, after validation of information module confirms, send emergency information to emergent emergency module, start to start after emergent emergency module receives emergency information, transmission voice messaging, note, multimedia message are cried for help to predefined particular person.
The present invention also provides a kind of robot man-machine interaction method based on gesture and speech recognition, and the method is divided into interactive mode and detecting pattern, and interactive mode is divided into: gesture mode, speech pattern and united mode.Wherein, detecting pattern is used for humiture, CH in testing environment 4gas content or CO gas content.As humiture, CH in environment 4when gas content or CO gas content exceed certain value, under detecting pattern, device of the present invention can send warning message to particular person, and warning message comprises voice messaging, note, photo (humiture, CH in environment 4when content or CO content exceed certain value, the scene photograph of camera shooting).Interactive mode is user by the pattern of gesture or voice and robot interactive.In such a mode, if user is by means of only gesture and robot interactive, this pattern is called gesture mode, when being applicable to user's control, need the situation of talk with other people or neighbourhood noise larger etc., be not suitable for the occasion of use speech recognition; If user is by means of only voice and robot interactive, this pattern is called speech pattern, is applicable to gesture major part or the restricted occasion of gesture identification interactive mode such as to be blocked completely; If user is by voice and gesture two kinds of passages and robot interactive, this pattern is called united mode, this pattern systhesis makes to use gesture and voice two kinds of passages, and these two kinds of passages in some cases can be complementary, adds the scope of application that human and computer people is mutual to a certain extent.
Man-machine interaction method of the present invention flow process as shown in Figure 3, specifically comprises the following steps:
The first step, has judged whether interactive object, if there is interactive object, then opens interactive mode.When robot receives the voice of greeting, as " hello ", " good afternoon " etc., or when robot identifies a certain action, during as " waving ", then judge to have near robot interactive object, open interactive mode.If do not have interactive object, robot is in detecting pattern.
Second step, information inputs.Gesture recognition module and sound identification module carry out information acquisition in real time, if the information of collecting, then perform the 3rd step, otherwise perform the 5th step;
3rd step, gesture recognition module, by RGB-D camera collection deep image information and RGB information, through steps such as Iamge Segmentation, feature extraction, training identifications, is carried out the identification of Pre-defined gesture, and recognition result is sent to information fusion module.Meanwhile, built-in for RGB-D camera Mike is gathered audio-frequency information by sound identification module, be converted to particular text information (SpeechtoText) by speech recognition software (CMUPocketSphinx of Carnegie Mellon University), and voice identification result is sent to information fusion module.Voice identification result and gesture identification result are carried out information fusion from semantic layer by information fusion module, obtain final fusion results.
Information fusion module is carried out semantic layer information fusion, and is sent to robot control module or validation of information module according to final fusion results after receiving the recognition result that gesture recognition module and sound identification module send.After robot control module receives the control information that information fusion module sends, control moves, and completes particular task.After validation of information module receives the emergent distress signals that information fusion module sends, validation of information module confirms to user with regard to whether sending distress signals.
Man-machine interaction method of the present invention has Three models in an interactive mode: gesture mode, speech pattern and united mode, then information fusion module is corresponding also has Three models: gesture mode, speech pattern and united mode.
A, information fusion module are under gesture mode, using gesture identification result as final fusion results, and send out message to robot control module or validation of information module according to the kind of information fusion net result, namely, if final fusion results is control information, then message is sent to robot control module, control completes particular task; If final fusion results is emergency information, then emergency information is sent to validation of information module, be confirmed whether to send emergency information to user.Namely, if user has done predefined emergency gesture, information fusion module receives the emergent distress signals that gesture recognition module is sent, for avoiding maloperation, send out message and carry out validation of information to validation of information module, whether the form that validation of information module is broadcasted with loudspeaker plays current task, and point out user to select to execute the task.When user uses certain gestures, as " OK " gesture, when confirming to perform current task, validation of information module sends message to emergent emergency module, emergent emergency module just sends voice, note and multimedia message to predefined particular person after receiving the emergency information that validation of information module sends over.Wherein, multimedia message is emergent emergency module when receiving the emergent message that validation of information module sends, by the scene photograph of the RGB-D camera shooting in sensor assembly.
B, information fusion module are in voice mode, using voice identification result as final fusion results, and send out message to robot control module or validation of information module according to the kind of information fusion net result, namely, if final fusion results is control information, then message is sent to robot control module, control completes particular task; If final fusion results is emergency information, then message is sent to validation of information module, be confirmed whether to send emergency information to user.Namely, if user uses predefined emergency phonetic order, information fusion module receives the emergent distress signals that sound identification module sends, for avoiding maloperation, send out message and carry out validation of information to validation of information module, whether the form that validation of information module is broadcasted with loudspeaker plays current task, and point out user to select to execute the task.When user uses special sound instruction, as instructions such as "Yes", when confirming to perform current task, validation of information module sends message to emergent emergency module, emergent emergency module just sends voice, note and multimedia message to predefined particular person after receiving the information that validation of information module sends over.Wherein, multimedia message is emergent emergency module when receiving the message that validation of information module sends, by the scene photograph of the RGB-D camera shooting in sensor assembly.
C, information fusion module, under united mode, using the result of gesture and speech recognition as final fusion results, and send out message to robot control module or validation of information module according to the kind of information fusion net result.The acquisition time of gesture and voice channel information may be inconsistent, therefore, after information fusion module obtains a certain channel information, can wait for the input of other channel informations, and setting time-out restriction t (as 10s).After each channel information all inputted, then carry out information fusion, make corresponding decision, trigger corresponding steering order, the motion of control or send information to validation of information module.
If when there is difference in the decision-making of information fusion module design task, as, voice messaging and gesture information afoul time, information fusion module can send out message to validation of information module, the form playing task title successively that validation of information module is broadcasted with loudspeaker, select for user, and point out user tasks clear.When validation of information module receives the asserts feedback of user, start to execute the task.Here asserts feedback available speech is expressed, and also gestures available represents.Voice: as vocabulary such as "Yes"; Gesture: as, " OK " gesture.
4th step, perform and feedback, the final fusion results according to information fusion module sends command adapted thereto.Final fusion results after information fusion is divided into two classes: a class is control information, and a class is emergency information.If control information, then send corresponding steering order, control moves.If emergency information, then send out message to validation of information module, validation of information module is broadcast to user by speech form (loudspeaker broadcast), whether inquiry user will perform, obtain positive reply or not response in setting-up time T (as 3 minutes), then start contingency mode, turn the 6th step.Described positive reply can be expressed by the form of voice or gesture: voice are as "Yes" etc.; Gesture is as gestures such as " OK ".
5th step, in interactive mode, if second step is not response in setting-up time T (as 3 minutes), namely, gesture recognition module can't detect gesture information, sound identification module also can't detect voice messaging, and after this situation certain time T (as 3 minutes), information fusion module sends emergency information to validation of information module, validation of information module is allowed to carry out indicative voice inquiry according to current task, confirm there is no interactive object near robot, then validation of information module is by voice (loudspeaker broadcast) inquiry user, if in special time T (as 3 minutes), second step still can't detect voice messaging and gesture information, then be considered as near robot without interactive object, close interactive mode, open detection pattern, turn the first step.
Described detecting pattern, is used for testing environment situation, as long as robot energising, robot is just in detecting pattern always.In this mode, emergent emergency module is by Temperature Humidity Sensor, the CH in sensor assembly 4detecting sensor and CO detecting sensor detected temperatures, CH 4content or CO content, as temperature, CH 4when content or CO content exceed certain value, after emergent emergency module detects the data that exceed standard, send voice messaging, note and multimedia message emergency to particular person.Wherein, when multimedia message refers to that emergency information sends, by the scene photograph of the RGB-D camera shooting in sensor assembly.
6th step, emergent emergency.Emergent emergency module receives the information that validation of information module sends over, or temperature, CH that sensor assembly collects 4when content or CO content exceed certain value, emergent emergency module sends voice, note and multimedia message emergency to the particular person of pre-registration.Wherein, voice messaging is the voice messaging of the text message synthesis adopting speech synthesis technique (Text-to-Speech) to export.When multimedia message refers to that emergent emergency module receives emergency information, by the scene photograph of the RGB-D camera shooting in sensor assembly.
Fig. 3 is six sufficient legs/arms combined mobile robot man-machine interaction method process flow diagrams of the invention process example, is described in detail the present invention in conjunction with an embodiment:
Embodiment one:
Be below the present invention in an interactive mode, user and six sufficient legs/arms combined mobile robots are based on the implementation procedure of gesture and speech recognition man-machine interaction method:
(1) judge whether have interactive object near robot.When robot has detected voice messaging by Mike, as " hello " " good morning " etc., or certain gestures information detected by camera, as waved, then think there is interactive object near robot, open interactive mode, and play voice messaging with the form of loudspeaker broadcast, as " hello, may I ask what instruction? "
(2) in an interactive mode, user can with predefined gesture or voice and robot interactive.Embodiment is provided below with regard to the gesture mode under interactive mode, speech pattern and united mode.
A, gesture mode, user only can use the embodiment of predefined gesture control as follows:
When user stretches out a finger, robot lifts that leg with clip, simultaneously, the form broadcasting voice of broadcasting with loudspeaker, as " may I ask and will carry out clipping operation? ", when user makes OK gesture, or carry out confirmation answer with voice, as "Yes" etc., then perform robot control module, control completes particular task.
When user stretches out two fingers, robot lifts two legs, start four-footed and support, two sufficient operations, and under the control of robot control module, control completes particular task.
When user stretches out three fingers, robot lifts three legs, when user's hand moves upward, robot travels forward with " 3+3 " gait, when user's hand is to left movement, robot turns left with " 3+3 " gait, when user's hand moves right, robot moves right with " 3+3 " gait, when user's hand moves backward, robot moves backward with " 3+3 " gait, when user stretches out four fingers, robot carries out wheel leg and switches, " wheel row " mode is switched to by " leg is capable " mode, when user's hand moves upward, robot travels forward in " wheel row " mode, when user's hand is to left movement, robot turns left in " wheel row " mode, when user's hand moves right, robot bends to right in " wheel row " mode, when user's hand moves downward, robot moves backward in " wheel row " mode, when user's hand pushes away forward, robot stop motion.
When user makes the gesture (namely just stretching out thumb and little finger of toe) of numeral 6, robot six legs land, when user stretches out 4 fingers, robot switches to " wheel row " mode by " leg is capable " mode, when user's hand moves upward, robot travels forward in " wheel row " mode, when user's hand is to left movement, robot turns left in " wheel row " mode, when user's hand moves right, robot moves right in " wheel row " mode, when user's hand moves backward, robot moves backward in " wheel row " mode, when user's hand pushes away forward, robot stop motion, when user stretches out 4 fingers, robot switches to " leg is capable " mode by " wheel row " mode.
When user's hand draws circle clockwise, robot original place turns clockwise, and when user's hand draws circle counterclockwise, robot original place is rotated counterclockwise.When user's hand pushes away forward, robot stop motion.
When user stretches out five fingers, whether the form inquiry user that robot broadcasts with loudspeaker will cry for help, if robot receive represent confirm gesture information (as, the gesture of OK) or not response in setting-up time T (as 3 minutes), robot then by representing the scene photograph of voice messaging, note and the camera shooting of crying for help, sends to the particular person of pre-registration.Wherein, the scene photograph of camera shooting, be robot when receiving distress signals, by the scene photograph of RGB-D camera in sensor assembly shooting.
If there is no gesture and voice messaging input in special time T (as 3 minutes), the form prompting user that robot broadcasts with loudspeaker inputs gesture or voice messaging, if still there is no gesture and voice messaging input in special time T (as 3 minutes), robot thinks that interactive object leaves, close interactive mode, open detection pattern, turns (1).
B, speech pattern, user only can use the embodiment of predefined Voice command robot as follows:
When user says " speech pattern ", after robot identifies these four words, voice are play with the form of loudspeaker broadcast, as, " you enter speech pattern now ", when user says " advance ", after robot identifies these two words, start to travel forward, when user says " acceleration ", start to accelerate to certain speed after robot identification, when user says " left-hand rotation ", turn left after robot identification, when user says " deceleration ", certain speed is reduced to after robot identification, when user says " right-hand rotation ", turn right after robot identification, when user says " stopping ", robot identifies stop motion.
When user is in reciprocal process, want to talk with other people, terminated speech mode control, that is, user thinks only to use gesture control, user can say " gesture mode ", after robot identifies these four words, play voice with the form of loudspeaker broadcast, as, " you enter gesture mode now, and this pattern can only use gesture control.", in this mode, user can only use gesture control, and instantiation is shown in a.
When user says " help ", after robot identifies these two words, whether the form inquiry user that validation of information module is broadcasted with loudspeaker sends emergency information, if obtain the positive reply of user, as, user says the vocabulary such as "Yes", or not response in setting-up time T (as 3 minutes), robot then by representing the scene photograph of voice messaging, note and the camera shooting of crying for help, sends to the particular person of pre-registration.Wherein, the scene photograph of camera shooting, be robot when receiving distress signals, by the scene photograph of RGB-D camera in sensor assembly shooting.
If there is no gesture and voice messaging input in special time T (as 3 minutes), the form prompting user that robot broadcasts with loudspeaker inputs gesture or voice messaging, if still there is no gesture and voice messaging input in special time T (as 3 minutes), robot thinks that interactive object leaves, close interactive mode, open detection pattern, turns (1).
C, united mode, user makes to use gesture and voice to jointly control the embodiment of robot as follows:
When user says " united mode ", voice are play with the form of loudspeaker broadcast after robot identification, as, " you enter gesture now and voice jointly control pattern ", in this mode, as long as user says a certain phonetic control command or does a certain gesture, robot can complete command adapted thereto, that is, user can use arbitrary voice or gesture control to complete specific action, and voice and gesture can cross-references, specific implementation is identical with a, b, just for the INQUIRE statement of robot, user can use gesture, and also can use voice answering.
When gesture is conflicted with phonetic order, robot plays gesture and semantic task title successively with the form that loudspeaker are broadcasted, often play a task and all inquire whether user performs, if obtain the positive reply of user, then perform this task, positive reply gestures available also available speech form is expressed, and gesture is as the gesture of OK, and voice are as "Yes" etc.After each playing task completes, if can not get user at special time T (as 3 minutes) to answer, then robot does not perform this task, if also have other tasks to need user to confirm, then play all tasks successively with the form of loudspeaker broadcast, and, all inquire after often playing whether user will perform this task, circulate successively, until all tasks all play with the form of loudspeaker broadcast, and all inquired whether user will perform, if there is no gesture and voice messaging input in special time T (as 3 minutes), the form prompting user that robot broadcasts with loudspeaker inputs gesture or voice messaging, if still there is no gesture and voice messaging input in special time T (as 3 minutes), robot thinks that interactive object leaves, close interactive mode, open detection pattern, turn (1).

Claims (5)

1. based on a robot human-computer interaction device for gesture and speech recognition, it is characterized in that: comprise sensor assembly, gesture recognition module, sound identification module, information fusion module, validation of information module, robot control module and emergent emergency module; Sensor assembly comprises RGB-D camera, Temperature Humidity Sensor, CH 4detecting sensor and CO detecting sensor, the RGB image information of described RGB-D camera collection and deep image information send to gesture recognition module or emergent emergency module; Described RGB-D camera has built-in Mike, and the voice messaging that Mike gathers sends to sound identification module; The form that described validation of information module is broadcasted by loudspeaker broadcasts emergency information to user, confirms to obtain user; Described Temperature Humidity Sensor, CH 4detecting sensor and CO detecting sensor are respectively used to gather CH in humiture, air 4with CO gas concentration, and the data of collection are sent to emergent emergency module; Gesture recognition module carries out gesture identification according to the RGB image information of RGB-D camera collection and deep image information, obtains gesture identification result and sends to information fusion module; The voice messaging that Mike gathers by sound identification module carries out speech recognition, obtains specific character information, as voice identification result, and recognition result is sent to information fusion module; Information fusion module is that gesture identification result and voice identification result are carried out information fusion at semantic layer, generate a final fusion results, when final fusion results is control information, control information is then sent to robot control module by information fusion module, and control completes particular task; When final fusion results is emergency information, information fusion module then sends emergency information to validation of information module, whether the form inquiry user that validation of information module is broadcasted by loudspeaker will perform, obtain positive reply or not response in setting-up time, validation of information module sends emergency information to emergent emergency module; After emergent emergency module receives the emergent message that validation of information module transmits, or temperature, CH that sensor assembly collects 4when content or CO content exceed certain value, then emergent emergency module sends voice, note and multimedia message emergency to the particular person of registration.
2. a kind of robot human-computer interaction device based on gesture and speech recognition according to claim 1, is characterized in that: described information fusion module adopts decision-making level's information fusion method.
3. a kind of robot human-computer interaction device based on gesture and speech recognition according to claim 1, is characterized in that: described information fusion module, if only have gesture information to input, using gesture identification result as final fusion results; If only have voice messaging to input, using voice identification result as final fusion results; If existing voice messaging, have again gesture information to input, and gesture identification result and voice identification result do not conflict, then the final fusion results of gesture and voice is optionally issued robot control module or validation of information module by information fusion module; When gesture identification result and voice identification result conflict, two kinds of results are directly sent to validation of information module, then confirms which kind of performs operates by validation of information module to user.
4., based on a robot man-machine interaction method for gesture and speech recognition, it is characterized in that: specifically comprise the following steps,
The first step, has judged whether interactive object, if there is interactive object, then opens interactive mode, turns second step; If there is no interactive object, robot open detection pattern;
Second step, information inputs: gesture recognition module and sound identification module carry out information acquisition in real time, if the information of collecting, then performs the 3rd step, otherwise performs the 5th step;
3rd step, gesture recognition module, by RGB-D camera collection deep image information and RGB image information, is carried out the identification of Pre-defined gesture, and gesture identification result is sent to information fusion module; Meanwhile, sound identification module gathers audio-frequency information by the built-in Mike of RGB-D camera, is converted to particular text information as voice identification result, and voice identification result is sent to information fusion module by speech recognition software; Voice identification result and gesture identification result are carried out information fusion from semantic layer by information fusion module, obtain final fusion results;
4th step, perform and feedback, the final fusion results according to information fusion module sends command adapted thereto;
If final fusion results is control information, then send corresponding steering order to robot control module, control moves; If final fusion results is emergency information, then send out emergency information to validation of information module, validation of information module is broadcast to user by speech form, whether inquiry user will perform, obtain positive reply or not response in setting-up time, then validation of information module sends emergency information to emergent emergency module, turns the 6th step.
5th step, if gesture recognition module and sound identification module do not have information to input in setting-up time, and after this situation certain time, information fusion module sends emergency information to validation of information module, allow validation of information module carry out indicative voice inquiry according to current task, turn the first step;
6th step, emergent emergency;
Emergent emergency module receives the emergency information that validation of information module sends over, or temperature, CH that sensor assembly collects 4when content or CO content exceed certain value, emergent emergency module sends voice, note or multimedia message emergency to the particular person of pre-registration.
5. a kind of robot man-machine interaction method based on gesture and speech recognition according to claim 4, is characterized in that: described detecting pattern, is used for testing environment situation; In detecting pattern, emergent emergency module is by Temperature Humidity Sensor, the CH in sensor assembly 4detecting sensor and CO detecting sensor detected temperatures, CH 4content or CO content, as temperature, CH 4when content or CO content exceed certain value, after emergent emergency module detects the data that exceed standard, send voice messaging, note and multimedia message emergency to particular person.
CN201510795938.1A 2015-11-18 2015-11-18 A kind of robot man-machine interaction method and device based on gesture and speech recognition Active CN105468145B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510795938.1A CN105468145B (en) 2015-11-18 2015-11-18 A kind of robot man-machine interaction method and device based on gesture and speech recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510795938.1A CN105468145B (en) 2015-11-18 2015-11-18 A kind of robot man-machine interaction method and device based on gesture and speech recognition

Publications (2)

Publication Number Publication Date
CN105468145A true CN105468145A (en) 2016-04-06
CN105468145B CN105468145B (en) 2019-05-28

Family

ID=55605929

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510795938.1A Active CN105468145B (en) 2015-11-18 2015-11-18 A kind of robot man-machine interaction method and device based on gesture and speech recognition

Country Status (1)

Country Link
CN (1) CN105468145B (en)

Cited By (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105955489A (en) * 2016-05-26 2016-09-21 苏州活力旺机器人科技有限公司 Robot gesture identification teaching apparatus and method
CN106019977A (en) * 2016-08-05 2016-10-12 易晓阳 Gesture and emotion recognition home control system
CN106056063A (en) * 2016-05-26 2016-10-26 吴昊 Recognition and control system of robot
CN106125566A (en) * 2016-08-05 2016-11-16 易晓阳 A kind of household background music control system
CN106125925A (en) * 2016-06-20 2016-11-16 华南理工大学 Method is arrested based on gesture and voice-operated intelligence
CN106125565A (en) * 2016-08-05 2016-11-16 易晓阳 A kind of motion and emotion recognition house control system
CN106200396A (en) * 2016-08-05 2016-12-07 易晓阳 A kind of appliance control method based on Motion Recognition
CN106200395A (en) * 2016-08-05 2016-12-07 易晓阳 A kind of multidimensional identification appliance control method
CN106228982A (en) * 2016-07-27 2016-12-14 华南理工大学 A kind of interactive learning system based on education services robot and exchange method
CN106297338A (en) * 2016-09-14 2017-01-04 深圳市喜悦智慧数据有限公司 A kind of traffic robot's control system and method
CN106372574A (en) * 2016-08-22 2017-02-01 湖南晖龙股份有限公司 ROS operation system-based robot object identification method
CN106426203A (en) * 2016-11-02 2017-02-22 旗瀚科技有限公司 Communication system and method of active trigger robot
CN106486122A (en) * 2016-12-26 2017-03-08 旗瀚科技有限公司 A kind of intelligent sound interacts robot
CN106527729A (en) * 2016-11-17 2017-03-22 科大讯飞股份有限公司 Non-contact type input method and device
CN106648054A (en) * 2016-10-08 2017-05-10 河海大学常州校区 Multi-mode interactive method for RealSense-based accompanying robot
CN106681508A (en) * 2016-12-29 2017-05-17 杭州电子科技大学 System for remote robot control based on gestures and implementation method for same
CN106847279A (en) * 2017-01-10 2017-06-13 西安电子科技大学 Man-machine interaction method based on robot operating system ROS
CN106873773A (en) * 2017-01-09 2017-06-20 北京奇虎科技有限公司 Robot interactive control method, server and robot
CN107150347A (en) * 2017-06-08 2017-09-12 华南理工大学 Robot perception and understanding method based on man-machine collaboration
CN107390714A (en) * 2017-06-13 2017-11-24 深圳市易成自动驾驶技术有限公司 Control method, device and the computer-readable recording medium of unmanned plane
CN107544271A (en) * 2017-09-18 2018-01-05 广东美的制冷设备有限公司 Terminal control method, device and computer-readable recording medium
WO2018023514A1 (en) * 2016-08-04 2018-02-08 易晓阳 Home background music control system
WO2018023523A1 (en) * 2016-08-04 2018-02-08 易晓阳 Motion and emotion recognizing home control system
WO2018023515A1 (en) * 2016-08-04 2018-02-08 易晓阳 Gesture and emotion recognition home control system
CN107688779A (en) * 2017-08-18 2018-02-13 北京航空航天大学 A kind of robot gesture interaction method and apparatus based on RGBD camera depth images
CN108124131A (en) * 2017-12-22 2018-06-05 合肥寰景信息技术有限公司 A kind of monitoring and alarming system of recognizable deliberate action
CN108320456A (en) * 2018-01-27 2018-07-24 西安交通大学 It is a kind of fusion multisensor the elderly fall down prediction technique and system
CN108399427A (en) * 2018-02-09 2018-08-14 华南理工大学 Natural interactive method based on multimodal information fusion
CN108415561A (en) * 2018-02-11 2018-08-17 北京光年无限科技有限公司 Gesture interaction method based on visual human and system
CN108509026A (en) * 2018-02-06 2018-09-07 西安电子科技大学 Tele-Maintenance Support System and method based on enhancing interactive mode
CN108595012A (en) * 2018-05-10 2018-09-28 北京光年无限科技有限公司 Visual interactive method and system based on visual human
CN108766435A (en) * 2018-05-17 2018-11-06 东莞市华睿电子科技有限公司 A kind of robot for space control method based on non-touch
CN108833766A (en) * 2018-03-21 2018-11-16 北京猎户星空科技有限公司 Control method, device, smart machine and the storage medium of smart machine
CN108847228A (en) * 2018-05-17 2018-11-20 东莞市华睿电子科技有限公司 A kind of robot for space control method based on double sounding
CN109088802A (en) * 2018-09-13 2018-12-25 天津西青区瑞博生物科技有限公司 A kind of speech recognition household robot based on Android control platform
WO2019019146A1 (en) * 2017-07-28 2019-01-31 李庆远 Video social interaction system with space robot
CN109410940A (en) * 2018-12-05 2019-03-01 湖北安心智能科技有限公司 A kind of man-machine interaction method and system based on indication control board
WO2019041301A1 (en) * 2017-09-01 2019-03-07 李庆远 Video-based social interaction system employing robotic device
CN109623848A (en) * 2019-02-26 2019-04-16 江苏艾萨克机器人股份有限公司 A kind of hotel service robot
CN109676621A (en) * 2019-01-04 2019-04-26 中船第九设计研究院工程有限公司 A kind of man machine language's exchange method based on ROS robot operating system
CN109993945A (en) * 2019-04-04 2019-07-09 清华大学 For gradually freezing the alarm system and alarm method of disease patient monitoring
CN110096154A (en) * 2019-05-08 2019-08-06 北京百度网讯科技有限公司 For handling the method and device of information
CN110202592A (en) * 2019-07-02 2019-09-06 江苏博子岛智能产业技术研究院有限公司 A kind of AI mobile medical service robot
CN110228065A (en) * 2019-04-29 2019-09-13 北京云迹科技有限公司 Motion planning and robot control method and device
CN110232922A (en) * 2019-04-29 2019-09-13 北京云迹科技有限公司 Robot voice halts control method and device
CN110675582A (en) * 2018-07-03 2020-01-10 北京蜂盒科技有限公司 Automatic alarm method and device
CN110797017A (en) * 2019-07-30 2020-02-14 深圳市南和移动通信科技股份有限公司 Voice help calling method, intelligent sound box and storage medium
CN110786859A (en) * 2018-08-03 2020-02-14 格力电器(武汉)有限公司 Emergency alarm method, device and system
CN110827931A (en) * 2020-01-13 2020-02-21 四川大学华西医院 Method and device for managing clinical terms and readable storage medium
CN110857067A (en) * 2018-08-24 2020-03-03 上海汽车集团股份有限公司 Human-vehicle interaction device and human-vehicle interaction method
CN111225028A (en) * 2018-11-27 2020-06-02 阿瓦亚公司 System and method for providing enhanced routing in a contact center
CN111324206A (en) * 2020-02-28 2020-06-23 重庆百事得大牛机器人有限公司 Gesture interaction-based confirmation information identification system and method
CN111382723A (en) * 2020-03-30 2020-07-07 北京云住养科技有限公司 Method, device and system for identifying help
CN111931605A (en) * 2020-07-23 2020-11-13 中核核电运行管理有限公司 Intelligent monitoring system and method for high-risk operation of nuclear power plant
CN112099632A (en) * 2020-09-16 2020-12-18 济南大学 Human-robot cooperative interaction method for assistant accompanying
CN112224304A (en) * 2020-10-28 2021-01-15 北京理工大学 Wheel step composite mobile platform and gesture and voice control method thereof
CN112568692A (en) * 2019-09-29 2021-03-30 浙江苏泊尔家电制造有限公司 Control method of cooking appliance, cooking appliance and computer storage medium
CN112866064A (en) * 2021-01-04 2021-05-28 欧普照明电器(中山)有限公司 Control method, control system and electronic equipment
CN113192503A (en) * 2021-04-28 2021-07-30 深圳市金画王技术有限公司 Intelligent voice control system of water surface lifesaving robot
CN113341820A (en) * 2021-06-16 2021-09-03 江苏纬信工程咨询有限公司 Intelligent construction site safety monitoring device based on Internet of things and monitoring method thereof
CN113359538A (en) * 2020-03-05 2021-09-07 东元电机股份有限公司 Voice control robot
CN113848790A (en) * 2021-09-28 2021-12-28 德州学院 Intelligent nursing type robot system and control method thereof
WO2022001120A1 (en) * 2020-06-30 2022-01-06 江苏科技大学 Multi-agent system and control method therefor
US11654573B2 (en) 2020-05-29 2023-05-23 Tata Consultancy Services Limited Methods and systems for enabling human robot interaction by sharing cognition

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013150076A1 (en) * 2012-04-04 2013-10-10 Aldebaran Robotics Robot capable of incorporating natural dialogues with a user into the behaviour of same, and methods of programming and using said robot
CN103885585A (en) * 2014-02-20 2014-06-25 深圳市贝特尔机电有限公司 Robot manual navigation method based on single-person gestures and voice information
CN103984315A (en) * 2014-05-15 2014-08-13 成都百威讯科技有限责任公司 Domestic multifunctional intelligent robot

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013150076A1 (en) * 2012-04-04 2013-10-10 Aldebaran Robotics Robot capable of incorporating natural dialogues with a user into the behaviour of same, and methods of programming and using said robot
CN103885585A (en) * 2014-02-20 2014-06-25 深圳市贝特尔机电有限公司 Robot manual navigation method based on single-person gestures and voice information
CN103984315A (en) * 2014-05-15 2014-08-13 成都百威讯科技有限责任公司 Domestic multifunctional intelligent robot

Cited By (75)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105955489A (en) * 2016-05-26 2016-09-21 苏州活力旺机器人科技有限公司 Robot gesture identification teaching apparatus and method
CN106056063A (en) * 2016-05-26 2016-10-26 吴昊 Recognition and control system of robot
CN106125925B (en) * 2016-06-20 2019-05-14 华南理工大学 Intelligence based on gesture and voice control arrests method
CN106125925A (en) * 2016-06-20 2016-11-16 华南理工大学 Method is arrested based on gesture and voice-operated intelligence
CN106228982B (en) * 2016-07-27 2019-11-15 华南理工大学 A kind of interactive learning system and exchange method based on education services robot
CN106228982A (en) * 2016-07-27 2016-12-14 华南理工大学 A kind of interactive learning system based on education services robot and exchange method
WO2018023515A1 (en) * 2016-08-04 2018-02-08 易晓阳 Gesture and emotion recognition home control system
WO2018023523A1 (en) * 2016-08-04 2018-02-08 易晓阳 Motion and emotion recognizing home control system
WO2018023514A1 (en) * 2016-08-04 2018-02-08 易晓阳 Home background music control system
CN106200395A (en) * 2016-08-05 2016-12-07 易晓阳 A kind of multidimensional identification appliance control method
CN106019977A (en) * 2016-08-05 2016-10-12 易晓阳 Gesture and emotion recognition home control system
CN106125566A (en) * 2016-08-05 2016-11-16 易晓阳 A kind of household background music control system
CN106125565A (en) * 2016-08-05 2016-11-16 易晓阳 A kind of motion and emotion recognition house control system
CN106200396A (en) * 2016-08-05 2016-12-07 易晓阳 A kind of appliance control method based on Motion Recognition
CN106372574A (en) * 2016-08-22 2017-02-01 湖南晖龙股份有限公司 ROS operation system-based robot object identification method
CN106297338A (en) * 2016-09-14 2017-01-04 深圳市喜悦智慧数据有限公司 A kind of traffic robot's control system and method
CN106648054A (en) * 2016-10-08 2017-05-10 河海大学常州校区 Multi-mode interactive method for RealSense-based accompanying robot
CN106648054B (en) * 2016-10-08 2019-07-16 河海大学常州校区 A kind of Multimodal interaction method of the company robot based on RealSense
CN106426203A (en) * 2016-11-02 2017-02-22 旗瀚科技有限公司 Communication system and method of active trigger robot
CN106527729A (en) * 2016-11-17 2017-03-22 科大讯飞股份有限公司 Non-contact type input method and device
CN106486122A (en) * 2016-12-26 2017-03-08 旗瀚科技有限公司 A kind of intelligent sound interacts robot
CN106681508A (en) * 2016-12-29 2017-05-17 杭州电子科技大学 System for remote robot control based on gestures and implementation method for same
CN106873773A (en) * 2017-01-09 2017-06-20 北京奇虎科技有限公司 Robot interactive control method, server and robot
CN106847279A (en) * 2017-01-10 2017-06-13 西安电子科技大学 Man-machine interaction method based on robot operating system ROS
CN107150347A (en) * 2017-06-08 2017-09-12 华南理工大学 Robot perception and understanding method based on man-machine collaboration
CN107390714A (en) * 2017-06-13 2017-11-24 深圳市易成自动驾驶技术有限公司 Control method, device and the computer-readable recording medium of unmanned plane
WO2019019146A1 (en) * 2017-07-28 2019-01-31 李庆远 Video social interaction system with space robot
CN107688779A (en) * 2017-08-18 2018-02-13 北京航空航天大学 A kind of robot gesture interaction method and apparatus based on RGBD camera depth images
WO2019041301A1 (en) * 2017-09-01 2019-03-07 李庆远 Video-based social interaction system employing robotic device
CN107544271A (en) * 2017-09-18 2018-01-05 广东美的制冷设备有限公司 Terminal control method, device and computer-readable recording medium
CN107544271B (en) * 2017-09-18 2020-08-14 广东美的制冷设备有限公司 Terminal control method, device and computer readable storage medium
US11417331B2 (en) 2017-09-18 2022-08-16 Gd Midea Air-Conditioning Equipment Co., Ltd. Method and device for controlling terminal, and computer readable storage medium
CN108124131A (en) * 2017-12-22 2018-06-05 合肥寰景信息技术有限公司 A kind of monitoring and alarming system of recognizable deliberate action
CN108320456A (en) * 2018-01-27 2018-07-24 西安交通大学 It is a kind of fusion multisensor the elderly fall down prediction technique and system
CN108509026B (en) * 2018-02-06 2020-04-14 西安电子科技大学 Remote maintenance support system and method based on enhanced interaction mode
CN108509026A (en) * 2018-02-06 2018-09-07 西安电子科技大学 Tele-Maintenance Support System and method based on enhancing interactive mode
CN108399427A (en) * 2018-02-09 2018-08-14 华南理工大学 Natural interactive method based on multimodal information fusion
CN108415561A (en) * 2018-02-11 2018-08-17 北京光年无限科技有限公司 Gesture interaction method based on visual human and system
CN108833766A (en) * 2018-03-21 2018-11-16 北京猎户星空科技有限公司 Control method, device, smart machine and the storage medium of smart machine
CN108595012A (en) * 2018-05-10 2018-09-28 北京光年无限科技有限公司 Visual interactive method and system based on visual human
CN108847228A (en) * 2018-05-17 2018-11-20 东莞市华睿电子科技有限公司 A kind of robot for space control method based on double sounding
CN108766435A (en) * 2018-05-17 2018-11-06 东莞市华睿电子科技有限公司 A kind of robot for space control method based on non-touch
CN110675582A (en) * 2018-07-03 2020-01-10 北京蜂盒科技有限公司 Automatic alarm method and device
CN110786859A (en) * 2018-08-03 2020-02-14 格力电器(武汉)有限公司 Emergency alarm method, device and system
CN110857067A (en) * 2018-08-24 2020-03-03 上海汽车集团股份有限公司 Human-vehicle interaction device and human-vehicle interaction method
CN109088802A (en) * 2018-09-13 2018-12-25 天津西青区瑞博生物科技有限公司 A kind of speech recognition household robot based on Android control platform
CN111225028A (en) * 2018-11-27 2020-06-02 阿瓦亚公司 System and method for providing enhanced routing in a contact center
CN111225028B (en) * 2018-11-27 2023-04-18 阿瓦亚公司 System and method for providing enhanced routing in a contact center
CN109410940A (en) * 2018-12-05 2019-03-01 湖北安心智能科技有限公司 A kind of man-machine interaction method and system based on indication control board
CN109676621A (en) * 2019-01-04 2019-04-26 中船第九设计研究院工程有限公司 A kind of man machine language's exchange method based on ROS robot operating system
CN109623848A (en) * 2019-02-26 2019-04-16 江苏艾萨克机器人股份有限公司 A kind of hotel service robot
CN109993945A (en) * 2019-04-04 2019-07-09 清华大学 For gradually freezing the alarm system and alarm method of disease patient monitoring
CN110232922A (en) * 2019-04-29 2019-09-13 北京云迹科技有限公司 Robot voice halts control method and device
CN110228065A (en) * 2019-04-29 2019-09-13 北京云迹科技有限公司 Motion planning and robot control method and device
CN110096154A (en) * 2019-05-08 2019-08-06 北京百度网讯科技有限公司 For handling the method and device of information
CN110202592A (en) * 2019-07-02 2019-09-06 江苏博子岛智能产业技术研究院有限公司 A kind of AI mobile medical service robot
WO2021000657A1 (en) * 2019-07-02 2021-01-07 江苏博子岛智能产业技术研究院有限公司 Ai mobile medical service robot
CN110797017A (en) * 2019-07-30 2020-02-14 深圳市南和移动通信科技股份有限公司 Voice help calling method, intelligent sound box and storage medium
CN112568692A (en) * 2019-09-29 2021-03-30 浙江苏泊尔家电制造有限公司 Control method of cooking appliance, cooking appliance and computer storage medium
CN110827931A (en) * 2020-01-13 2020-02-21 四川大学华西医院 Method and device for managing clinical terms and readable storage medium
CN111324206B (en) * 2020-02-28 2023-07-18 重庆百事得大牛机器人有限公司 System and method for identifying confirmation information based on gesture interaction
CN111324206A (en) * 2020-02-28 2020-06-23 重庆百事得大牛机器人有限公司 Gesture interaction-based confirmation information identification system and method
CN113359538A (en) * 2020-03-05 2021-09-07 东元电机股份有限公司 Voice control robot
CN111382723A (en) * 2020-03-30 2020-07-07 北京云住养科技有限公司 Method, device and system for identifying help
US11654573B2 (en) 2020-05-29 2023-05-23 Tata Consultancy Services Limited Methods and systems for enabling human robot interaction by sharing cognition
WO2022001120A1 (en) * 2020-06-30 2022-01-06 江苏科技大学 Multi-agent system and control method therefor
CN111931605A (en) * 2020-07-23 2020-11-13 中核核电运行管理有限公司 Intelligent monitoring system and method for high-risk operation of nuclear power plant
CN111931605B (en) * 2020-07-23 2024-05-14 中核核电运行管理有限公司 Nuclear power plant high-risk operation intelligent monitoring system and method
CN112099632A (en) * 2020-09-16 2020-12-18 济南大学 Human-robot cooperative interaction method for assistant accompanying
CN112099632B (en) * 2020-09-16 2024-04-05 济南大学 Human-robot cooperative interaction method for helping old accompany
CN112224304A (en) * 2020-10-28 2021-01-15 北京理工大学 Wheel step composite mobile platform and gesture and voice control method thereof
CN112866064A (en) * 2021-01-04 2021-05-28 欧普照明电器(中山)有限公司 Control method, control system and electronic equipment
CN113192503A (en) * 2021-04-28 2021-07-30 深圳市金画王技术有限公司 Intelligent voice control system of water surface lifesaving robot
CN113341820A (en) * 2021-06-16 2021-09-03 江苏纬信工程咨询有限公司 Intelligent construction site safety monitoring device based on Internet of things and monitoring method thereof
CN113848790A (en) * 2021-09-28 2021-12-28 德州学院 Intelligent nursing type robot system and control method thereof

Also Published As

Publication number Publication date
CN105468145B (en) 2019-05-28

Similar Documents

Publication Publication Date Title
CN105468145A (en) Robot man-machine interaction method and device based on gesture and voice recognition
EP3392181B1 (en) A holographic elevator assistance system
CN106406119B (en) Service robot based on interactive voice, cloud and integrated intelligent Household monitor
Yoshimi et al. Development of a person following robot with vision based target detection
TWI469910B (en) Control method and device of a simple node transportation system
Gross et al. Living with a mobile companion robot in your own apartment-final implementation and results of a 20-weeks field study with 20 seniors
US11580385B2 (en) Artificial intelligence apparatus for cleaning in consideration of user's action and method for the same
US20200008639A1 (en) Artificial intelligence monitoring device and method of operating the same
US11307593B2 (en) Artificial intelligence device for guiding arrangement location of air cleaning device and operating method thereof
US20190392819A1 (en) Artificial intelligence device for providing voice recognition service and method of operating the same
CN109093633A (en) A kind of detachable robot and its control method
US11607801B2 (en) Artificial intelligence robot for managing movement of object using artificial intelligence and method of operating the same
KR20190104490A (en) Artificial intelligence apparatus and method for recognizing utterance voice of user
US11653805B2 (en) Robot cleaner for performing cleaning using artificial intelligence and method of operating the same
KR20160072621A (en) Artificial intelligence robot service system
US11748614B2 (en) Artificial intelligence device for controlling external device
CN208930273U (en) A kind of detachable robot
KR20190084912A (en) Artificial intelligence device that can be controlled according to user action
CN111300429A (en) Robot control system, method and readable storage medium
CN102895093A (en) Walker aid robot tracking system and walker aid robot tracking method based on RGB-D (red, green and blue-depth) sensor
CN111098307A (en) Intelligent patrol robot
KR20210020312A (en) Robot and method for controlling same
CN116572260A (en) Emotion communication accompanying and nursing robot system based on artificial intelligence generated content
CN115136174A (en) Augmented reality customer identification system and method
US20210401255A1 (en) Artificial intelligence robot and method of operating the same

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant