CN108406848A - A kind of intelligent robot and its motion control method based on scene analysis - Google Patents

A kind of intelligent robot and its motion control method based on scene analysis Download PDF

Info

Publication number
CN108406848A
CN108406848A CN201810210328.4A CN201810210328A CN108406848A CN 108406848 A CN108406848 A CN 108406848A CN 201810210328 A CN201810210328 A CN 201810210328A CN 108406848 A CN108406848 A CN 108406848A
Authority
CN
China
Prior art keywords
module
information
detection
scene
robot
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810210328.4A
Other languages
Chinese (zh)
Inventor
刘阳
赵强
张银磊
吴雄辉
佀昶
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui Fruit Intelligent Technology Co Ltd
Original Assignee
Anhui Fruit Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui Fruit Intelligent Technology Co Ltd filed Critical Anhui Fruit Intelligent Technology Co Ltd
Priority to CN201810210328.4A priority Critical patent/CN108406848A/en
Publication of CN108406848A publication Critical patent/CN108406848A/en
Pending legal-status Critical Current

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J19/00Accessories fitted to manipulators, e.g. for monitoring, for viewing; Safety devices combined with or specially adapted for use in connection with manipulators
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J19/00Accessories fitted to manipulators, e.g. for monitoring, for viewing; Safety devices combined with or specially adapted for use in connection with manipulators
    • B25J19/02Sensing devices
    • B25J19/04Viewing devices
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1679Programme controls characterised by the tasks executed
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Mechanical Engineering (AREA)
  • Robotics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Manipulator (AREA)

Abstract

The invention discloses a kind of intelligent robots and its motion control method based on scene analysis.By vision, the sense of hearing, tactile module, the information in current scene is obtained, such as:Human body information, object information, user speech information etc..Each module data is uniformly aggregated into scene analysis module, and information priorities arrangement and the fusion of several scenes information are carried out by scene analysis module.Scene information module by analysis after, complete modeling to current environment, and the generation of corresponding movement instruction, then pass to movement output module, send commands to each execution unit, carry out the execution of corresponding sports.The present invention is completed by the fusion of multisensor to the identification of scene rather than single vision or speech recognition.To the identification of people in scene Recognition, the identification to object, and scene information is uniformly synthesized to the identification of environment, enhance the adaptability of identification.Enhance the independence and interactivity of robot.

Description

A kind of intelligent robot and its motion control method based on scene analysis
Technical field
The present invention relates to robotic technology field more particularly to a kind of intelligent robots and its movement based on scene analysis Control method.
Background technology
Tool of the robot as neither one sense of independence can only be used for solving the very high work of some standardization level Make, the possibility that more working robot of single action substitutes is higher, and initial application is namely in automobile making.By The development of decades, robot technology constantly improve, intelligence degree is also higher and higher, and application is also extended to from manufacturing industry Consumption market.
Consumer level robot, form that there are two types of mainstreams:One kind is controlled based on movenent performance with mobile phone or controller System;Another kind is based on intelligence system, but most is to be mounted in common trolley or fixed structure.At present on the market There is not yet intelligently to interact with people, but also with the consumer level robot of good autokinetic movement performance.
Invention content
In view of the drawbacks described above of the prior art, technical problem to be solved by the invention is to provide one kind based on scene point The intelligent robot and its motion control method of analysis, respectively energy utonomous working, manual operation and human-computer interaction, intelligence degree Height can make corresponding reaction according to the difference of ambient enviroment.
To achieve the above object, the present invention provides a kind of intelligent robots based on scene analysis, including robot to regard Feel module, sense of hearing module, tactile module, scene analysis module, movement output module, the vision module, sense of hearing module, tactile Module is connect with scene analysis module, and the scene analysis module output end is connect with movement output module, wherein:Vision mould Block includes sequentially connected image reading submodule, image procossing submodule, object detection sub-module;Sense of hearing module include according to The environmental noise processing submodule of secondary connection, current signature detection submodule, auditory information collects submodule;Tactile module includes Sequentially connected tactile sensing submodule, tactile data summarizing module;Scene analysis module is believed comprising sequentially connected sensor Breath fusion submodule, priority screening submodule, action generate submodule.
Further, the object detection sub-module of the vision module include human detection module, object detection module, Detection of obstacles module and the visual information being connect with human detection module, object detection module, detection of obstacles module are converged Total module.
Further, the tactile sensing submodule of the tactile module includes pressure sensor and touch sensor, pressure For sensor arrangement in end effector of robot, touch sensor is arranged in robot matrix surface.
Further, the movement output module includes motor and detects the position sensor of motor position.
A kind of intelligent robot motion control method based on scene analysis, it is characterised in that:
Vision module obtains image information, including:Detection and people relevant information, detection other objects in addition to the human body It whether there is obstacle in front of information, detection robot;
Sense of hearing module obtains auditory information, including:By the voice signal filter preprocessing of the environment of detection, detection sound point Class, and semantic information is extracted to the sound of different classifications;
Tactile module obtains tactile data, including:Robot captures the information of object, and the information whether touched;
Scene analysis module synthesis vision module, sense of hearing module, tactile module information, are merged, priority arrangement, are led to It crosses deep neural network model and generates Scene Semantics information, according to Scene Semantics information, generate robot corresponding actions;
Robot motion is distributed to the movement of each motor by movement output module.
Further, the vision module obtains image information, specifically includes:
Camera constantly updates the image data of buffer area with the rate capture images of 30fps, later modules with The form of multithreading, while calling same frame image in buffer area;
Human detection module predominantly detects and the relevant information of people;First detect picture in whether someone, if nobody, Directly return the result;If someone, the posture and gesture information of human body are detected;Detect whether face simultaneously, if there is Face then detect the including but not limited to expression of face, face whether recognize, face gender, the information at face age;Later, will The information of all human detection modules is transmitted to visual information summarizing module;
Object detection module predominantly detects the information of other objects in addition to the human body, and system prestores the spy of familiar object It levies and marked, object detection module scans in the image collected, looks for whether that there are marked objects, such as Fruit detects market object, and the object number detected and the location information in picture, which are transmitted to visual information, converges Total module;
Obstacle detection module, which predominantly detects, whether there is obstacle in front of robot, this module according to image continuity information, Whether there is doubtful barrier in detection image, and returns to its position and size information to visual information in the picture and summarize mould Block;
Visual information summarizing module receives human detection module, object detection module, detection of obstacles module and inputs Information after information sorting, is transmitted to scene analysis module and is analyzed according to certain format.
Further, the sense of hearing module obtains auditory information, and specific steps include:
After microphone captures the voice signal of environment, environmental noise processing submodule filter preprocessing behaviour is first passed around Make, filters out ambient noise;
Voice and other stored sound are detected by current signature detection submodule;
For stored specific sound, corresponding semantic information is stored in system, which is sent directly to Auditory information collects submodule;For voice, extract what voice included according to the characteristic model of current signature detection submodule Text information is sent to auditory information collects submodule by keyword;
Auditory information collects submodule summarizes all acoustic informations, and scene analysis module is sent to according to fixed format It is analyzed.
Further, the tactile module obtains tactile data, specifically includes:
Pressure sensor is arranged in end effector of robot, by information of voltage judge paw whether capture object and The weight of object;
Touch sensor is arranged in robot matrix surface, by judging that the voltage pulse at each position changes, obtains machine Whether device people is touched and touching position information.
Further, the scene analysis module synthesis vision module, sense of hearing module, tactile module information, merged, Priority arrangement generates Scene Semantics information by deep neural network model, according to Scene Semantics information, generates robot phase It should act, specially:
Vision module, sense of hearing module, tactile module remember timestamp, scene before handling data, in pending data subscript Analysis module is after obtaining information, first by timestamp verification data, and all data is pressed newest time synchronization;
After time synchronization, by there are the progress of the information of correlation after vision module, sense of hearing module, tactile module detection process Fusion;
To after fusion information carry out priority arrangement, according to priority from high to low be people, animal, barrier, other The priority orders of object screen fuse information;
After being screened to fuse information according to priority orders, by advance trained deep neural network model, generate Current scene semantic information is converted to the instruction of robot execution action according still further to certain rule, is sent to action and generates son Module.
The beneficial effects of the invention are as follows:
1, it is completed to the identification of scene rather than single vision or speech recognition by the fusion of multisensor.
2, after obtaining scene information, robot can make reaction on limb action rather than single voice or Screen reacts.
3, to the identification of people in scene Recognition, the identification to object, and scene information is uniformly synthesized to the identification of environment, Enhance the adaptability of identification.
4, the independence and interactivity of robot are enhanced.
The technique effect of the design of the present invention, concrete structure and generation is described further below with reference to attached drawing, with It is fully understood from the purpose of the present invention, feature and effect.
Description of the drawings
Fig. 1 is the overall structure block diagram of the present invention.
Fig. 2 is the vision module work flow diagram of the present invention.
Fig. 3 is the sense of hearing module work flow diagram of the present invention.
Fig. 4 is the tactile module work flow diagram of the present invention.
Fig. 5 is the scene analysis module work flow diagram of the present invention.
Fig. 6 is the movement output module work flow diagram of the present invention.
Specific implementation mode
As shown in Figure 1, a kind of intelligent robot based on scene analysis, including robot vision module, sense of hearing module, touch Feel module, scene analysis module, movement output module, the vision module, sense of hearing module, tactile module with scene analysis mould Block connects, and the scene analysis module output end is connect with movement output module, wherein:Vision module includes sequentially connected figure As reading submodule, image procossing submodule, object detection sub-module;Sense of hearing module includes at sequentially connected environmental noise Manage submodule, current signature detection submodule, auditory information collects submodule;Tactile module includes sequentially connected tactile sensing Submodule, tactile data summarizing module;Scene analysis module includes sequentially connected sensor data fusion submodule, priority Screen submodule, action generates submodule.
In the present embodiment, the object detection sub-module of the vision module includes human detection module, object detection mould Block, detection of obstacles module and the visual information being connect with human detection module, object detection module, detection of obstacles module Summarizing module.
In the present embodiment, the tactile sensing submodule of the tactile module includes pressure sensor and touch sensor, pressure Force snesor is arranged in end effector of robot, and touch sensor is arranged in robot matrix surface.
In the present embodiment, the movement output module includes motor and detects the position sensor of motor position.
As shown in figures 2-6, a kind of intelligent robot motion control method based on scene analysis:
Vision module obtains image information, including:Detection and people relevant information, detection other objects in addition to the human body It whether there is obstacle in front of information, detection robot;
Sense of hearing module obtains auditory information, including:By the voice signal filter preprocessing of the environment of detection, detection sound point Class, and semantic information is extracted to the sound of different classifications;
Tactile module obtains tactile data, including:Robot captures the information of object, and the information whether touched;
Scene analysis module synthesis vision module, sense of hearing module, tactile module information, are merged, priority arrangement, are led to It crosses deep neural network model and generates Scene Semantics information, according to Scene Semantics information, generate robot corresponding actions;
Robot motion is distributed to the movement of each motor by movement output module.
In the present embodiment, the vision module obtains image information, specifically includes:
Camera constantly updates the image data of buffer area with the rate capture images of 30fps, later modules with The form of multithreading, while calling same frame image in buffer area;
Human detection module predominantly detects and the relevant information of people;First detect picture in whether someone, if nobody, Directly return the result;If someone, the posture and gesture information of human body are detected;Detect whether face simultaneously, if there is Face then detect the including but not limited to expression of face, face whether recognize, face gender, the information at face age;Later, will The information of all human detection modules is transmitted to visual information summarizing module;
Object detection module predominantly detects the information of other objects in addition to the human body, and system prestores the spy of familiar object It levies and marked, object detection module scans in the image collected, looks for whether that there are marked objects, such as Fruit detects market object, and the object number detected and the location information in picture, which are transmitted to visual information, converges Total module;
Obstacle detection module, which predominantly detects, whether there is obstacle in front of robot, this module according to image continuity information, Whether there is doubtful barrier in detection image, and returns to its position and size information to visual information in the picture and summarize mould Block;
Visual information summarizing module receives human detection module, object detection module, detection of obstacles module and inputs Information after information sorting, is transmitted to scene analysis module and is analyzed according to certain format.
In the present embodiment, the sense of hearing module obtains auditory information, and specific steps include:
After microphone captures the voice signal of environment, environmental noise processing submodule filter preprocessing behaviour is first passed around Make, filters out ambient noise;
Voice and other stored sound are detected by current signature detection submodule;
For stored specific sound, corresponding semantic information is stored in system, which is sent directly to Auditory information collects submodule;For voice, extract what voice included according to the characteristic model of current signature detection submodule Text information is sent to auditory information collects submodule by keyword;
Auditory information collects submodule summarizes all acoustic informations, and scene analysis module is sent to according to fixed format It is analyzed.
In the present embodiment, the tactile module obtains tactile data, specifically includes:
Pressure sensor is arranged in end effector of robot, by information of voltage judge paw whether capture object and The weight of object;
Touch sensor is arranged in robot matrix surface, by judging that the voltage pulse at each position changes, obtains machine Whether device people is touched and touching position information.
In the present embodiment, the scene analysis module synthesis vision module, sense of hearing module, tactile module information are melted It closes, priority arrangement, Scene Semantics information is generated by deep neural network model, according to Scene Semantics information, generate machine People's corresponding actions, specially:
Vision module, sense of hearing module, tactile module remember timestamp, scene before handling data, in pending data subscript Analysis module is after obtaining information, first by timestamp verification data, and all data is pressed newest time synchronization;
After time synchronization, by there are the progress of the information of correlation after vision module, sense of hearing module, tactile module detection process Fusion;
To after fusion information carry out priority arrangement, according to priority from high to low be people, animal, barrier, other The priority orders of object screen fuse information;
After being screened to fuse information according to priority orders, by advance trained deep neural network model, generate Current scene semantic information is converted to the instruction of robot execution action according still further to certain rule, is sent to action and generates son Module.
The detailed description below principle of the invention:
As shown in Fig. 2, the information that vision module obtains includes:In picture whether someone, whether had robot in picture Object through memory.Such as someone, if there is face.If any face, whether face has recognized, if has mood.People and Posture of the target object relative to robot.
Visual processes are primarily referred to as capturing image to the process for obtaining semantic information in image from camera.First, it takes the photograph As head is with the rate capture images of 30fps, and constantly update the image data of buffer area.Later, modules are with multithreading Form, while calling same frame image in buffer area.
Human detection module predominantly detects and the relevant information of people.First detect picture in whether someone, if nobody, Directly return the result.If someone, the posture and gesture information of human body are detected.Detect whether face simultaneously, if there is Face then detects the expression of face, and whether face recognizes, face gender, the information such as face age.Later, by all human testings The information of module is transmitted to visual information summarizing module.
Object detection module predominantly detects the information of other objects in addition to the human body.System is advance and stores common object The feature of body by scanning on the image, looks for whether there are marked object, such as pet in object detection module, Flowerpot, desk, dustbin etc..Later, the object number that will be detected, and location information in picture are transmitted to vision letter Cease summarizing module.
Obstacle detection module, which predominantly detects, whether there is obstacle in front of robot.This module according to image continuity information, Whether there is doubtful barrier in detection image, and returns to its position and size information to visual information in the picture and summarize mould Block.
Visual information summarizing module receives the information of modules input, according to certain format, after information sorting, Scene analysis module is transmitted to be analyzed.
As shown in figure 3, the information that sense of hearing module obtains includes:Whether sound letter that robot remembered is had in environment Number, if having whether human voice signal, voice signal contain semantic information.Auditory processing refers mainly to be collected into sound from microphone Data are to the process for getting in sound semantic and text message.
After microphone captures the voice signal of environment, the pretreatment operations such as filtering are first passed around, environment is filtered out and makes an uproar Sound detects voice and other stored sound by specific frequency later (such as mewing is barked).For stored Specific sound has stored corresponding semantic information in system, which is sent directly to auditory information summarizing module.For Voice, the keyword that voice includes is extracted according to characteristic model, and text information is sent to auditory information summarizing module.It Auditory information summarizing module summarizes all information afterwards, and being sent to scene analysis module according to fixed format is analyzed.
As shown in figure 4, the information that tactile module obtains includes:Whether robot is touched, and the number touched is with timely Between.Tactile processing refers to that the data obtained from pressure sensor and touch sensor obtain the object being in direct contact with robot Information.
Pressure sensor is arranged in end effector of robot, i.e. robot hand.For judging hand by information of voltage Whether pawl captures object, the weight etc. of object.
Touch sensor is arranged in robot matrix surface, by judging that the voltage pulse at each position changes, obtains machine Whether device people is touched and touching position information.
Tactile data summarizing module will capture information and touch information summarizes, and scene analysis is sent to according to fixed format Module is analyzed.
As shown in figure 5, the effect of scene analysis module is to integrate each module information, carries out priority arrangement and merge; Judge residing scene at present by intelligent algorithm;According to scene serial number, corresponding actions are generated.
Scene analysis is primarily referred to as summarizing the information that vision, the sense of hearing, tactile obtain, and merges, screening, and final Go out the process of the semantic description best to current scene.
The speed of each module process data is different, therefore is passed to the vision of scene analysis module simultaneously, the sense of hearing, tactile letter It ceases asynchronous.In order to solve this problem, each module all can remember timestamp before handling data in pending data subscript. Scene analysis module is after obtaining information, first by timestamp verification data, and all data is pressed newest time synchronization.
Since the information in modules can have correlation, so to be merged to information.As detected in picture There is happy expression, while hearing there is laugh, this two information will be fused to happy mood simultaneously.
Meanwhile the information that modules obtain is very more, can not be handled in real time, so to carry out priority to information Arrangement.After each module information carries out synthesis, people, animal, barrier, other four classes of object, later according to people can be divided into>Animal >Barrier>The priority of other objects selects information.
According to priority to information sifting after, by advance trained deep neural network model, generate according to subject+ One information of predicate (+object).Such as Ren Mo robots, people laughs at, and barrier occurs etc..Later by current scene semantic information, Robot is converted to according to certain rule and needs the instruction executed, is sent to execution module.
As shown in fig. 6, movement output is mainly that the robot motion instruction that other modules generate is transmitted to motion control Device, motion controller generate the electrical parameter needed for motor and pass to actuating motor, and motor drives executing agency to complete corresponding machine Human action.
In order to allow robot precise motion, each motor all to carry position-detection sensor, acquire the real-time appearance of robot State information.Since robot different parts deposit relevance, in order to ensure that robot motion is consistent with expection, motion controller is also wanted Export the priority time of each actuating motor work.
To sum up, the present invention has the advantage that:
1, it is completed to the identification of scene rather than single vision or speech recognition by the fusion of multisensor.
2, after obtaining scene information, robot can make reaction on limb action rather than single voice or Screen reacts.
3, to the identification of people in scene Recognition, the identification to object, and scene information is uniformly synthesized to the identification of environment, Enhance the adaptability of identification.
4, the independence and interactivity of robot are enhanced.
The preferred embodiment of the present invention has been described in detail above.It should be appreciated that those skilled in the art without It needs creative work according to the present invention can conceive and makes many modifications and variations.Therefore, all technologies in the art Personnel are available by logical analysis, reasoning, or a limited experiment on the basis of existing technology under this invention's idea Technical solution, all should be in the protection domain being defined in the patent claims.

Claims (9)

1. a kind of intelligent robot based on scene analysis, which is characterized in that including robot vision module, sense of hearing module, touch Feel module, scene analysis module, movement output module, the vision module, sense of hearing module, tactile module with scene analysis mould Block connects, and the scene analysis module output end is connect with movement output module, wherein:Vision module includes sequentially connected figure As reading submodule, image procossing submodule, object detection sub-module;Sense of hearing module includes at sequentially connected environmental noise Manage submodule, current signature detection submodule, auditory information collects submodule;Tactile module includes sequentially connected tactile sensing Submodule, tactile data summarizing module;Scene analysis module includes sequentially connected sensor data fusion submodule, priority Screen submodule, action generates submodule.
2. a kind of intelligent robot based on scene analysis as described in claim 1, it is characterised in that:The vision module Object detection sub-module include human detection module, object detection module, detection of obstacles module and with human testing mould Block, object detection module, the visual information summarizing module of detection of obstacles module connection.
3. a kind of intelligent robot based on scene analysis as described in claim 1, it is characterised in that:The tactile module Tactile sensing submodule includes pressure sensor and touch sensor, and pressure sensor is arranged in end effector of robot, is touched Sensor arrangement is touched in robot matrix surface.
4. a kind of intelligent robot based on scene analysis as described in claim 1, it is characterised in that:The movement output mould Block includes motor and detects the position sensor of motor position.
5. a kind of intelligent robot motion control method based on scene analysis, it is characterised in that:
Vision module obtains image information, including:Detection and the relevant information of people, the letter for detecting other objects in addition to the human body It whether there is obstacle in front of breath, detection robot;
Sense of hearing module obtains auditory information, including:By the voice signal filter preprocessing of the environment of detection, sound classification is detected, And semantic information is extracted to the sound of different classifications;
Tactile module obtains tactile data, including:Robot captures the information of object, and the information whether touched;
Scene analysis module synthesis vision module, sense of hearing module, tactile module information, are merged, priority arrangement, pass through depth It spends neural network model and generates Scene Semantics information, according to Scene Semantics information, generate robot corresponding actions;
Robot motion is distributed to the movement of each motor by movement output module.
6. a kind of intelligent robot motion control method based on scene analysis as claimed in claim 5, which is characterized in that institute It states vision module and obtains image information, specifically include:
Camera constantly updates the image data of buffer area with the rate capture images of 30fps, and modules are with multi-thread later The form of journey, while calling same frame image in buffer area;
Human detection module predominantly detects and the relevant information of people;First detect picture in whether someone, if nobody, directly It returns the result;If someone, the posture and gesture information of human body are detected;Face is detected whether simultaneously, if there is face Then detection include but not limited to the expression of face, face whether recognize, face gender, the information at face age;Later, will own The information of human detection module is transmitted to visual information summarizing module;
Object detection module predominantly detects the information of other objects in addition to the human body, and system prestores the feature of familiar object simultaneously And it is marked, object detection module scans in the image collected, looks for whether there are marked object, if inspection Market object is measured, the object number detected and the location information in picture, which are transmitted to visual information, summarizes mould Block;
Obstacle detection module, which predominantly detects, whether there is obstacle in front of robot, this module is according to image continuity information, detection Whether there is doubtful barrier in image, and returns to its position and size information in the picture to visual information summarizing module;
Visual information summarizing module receives human detection module, object detection module, the information of detection of obstacles module input, According to certain format, after information sorting, it is transmitted to scene analysis module and is analyzed.
7. a kind of intelligent robot motion control method based on scene analysis as claimed in claim 5, which is characterized in that institute It states sense of hearing module and obtains auditory information, specific steps include:
After microphone captures the voice signal of environment, environmental noise processing submodule filter preprocessing operation is first passed around, Filter out ambient noise;
Voice and other stored sound are detected by current signature detection submodule;
For stored specific sound, corresponding semantic information is stored in system, which is sent directly to the sense of hearing Information collects submodule;For voice, the key that voice includes is extracted according to the characteristic model of current signature detection submodule Text information is sent to auditory information collects submodule by word;
Auditory information collects submodule summarizes all acoustic informations, and being sent to scene analysis module according to fixed format carries out Analysis.
8. a kind of intelligent robot motion control method based on scene analysis as claimed in claim 5, which is characterized in that institute It states tactile module and obtains tactile data, specifically include:
Pressure sensor is arranged in end effector of robot, judges whether paw captures object and object by information of voltage Weight;
Touch sensor is arranged in robot matrix surface, by judging that the voltage pulse at each position changes, obtains robot Whether touched and touching position information.
9. a kind of intelligent robot motion control method based on scene analysis as claimed in claim 5, which is characterized in that institute Scene analysis module synthesis vision module, sense of hearing module, tactile module information are stated, is merged, priority arrangement, passes through depth Neural network model generates Scene Semantics information, according to Scene Semantics information, generates robot corresponding actions, specially:
Vision module, sense of hearing module, tactile module remember timestamp, scene analysis before handling data, in pending data subscript Module is after obtaining information, first by timestamp verification data, and all data is pressed newest time synchronization;
After time synchronization, by there are the information of correlation to melt after vision module, sense of hearing module, tactile module detection process It closes;
Priority arrangement is carried out to the information after fusion, is people, animal, barrier, other objects from high to low according to priority Priority orders screen fuse information;
After being screened to fuse information according to priority orders, by advance trained deep neural network model, generate current Scene Semantics information is converted to the instruction of robot execution action according still further to certain rule, is sent to action and generates submodule.
CN201810210328.4A 2018-03-14 2018-03-14 A kind of intelligent robot and its motion control method based on scene analysis Pending CN108406848A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810210328.4A CN108406848A (en) 2018-03-14 2018-03-14 A kind of intelligent robot and its motion control method based on scene analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810210328.4A CN108406848A (en) 2018-03-14 2018-03-14 A kind of intelligent robot and its motion control method based on scene analysis

Publications (1)

Publication Number Publication Date
CN108406848A true CN108406848A (en) 2018-08-17

Family

ID=63131508

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810210328.4A Pending CN108406848A (en) 2018-03-14 2018-03-14 A kind of intelligent robot and its motion control method based on scene analysis

Country Status (1)

Country Link
CN (1) CN108406848A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109605403A (en) * 2019-01-25 2019-04-12 北京番茄时代科技有限公司 Robot and its operating system, control device, control method and storage medium
CN109741747A (en) * 2019-02-19 2019-05-10 珠海格力电器股份有限公司 Voice scene recognition method and device, sound control method and equipment, air-conditioning
CN110861853A (en) * 2019-11-29 2020-03-06 三峡大学 Intelligent garbage classification method combining vision and touch
CN111098307A (en) * 2019-12-31 2020-05-05 航天信息股份有限公司 Intelligent patrol robot
CN111357011A (en) * 2019-01-31 2020-06-30 深圳市大疆创新科技有限公司 Environment sensing method and device, control method and device and vehicle
CN111604899A (en) * 2020-05-15 2020-09-01 深圳国信泰富科技有限公司 Data transmission system of high intelligent robot
CN111618856A (en) * 2020-05-27 2020-09-04 山东交通学院 Robot control method and system based on visual excitation points and robot
CN111723598A (en) * 2019-03-18 2020-09-29 北京邦天信息技术有限公司 Machine vision system and implementation method thereof
CN112578909A (en) * 2020-12-15 2021-03-30 北京百度网讯科技有限公司 Equipment interaction method and device
CN112825014A (en) * 2019-11-21 2021-05-21 王炼 Artificial intelligence brain
CN113082268A (en) * 2021-03-12 2021-07-09 浙江创力电子股份有限公司 Handheld sterilizer of networking based on 4G
CN113867163A (en) * 2021-10-09 2021-12-31 深圳康佳电子科技有限公司 Intelligent household scene switching method and device, intelligent terminal and storage medium
WO2022111443A1 (en) * 2020-11-26 2022-06-02 苏州中科先进技术研究院有限公司 Intelligent blind assisting system and method, computer device, and storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060126918A1 (en) * 2004-12-14 2006-06-15 Honda Motor Co., Ltd. Target object detection apparatus and robot provided with the same
CN105118518A (en) * 2015-07-15 2015-12-02 百度在线网络技术(北京)有限公司 Sound semantic analysis method and device
CN105598972A (en) * 2016-02-04 2016-05-25 北京光年无限科技有限公司 Robot system and interactive method
CN105700438A (en) * 2016-03-18 2016-06-22 北京光年无限科技有限公司 Electronic control system for multi-joint small robot
CN105912725A (en) * 2016-05-12 2016-08-31 上海劲牛信息技术有限公司 System for calling vast intelligence applications through natural language interaction
CN106529375A (en) * 2015-09-11 2017-03-22 上海乐今通信技术有限公司 Mobile terminal and object feature identification method for image of mobile terminal
CN106570491A (en) * 2016-11-11 2017-04-19 华南智能机器人创新研究院 Robot intelligent interaction method and intelligent robot
CN106821694A (en) * 2017-01-18 2017-06-13 西南大学 A kind of mobile blind guiding system based on smart mobile phone
CN106997236A (en) * 2016-01-25 2017-08-01 亮风台(上海)信息科技有限公司 Based on the multi-modal method and apparatus for inputting and interacting

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060126918A1 (en) * 2004-12-14 2006-06-15 Honda Motor Co., Ltd. Target object detection apparatus and robot provided with the same
CN105118518A (en) * 2015-07-15 2015-12-02 百度在线网络技术(北京)有限公司 Sound semantic analysis method and device
CN106529375A (en) * 2015-09-11 2017-03-22 上海乐今通信技术有限公司 Mobile terminal and object feature identification method for image of mobile terminal
CN106997236A (en) * 2016-01-25 2017-08-01 亮风台(上海)信息科技有限公司 Based on the multi-modal method and apparatus for inputting and interacting
CN105598972A (en) * 2016-02-04 2016-05-25 北京光年无限科技有限公司 Robot system and interactive method
CN105700438A (en) * 2016-03-18 2016-06-22 北京光年无限科技有限公司 Electronic control system for multi-joint small robot
CN105912725A (en) * 2016-05-12 2016-08-31 上海劲牛信息技术有限公司 System for calling vast intelligence applications through natural language interaction
CN106570491A (en) * 2016-11-11 2017-04-19 华南智能机器人创新研究院 Robot intelligent interaction method and intelligent robot
CN106821694A (en) * 2017-01-18 2017-06-13 西南大学 A kind of mobile blind guiding system based on smart mobile phone

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109605403A (en) * 2019-01-25 2019-04-12 北京番茄时代科技有限公司 Robot and its operating system, control device, control method and storage medium
CN109605403B (en) * 2019-01-25 2020-12-11 北京妙趣伙伴科技有限公司 Robot, robot operating system, robot control device, robot control method, and storage medium
CN111357011A (en) * 2019-01-31 2020-06-30 深圳市大疆创新科技有限公司 Environment sensing method and device, control method and device and vehicle
CN111357011B (en) * 2019-01-31 2024-04-30 深圳市大疆创新科技有限公司 Environment sensing method and device, control method and device and vehicle
CN109741747A (en) * 2019-02-19 2019-05-10 珠海格力电器股份有限公司 Voice scene recognition method and device, sound control method and equipment, air-conditioning
CN111723598A (en) * 2019-03-18 2020-09-29 北京邦天信息技术有限公司 Machine vision system and implementation method thereof
CN112825014A (en) * 2019-11-21 2021-05-21 王炼 Artificial intelligence brain
CN110861853A (en) * 2019-11-29 2020-03-06 三峡大学 Intelligent garbage classification method combining vision and touch
CN111098307A (en) * 2019-12-31 2020-05-05 航天信息股份有限公司 Intelligent patrol robot
CN111604899A (en) * 2020-05-15 2020-09-01 深圳国信泰富科技有限公司 Data transmission system of high intelligent robot
CN111618856B (en) * 2020-05-27 2021-11-05 山东交通学院 Robot control method and system based on visual excitation points and robot
CN111618856A (en) * 2020-05-27 2020-09-04 山东交通学院 Robot control method and system based on visual excitation points and robot
WO2022111443A1 (en) * 2020-11-26 2022-06-02 苏州中科先进技术研究院有限公司 Intelligent blind assisting system and method, computer device, and storage medium
CN112578909A (en) * 2020-12-15 2021-03-30 北京百度网讯科技有限公司 Equipment interaction method and device
CN112578909B (en) * 2020-12-15 2024-05-31 北京百度网讯科技有限公司 Method and device for equipment interaction
CN113082268A (en) * 2021-03-12 2021-07-09 浙江创力电子股份有限公司 Handheld sterilizer of networking based on 4G
CN113867163A (en) * 2021-10-09 2021-12-31 深圳康佳电子科技有限公司 Intelligent household scene switching method and device, intelligent terminal and storage medium

Similar Documents

Publication Publication Date Title
CN108406848A (en) A kind of intelligent robot and its motion control method based on scene analysis
EP2877909B1 (en) Multimodal interaction with near-to-eye display
CN105739688A (en) Man-machine interaction method and device based on emotion system, and man-machine interaction system
CN104956292A (en) Interaction of multiple perceptual sensing inputs
CN101825947A (en) Method and device for intelligently controlling mobile terminal and mobile terminal thereof
US11825278B2 (en) Device and method for auto audio and video focusing
CN107944434A (en) A kind of alarm method and terminal based on rotating camera
Tsai et al. Spatial temporal variation graph convolutional networks (STV-GCN) for skeleton-based emotional action recognition
Wachs et al. Real-time hand gesture telerobotic system using fuzzy c-means clustering
KR20160072621A (en) Artificial intelligence robot service system
CN114821753B (en) Eye movement interaction system based on visual image information
CN113377193A (en) Vending machine interaction method and system based on reliable gesture recognition
CN105867595A (en) Human-machine interaction mode combing voice information with gesture information and implementation device thereof
CN114255508A (en) OpenPose-based student posture detection analysis and efficiency evaluation method
Joslin et al. Dynamic gesture recognition
Kheratkar et al. Gesture controlled home automation using CNN
Sisodia et al. Image pixel intensity and artificial neural network based method for pattern recognition
Dhamanskar et al. Human computer interaction using hand gestures and voice
Srisuphab et al. Artificial neural networks for gesture classification with inertial motion sensing armbands
Simão et al. Unsupervised gesture segmentation of a real-time data stream in MATLAB
Sawaragi et al. Self-reflective segmentation of human bodily motions using recurrent neural networks
CN113894779A (en) Multi-mode data processing method applied to robot interaction
CN113807280A (en) Kinect-based virtual ship cabin system and method
Chanhan et al. Gestures based wireless robotic control using image processing
CN112308041A (en) Unmanned platform gesture control method based on vision

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180817