CN207669312U - Intelligent robot - Google Patents
Intelligent robot Download PDFInfo
- Publication number
- CN207669312U CN207669312U CN201721669686.9U CN201721669686U CN207669312U CN 207669312 U CN207669312 U CN 207669312U CN 201721669686 U CN201721669686 U CN 201721669686U CN 207669312 U CN207669312 U CN 207669312U
- Authority
- CN
- China
- Prior art keywords
- module
- robot
- intelligent robot
- laser
- power supply
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Abstract
The utility model discloses a kind of intelligent robot, which includes head and fuselage, and the intelligent robot further includes main control module and the drive module being electrically connected respectively with the main control module, obstacle avoidance module, power supply module and emergency stop module;Wherein, the drive module includes driving motor, and the drive module is for driving the robot to move forward or back;The obstacle avoidance module includes avoidance sensor, and the obstacle avoidance module is for obtaining the obstacle information in robot traveling process to get around the barrier;The power supply module is used to provide the electric energy of the robot normal work;The emergency stop module includes scram button, and the scram button is set to the robot back, and the scram button is for cutting off the robot interior power supply.The utility model intelligent robot can effectively extend the service life of robot.
Description
Technical field
The utility model is related to robot fields, and in particular to a kind of intelligent robot
Background technology
With the development of science and technology, a large amount of intelligent robot is designed and manufactured out and applied to the production of society
In life, to improve social productive forces and promote people’s lives quality.Robot can pass through infrared biography during traveling
Sensor, laser sensor, visual sensor or ultrasonic sensor detect front obstacle, are moved on getting around barrier.
However, in actual use, sensor will appear the situation of failure.As being one piece fully transparent in front of robot
Glass, at this point, can't detect the glass, robot using the robot of infrared sensor, laser sensor or visual sensor
It will move on, collide with the glass in front;And avoiding obstacles by supersonic wave is generally required through ultrasonic array, between array
Sensor works at the same time, and easy tos produce interference, is impacted to avoidance.Therefore, in the case of sensor failure, robot without
Method judges barrier existing for front, is easy to collide with barrier, causes robot to damage, influence the use of robot
Service life.
Utility model content
The main purpose of the utility model is that a kind of intelligent robot is proposed, to solve existing robot in sensor
The technical issues of can not halting in time when failure.
In order to solve the above technical problems, the utility model proposes a kind of intelligent robot, the intelligent robot, including head
And fuselage, the intelligent robot further include main control module and the drive module being electrically connected respectively with the main control module, avoidance mould
Block, power supply module and emergency stop module;Wherein,
The drive module includes driving motor, and the drive module is for driving the robot to move forward or back;
The obstacle avoidance module includes avoidance sensor, and the obstacle avoidance module is used to obtain the obstacle in robot traveling process
Object information is to get around the barrier;
The power supply module is used to provide the electric energy of the robot normal work;
The emergency stop module includes scram button, and the scram button is set to the robot back, the scram button
For cutting off the robot interior power supply.
Preferably, the avoidance sensor includes infrared sensor and/or ultrasonic sensor.
Preferably, the intelligent robot further includes the auditory localization module set on the robot head, the sound source
Locating module includes multiple microphones, and the microphone is respectively used to receive the voice and localization of sound source of different directions.
Preferably, the intelligent robot further includes the interactive module being electrically connected with the main control module, the interactive module
Including voice recognition unit, the interactive module is used to receive the speech-input instructions of user, and with voice, image or video side
Formula exports corresponding information.
Preferably, the intelligent robot further includes the photographing module being electrically connected with the main control module, the photographing module
Include depth camera unit for measuring scene depth information and profile information in visual range and common for obtaining
The main camera unit of image.
Preferably, the depth camera unit includes the laser emitter being connected with each other, laser pickoff and measuring and calculating chip;
The laser emitter is used to receive the laser of reflection, institute for emitting laser, the laser pickoff member into measurement range
Measuring and calculating chip is stated for calculating the scene depth information and profile information.
Preferably, the intelligent robot further includes the communication module being connect with the main control module, the communication module
For being connect with external equipment and internet.
Preferably, the communication module is connected by WIFI, bluetooth or ZigBee modes.
For the utility model intelligent robot in sensor failure, user can press the scram button at robot back, this
When, robot interior power supply is cut off, so that robot halts when that can not judge that traveling has barrier on the way, is avoided
Robot collides with barrier, to extend the service life of robot.
Description of the drawings
Fig. 1 is the structural schematic diagram of one embodiment of the utility model intelligent robot;
Fig. 2 is the structural schematic diagram of another embodiment of the utility model intelligent robot.
Drawing reference numeral explanation:
Specific implementation mode
The embodiments of the present invention are described below in detail, examples of the embodiments are shown in the accompanying drawings, wherein from beginning
Same or similar element or element with the same or similar functions are indicated to same or similar label eventually.Below by ginseng
The embodiment for examining attached drawing description is exemplary, it is intended to for explaining the utility model, and should not be understood as to the utility model
Limitation, be based on the embodiments of the present invention, those of ordinary skill in the art are without creative efforts
The every other embodiment obtained, shall fall within the protection scope of the present invention.
As depicted in figs. 1 and 2, the utility model proposes intelligent robot, including head 100 and fuselage 200, intelligent machine
Device people further includes main control module 300 and the drive module being electrically connected respectively with main control module 400, obstacle avoidance module 500, power supply module
600 and emergency stop module 700;Wherein,
Drive module 400 includes driving motor 410, and drive module 400 is for driving robot to move forward or back;
Obstacle avoidance module 500 includes avoidance sensor 510, and obstacle avoidance module 500 is used to obtain the barrier in robot traveling process
Hinder object information to get around barrier;
Power supply module 600 is used to provide the electric energy of robot normal work;
Emergency stop module 700 includes scram button 710, and scram button 710 is set to robot back, and scram button 710 is used for
Cut off robot internal electric source.
For the utility model intelligent robot in sensor failure, user can press the scram button 710 at robot back,
At this point, robot interior power supply is cut off, so that robot halts when that can not judge that traveling has barrier on the way, keep away
Exempt from robot to collide with barrier, to extend the service life of robot.
In one preferred embodiment of the utility model, avoidance sensor 510 includes infrared sensor and/or supersonic sensing
Device.Avoidance refer to robot in the process of walking, perceived by avoidance sensor 510 exist on its programme path it is static or
When dynamic barrier, according to certain algorithm real-time update path with cut-through object, avoid colliding with it.This practicality is new
Type avoidance sensor 510 includes infrared sensor and/or ultrasonic sensor, but not limited to this, laser sensor also can be used
Or visual sensor etc. is to achieve the purpose of obstacle avoidance.The array of the preferred infrared sensor of the utility model and ultrasonic sensor with
Avoid the barrier that robot advances on the way, wherein ultrasonic sensor is set in robot shells, and close to the bottom of robot
At portion position, front and back totally 6 groups, since the robot shells at setting ultrasound senor position have certain angle,
Its investigative range is wider;And infrared sensor then be set to robot bottom position, front and back totally 4 groups, ultrasonic sensor with
Infrared sensor cooperates to prevent robot from colliding with barrier, extends its service life.
In another preferred embodiment of the utility model, intelligent robot further includes the auditory localization set on robot head
Module 800, auditory localization module 800 include multiple microphones 810, and microphone 810 is respectively used to receive the voice of different directions
And localization of sound source.Shown in Figure 2, in the present embodiment, auditory localization module 800 includes multiple array distributions and is located at each side
To microphone 810, which can judge the specific orientation of user according to the time difference of the voice received.Such as, user
It makes a sound to call robot, at this point, what robot sent out the multiple microphones 810 being arranged by its head reception user
Sound judges the direction of propagation of the sound by calculating, so that it is determined that the specific orientation of user, and head 100 is turned to
User then follows user to walk.
In the utility model another embodiment, intelligent robot further includes the interaction being electrically connected with main control module 300
Module 900.Interactive module 900 includes voice recognition unit 910, and interactive module 900 is used to receive the speech-input instructions of user,
And corresponding information is exported with voice, image or video mode.Interactive module 900 includes voice recognition unit 910, Yong Hufa
The phonetic order gone out will be identified and parsed by voice recognition unit 910, and after being parsed, robot passes through loud speaker
Or liquid crystal display exports corresponding information.Such as, user sends out " I is hungry ", at this point, robot is received by interactive module 900
To the phonetic order and it is parsed, after being parsed, robot is made a reservation interface by liquid crystal display output, and according to
User preferences recommend related content by loud speaker.
In the another preferred embodiment of the utility model, intelligent robot further includes the camera shooting being electrically connected with main control module 300
Module 920, photographing module 920 include the depth camera list for measuring scene depth information and profile information in visual range
Member 930 and the main camera unit 940 for obtaining common image.In the present embodiment, photographing module 920 is taken the photograph including depth
As unit 930, which mainly uses light technology, the depth data for obtaining environment in the visual field, to structure
Build 3D depth images.In addition, eyes location technology also can be used in depth camera unit 930, i.e., apish two progress ranging
Technology, including two RGB cameras arranged side by side, according to the angle of object under test in the image data of two RGB cameras shooting
Degree difference calculates actual range, has the characteristics that high certainty of measurement.Flight time skill can also be used in depth camera unit 930
Art also can quickly calculate depth information to carry out accurate three-dimensional detection, in body-sensing using upper relatively broad.Main camera shooting is single
Member 940 includes the 4K cameras of a ultra high-definition, which is used to obtain the image information in environment, to improve machine
The visual capacity of head part.
In above-described embodiment, depth camera unit 930 includes the laser emitter 950 being connected with each other, laser pickoff 960
With measuring and calculating chip 970;Laser emitter 950 is reflected for emitting laser, laser pickoff 960 into measurement range for receiving
Laser, measuring and calculating chip 970 is for calculating scene depth information and profile information.In the present embodiment, depth camera unit 930
Laser emitter 950, laser pickoff 960 including interconnection and measuring and calculating chip 970.Laser emitter 950 generates one
Infrared laser code pattern (IR Light Coding Image), irradiation is in space so that the object in space is coupled with mark
Note, since speckle has the randomness of height, any two in space at speckle can all be different pattern.Laser pick-off
Device 960 is given measuring and calculating chip 970 after the laser image in intercepting space, image and is calculated, and depth map is calculated.This
Technology is put into scene as parlor, is exactly that depth camera unit 930 has sent out volume by laser emitter 950 in simple terms
The infrared laser of code, after this series excitation light is got on the object in monitoring scene, that is, described scene is by this invisible
Encoded infrared laser to labeled, then laser pickoff 960 (the CMOS inductors of a standard) is received and is returned
Infrared laser back, and give measuring and calculating chip 970 the information received and handled, finally return result to master control mould
Block 300.In the present embodiment, more accurate it can be obtained by the way that laser emitter 950, laser pickoff 960 and measuring and calculating chip 970 is arranged
The spatial data for obtaining depth, improves the space exploration ability of robot head.
In the another preferred embodiment of the utility model, intelligent robot further includes the communication being electrically connected with main control module 300
Module 980, communication module 980 with external equipment and internet for connecting.In the present embodiment, communication module 980 includes logical
The network communication unit of cable broadband access network is crossed, main use includes ADSL, PON technology etc..Include additionally being used for
Local communication unit, including wifi units, zigbee unit for being connect with external equipment etc..Further include connecting with customer mobile terminal
Bluetooth ble units connect etc..In the embodiment of the present invention, head and the interconnection of robot can be realized by the way that communication module 980 is arranged
The connection of net obtains magnanimity information resource;Other smart machines in local area network control man can be also established in user family;Or
The mobile device of person and user connect, and realize wireless control at home, can also be realized in mobile terminal by network communication unit
The upper function of carrying out remote control.
Above-described is only the part of the utility model or preferred embodiment, and either word or attached drawing all cannot be because
The range of this limitation the utility model protection utilizes the utility model under every design with one entirety of the utility model
Equivalent structure transformation made by specification and accompanying drawing content, or directly/be used in other related technical areas indirectly and include
In the range of the utility model is protected.
Claims (8)
1. a kind of intelligent robot, including head and fuselage, which is characterized in that the intelligent robot further include main control module and
Drive module, obstacle avoidance module, power supply module and the emergency stop module being electrically connected respectively with the main control module;Wherein,
The drive module includes driving motor, and the drive module is for driving the robot to move forward or back;
The obstacle avoidance module includes avoidance sensor, and the obstacle avoidance module is used to obtain the barrier letter in robot traveling process
Breath is to get around the barrier;
The power supply module is used to provide the electric energy of the robot normal work;
The emergency stop module includes scram button, and the scram button is set to the robot back, and the scram button is used for
Cut off the robot interior power supply.
2. intelligent robot as described in claim 1, which is characterized in that the avoidance sensor include infrared sensor and/
Or ultrasonic sensor.
3. intelligent robot as described in claim 1, which is characterized in that the intelligent robot further includes being set to the machine
The auditory localization module of head part, the auditory localization module include multiple microphones, and the microphone is respectively used to receive not
Equidirectional voice and localization of sound source.
4. intelligent robot as described in claim 1, which is characterized in that the intelligent robot further includes and the master control mould
The interactive module that block is electrically connected, the interactive module include voice recognition unit, and the interactive module is used to receive the voice of user
Input instruction, and corresponding information is exported with voice, image or video mode.
5. intelligent robot as described in claim 1, which is characterized in that the intelligent robot further includes and the master control mould
The photographing module that block is electrically connected, the photographing module include for measuring scene depth information and profile information in visual range
Depth camera unit and main camera unit for obtaining common image.
6. intelligent robot as claimed in claim 5, which is characterized in that the depth camera unit includes swashing for interconnection
Optical transmitting set, laser pickoff and measuring and calculating chip;The laser emitter is for emitting laser, the laser into measurement range
Receiver member is used to receive the laser of reflection, and the measuring and calculating chip is for calculating the scene depth information and profile information.
7. intelligent robot as described in claim 1, which is characterized in that the intelligent robot further includes and the master control mould
The communication module of block connection, the communication module with external equipment and internet for connecting.
8. intelligent robot as claimed in claim 7, which is characterized in that the communication module by WIFI, bluetooth or
ZigBee modes connect.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201721669686.9U CN207669312U (en) | 2017-12-04 | 2017-12-04 | Intelligent robot |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201721669686.9U CN207669312U (en) | 2017-12-04 | 2017-12-04 | Intelligent robot |
Publications (1)
Publication Number | Publication Date |
---|---|
CN207669312U true CN207669312U (en) | 2018-07-31 |
Family
ID=62967477
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201721669686.9U Expired - Fee Related CN207669312U (en) | 2017-12-04 | 2017-12-04 | Intelligent robot |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN207669312U (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108714886A (en) * | 2018-08-13 | 2018-10-30 | 四川桑瑞思环境技术工程有限公司 | A kind of data center's crusing robot |
-
2017
- 2017-12-04 CN CN201721669686.9U patent/CN207669312U/en not_active Expired - Fee Related
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108714886A (en) * | 2018-08-13 | 2018-10-30 | 四川桑瑞思环境技术工程有限公司 | A kind of data center's crusing robot |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Bai et al. | Smart guiding glasses for visually impaired people in indoor environment | |
US10872584B2 (en) | Providing positional information using beacon devices | |
US9432593B2 (en) | Target object information acquisition method and electronic device | |
US10643389B2 (en) | Mechanism to give holographic objects saliency in multiple spaces | |
US8965741B2 (en) | Context aware surface scanning and reconstruction | |
US7598976B2 (en) | Method and apparatus for a multisensor imaging and scene interpretation system to aid the visually impaired | |
CN110147106A (en) | Has the intelligent Mobile Service robot of laser and vision fusion obstacle avoidance system | |
CN106859929A (en) | A kind of Multifunctional blind person guiding instrument based on binocular vision | |
US11260533B2 (en) | Robot and robot system comprising same | |
KR101691564B1 (en) | Method for Providing Augmented Reality by using Tracking Eyesight | |
CN106415444A (en) | Gaze swipe selection | |
WO2015200406A1 (en) | Digital action in response to object interaction | |
Patel et al. | Multisensor-based object detection in indoor environment for visually impaired people | |
CN106687850A (en) | Scanning laser planarity detection | |
CN207736294U (en) | Intelligent robot | |
CN107976199A (en) | Navigation of Pilotless Aircraft method, system and unmanned plane | |
US20180075302A1 (en) | Wayfinding and Obstacle Avoidance System | |
CN108693532A (en) | Wearable barrier-avoiding method and device based on enhanced binocular camera Yu 3D millimetre-wave radars | |
Rajendran et al. | Design and implementation of voice assisted smart glasses for visually impaired people using google vision api | |
CN110383289A (en) | Device, method and the electronic equipment of recognition of face | |
EP4172740A1 (en) | Augmented reality eyewear with speech bubbles and translation | |
US20210370505A1 (en) | Method of detecting human and/or animal motion and performing mobile disinfection | |
KR20190111262A (en) | Portable device for measuring distance from obstacle for blind person | |
CN207669312U (en) | Intelligent robot | |
WO2022068193A1 (en) | Wearable device, intelligent guidance method and apparatus, guidance system and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20180731 Termination date: 20191204 |
|
CF01 | Termination of patent right due to non-payment of annual fee |