CN110442243A - A kind of man-machine interaction method and system - Google Patents
A kind of man-machine interaction method and system Download PDFInfo
- Publication number
- CN110442243A CN110442243A CN201910749386.9A CN201910749386A CN110442243A CN 110442243 A CN110442243 A CN 110442243A CN 201910749386 A CN201910749386 A CN 201910749386A CN 110442243 A CN110442243 A CN 110442243A
- Authority
- CN
- China
- Prior art keywords
- human body
- man
- unit
- posture information
- interactive system
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 16
- 230000003993 interaction Effects 0.000 title abstract description 16
- 230000009471 action Effects 0.000 claims abstract description 51
- 239000000284 extract Substances 0.000 claims abstract description 20
- 238000000605 extraction Methods 0.000 claims abstract description 17
- 230000033001 locomotion Effects 0.000 claims abstract description 14
- 230000002452 interceptive effect Effects 0.000 claims description 30
- 230000000007 visual effect Effects 0.000 claims description 7
- 210000002414 leg Anatomy 0.000 claims description 6
- 210000003128 head Anatomy 0.000 claims description 4
- 210000001217 buttock Anatomy 0.000 claims description 3
- 238000001514 detection method Methods 0.000 claims description 3
- 238000003384 imaging method Methods 0.000 claims description 3
- 230000005540 biological transmission Effects 0.000 claims description 2
- 238000005452 bending Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
Abstract
The present invention is applicable in human-computer interaction technique field, provides a kind of man-machine interaction method and system;Realtime graphic is acquired by image acquisition units and corresponding operational order is generated according to action message and posture information using the action message and posture information of image extraction unit extract real-time human body from image, and using instruction generation unit;To realize that movement and posture according to identification human body generate corresponding operational order, and also by instruction display unit display operation instruction, realization inputs operation instruction without using hardware, greatly improves work efficiency, hardware cost is reduced, the experience of user is improved.
Description
Technical field
The invention belongs to human-computer interaction technique field more particularly to a kind of man-machine interaction method and systems.
Background technique
Man-machine interactive system is usually made of interactive device and controlled device, generates control instruction by interactive device, then
Controlled device is sent by control instruction by wired or wireless mode.Traditional interactive device, such as mouse, touch screen or key
Disk etc., user executes click or the operation such as mobile target object or mobile cursor by these interactive devices on interactive interface.
By taking mouse as an example, when the mouse key that is pressed then generates corresponding control instruction, then by corresponding control instruction
It is sent to controlled device.
However, needing touch interaction equipment, such as mouse when interacting using traditional interactive device, need often
The mechanical button on mouse is pressed, because frequently pressing, the reduced service life of mouse need more frequently to replace mouse, is increased
Cost is added.
Summary of the invention
The purpose of the present invention is to provide a kind of man-machine interaction method and systems, it is intended to solve not mentioning due to the prior art
For a kind of man-machine interaction method and system, lead to the problem that hardware cost is high, working efficiency is low, user experience is bad.
On the one hand, the present invention provides a kind of man-machine interactive systems, comprising:
Image acquisition units are used for real-time image acquisition;
Image extraction unit, action message and posture information for the extract real-time human body from described image;
Instruction generation unit, for generating corresponding operational order according to the action message and the posture information;
Display unit is instructed, for showing the operational order.
It is further preferred that described image acquisition unit includes multiple images sensor, multiple described image sensors are in
The middle line at annular setting and the visual angle of two adjacent described image sensors is formed by angle and is equal;Multiple described images
The visual angle of sensor is towards the inner ring of the annular.
Preferably, described image acquisition unit generates three-dimensional figure also according to multiple described image sensor acquired images
Picture.
Preferably, action message and the posture letter of described image extraction unit extract real-time human body from the 3-D image
Breath.
Preferably, the action message includes headwork, hand motion, arm action and the leg action of the human body
And foot action;The posture information acts according to the waist of the human body, buttocks movement, leg action and headwork with
And arm action generates.
It is further preferred that the man-machine interactive system further includes face identification unit, the face identification unit is from institute
State the face feature data of human body described in extract real-time in image and be compared with preset face feature data, compare at
The action message and posture information of power control described image extraction unit extract real-time human body from described image, comparison are unsuccessfully sent out
Warning message is sent to carry out corresponding display to described instruction display unit.
Preferably, the face identification unit extracted from described image multiple human bodies face feature data,
And it is extracted after succeeding with preset face feature comparing, according to preset two priority classes described image extraction unit excellent
The action message and posture information of the high human body of first rank.
It is further preferred that described instruction generation unit also according to the sound of the human body generate corresponding operational order or
And text information.
Preferably, the man-machine interactive system further include:
Customized unit is instructed, for customized with the action message of the human body or operation corresponding with posture information
Instruction;
Controlled unit, for executing corresponding operation according to the operational order;
Audible alarm unit, for carrying out acoustic alarm.
It is further preferred that the man-machine interactive system further include:
Human eye detection unit, for obtaining the eye state of human body in real time from described image, at the eyes of the human body
In state of opening, then controls described instruction generation unit corresponding operation is generated according to the action message and the posture information
Instruction;The eyes of the human body are in closed state and then start timing, and the eyes closed time of the human body is greater than preset close
The time is closed, then sends warning message and carries out corresponding display and audible alarm unit carry out sound report to described instruction display unit
It is alert.
On the other hand, the present invention provides a kind of man-machine interaction method, the method includes the following steps:
Real-time image acquisition;
The action message and posture information of extract real-time human body from described image;
Corresponding operational order is generated according to the action message and the posture information;
Show the operational order.
It is of the invention beneficial to thinking that effect is to acquire realtime graphic by image acquisition units and uses image extraction unit
The action message and posture information of extract real-time human body from image, and using instruction generation unit according to action message and appearance
State information generates corresponding operational order;To realize that movement and posture according to identification human body generate corresponding operational order,
And also by instruction display unit display operation instruction, realization is input operation instruction without using hardware, substantially increases work
Efficiency reduces hardware cost, improves the experience of user.
Detailed description of the invention
Fig. 1 is the structural schematic diagram for the man-machine interactive system that the embodiment of the present invention one provides;
Fig. 2 is the flow chart of man-machine interaction method provided by Embodiment 2 of the present invention.
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, with reference to the accompanying drawings and embodiments, right
The present invention is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, and
It is not used in the restriction present invention.
Specific implementation of the invention is described in detail below in conjunction with specific embodiment:
Embodiment one:
The structure that Fig. 1 shows the man-machine interactive system of the offer of the embodiment of the present invention one illustrates only for ease of description
Part related to the embodiment of the present invention.
A kind of man-machine interactive system, comprising:
Image acquisition units 101 are used for real-time image acquisition;
Image extraction unit 102, action message and posture information for the extract real-time human body from image;
Instruction generation unit 103, for generating corresponding operational order according to action message and posture information;
Display unit 104 is instructed, is instructed for display operation.
It is further preferred that image acquisition units 101 include multiple images sensor (not shown), multiple images are passed
Sensor is arranged in a ring and the middle line at the visual angle of two adjacent imaging sensors is formed by angle and is equal;Multiple images pass
Towards the inner ring of annular, i.e. annular inner ring is image acquisition region at the visual angle of sensor, realizes 360 degree of nothings in the same plane
Dead angle acquires image;Multiple images sensor can also be arranged in multi-layer annular simultaneously, i.e., upper layer, middle layer, lower layer realize
Image is acquired in the plane of three different heights, realizes comprehensive no dead angle;It can also be selected simultaneously according to specific application scenarios
It is the imaging sensor of wide-angle or narrow angle with visual angle.
Preferably, image acquisition units 101 generate 3-D image also according to multiple images sensor acquired image;Just
The action message and posture information of extraction human body that can be more accurate in image extraction unit 102.
Preferably, the action message and posture information of the extract real-time human body from 3-D image of image extraction unit 102;
In an embodiment of the present invention, action message include human body headwork (such as: shake the head, inclined head faces upward head, low
Head etc.), hand motion (i.e. sign language acts, such as shows OK gesture, thumbs up gesture etc.), arm action (such as: in specific direction
Stretch or and bending etc.) and leg action (such as: stretched in specific direction or and be bent etc.) and foot action (such as: foot
Point lands, and heel lands, and single foot lands, and shows foot of standing at ease etc.);Posture information is according to the movement of the waist of human body, buttocks
Movement, leg action and headwork and arm action generate (such as: do attention position, goose-step posture, squat appearance
Gesture).
It is further preferred that man-machine interactive system further includes face identification unit 210, face identification unit 210 is from image
The face feature data of middle extract real-time human body are simultaneously compared with preset face feature data, compare and successfully control image
Extraction unit 102 action message of extract real-time human body and posture information, the subsequent operation processing of progress from image, compare and lose
It loses transmission warning message and carries out corresponding display to instruction display unit 104, or warning message can also be sent to audible alarm list
Member 240 is alarmed;The safety that human-computer interaction is improved equipped with face identification unit 210 prevents from arbitrarily being manipulated.
Preferably, face identification unit 210 extracted from image multiple human bodies face feature data and with it is default
Face feature comparing success after, according to preset two priority classes image extraction unit 102 extract priority level it is high
The action message and posture information of human body;It avoids the occurrence of operation confusion or crashes;
It is further preferred that instruction generation unit 103 generates corresponding operational order or and text also according to the sound of human body
It is single can also to be sent to instruction display by word information (microphone for being provided with the sound of captured image pickup area) for text information
Member 104 show or carries out corresponding text input with controlled unit 230;Improve intelligence degree, realization movement and language
The comprehensive human-computer interaction of sound.
In an embodiment of the present invention, man-machine interactive system further include:
Customized unit 220 is instructed, for customized with the action message of human body or operation corresponding with posture information
Instruction;It realizes and corresponding operational order is generated according to the action message and posture information of human body
Controlled unit 230, for executing corresponding operation according to operational order;Realize human-computer interaction;
Audible alarm unit 240, for carrying out acoustic alarm;Meet different use demands.
It is further preferred that man-machine interactive system further include:
Human eye detection unit 250, for obtaining the eye state of human body in real time from image, the eyes of human body, which are in, to be opened
State, then control instruction generation unit 103 generate corresponding operational order according to action message and posture information;The eyes of human body
Then start timing in closed state, the eyes closed time of human body be greater than preset closing time, then send warning message to
Instruction display unit 104 carries out corresponding display and audible alarm unit 240 carries out acoustic alarm;Prevent user from carrying out human-computer interaction
When sleeping (acoustic alarm is carried out by audible alarm unit 240, and then wakes up the user) or user when accident occurs (such as swoon
Or die suddenly) it alarms so as to cause the attention of surrounding population.
Embodiment two:
Fig. 2 shows the implementation processes of man-machine interaction method provided by Embodiment 2 of the present invention, for ease of description, only show
Go out part related to the embodiment of the present invention, details are as follows:
In step s 201, real-time image acquisition.
In step S202, the action message and posture information of extract real-time human body from image.
In step S203, corresponding operational order is generated according to action message and posture information.
In step S204, display operation instruction.
In an embodiment of the present invention: acquiring realtime graphic by image acquisition units 101 and use image extraction unit
102 from image extract real-time human body action message and posture information, and using instruction generation unit 103 according to movement believe
Breath and posture information generate corresponding operational order;To realize that movement and posture according to identification human body generate corresponding operation
Instruction, and also by instruction 104 display operation of display unit instruction, realization inputs operation instruction without using hardware, mentions significantly
High working efficiency, reduces hardware cost, improves the experience of user.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all in essence of the invention
Made any modifications, equivalent replacements, and improvements etc., should all be included in the protection scope of the present invention within mind and principle.
Claims (10)
1. a kind of man-machine interactive system, comprising:
Image acquisition units are used for real-time image acquisition;
Image extraction unit, action message and posture information for the extract real-time human body from described image;
Instruction generation unit, for generating corresponding operational order according to the action message and the posture information;
Display unit is instructed, for showing the operational order.
2. man-machine interactive system as described in claim 1, which is characterized in that described image acquisition unit includes that multiple images pass
Sensor, multiple described image sensors are arranged in a ring and the middle line at the visual angle of two adjacent described image sensors is formed
Angle be equal;The visual angle of multiple described image sensors is towards the inner ring of the annular.
3. man-machine interactive system as claimed in claim 2, which is characterized in that described image acquisition unit is also according to multiple described
Imaging sensor acquired image generates 3-D image.
4. man-machine interactive system as claimed in claim 3, which is characterized in that described image extraction unit is from the 3-D image
The action message and posture information of middle extract real-time human body.
5. man-machine interactive system as claimed in claim 4, which is characterized in that the action message includes the head of the human body
Movement, hand motion, arm action and leg action and foot action;The posture information is dynamic according to the waist of the human body
Work, buttocks movement, leg action and headwork and arm action generate.
6. man-machine interactive system as described in claim 1, which is characterized in that the man-machine interactive system further includes recognition of face
Unit, the face identification unit from described image the face feature data of human body described in extract real-time and with preset face
Portion's characteristic is compared, and compares the movement for successfully controlling described image extraction unit extract real-time human body from described image
Information and posture information compare unsuccessfully transmission warning message and carry out corresponding to display to described instruction display unit.
7. man-machine interactive system as claimed in claim 6, which is characterized in that the face identification unit is mentioned from described image
After getting the face feature data of multiple human bodies and succeeding with preset face feature comparing, according to preset
Two priority classes described image extraction unit extracts the action message and posture information of the high human body of priority level;
Described instruction generation unit generates corresponding operational order or and text information also according to the sound of the human body.
8. man-machine interactive system as described in claim 1, which is characterized in that the man-machine interactive system further include:
Customized unit is instructed, is referred to for customized with the action message of the human body or operation corresponding with posture information
It enables;
Controlled unit, for executing corresponding operation according to the operational order;
Audible alarm unit, for carrying out acoustic alarm.
9. man-machine interactive system as claimed in claim 8, which is characterized in that the man-machine interactive system further include:
Human eye detection unit, for obtaining the eye state of human body in real time from described image, the eyes of the human body, which are in, opens
Open state then controls described instruction generation unit and generates corresponding operation according to the action message and the posture information and refers to
It enables;The eyes of the human body are in closed state and then start timing, and the eyes closed time of the human body is greater than preset closure
Time then sends warning message and carries out corresponding display and audible alarm unit carry out sound report to described instruction display unit
It is alert.
10. a kind of method based on man-machine interactive system described in any one of claim 1-9, which is characterized in that the side
Method includes the following steps:
Real-time image acquisition;
The action message and posture information of extract real-time human body from described image;
Corresponding operational order is generated according to the action message and the posture information;
Show the operational order.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910749386.9A CN110442243A (en) | 2019-08-14 | 2019-08-14 | A kind of man-machine interaction method and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910749386.9A CN110442243A (en) | 2019-08-14 | 2019-08-14 | A kind of man-machine interaction method and system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110442243A true CN110442243A (en) | 2019-11-12 |
Family
ID=68435430
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910749386.9A Pending CN110442243A (en) | 2019-08-14 | 2019-08-14 | A kind of man-machine interaction method and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110442243A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111136662A (en) * | 2020-02-25 | 2020-05-12 | 上海擎朗智能科技有限公司 | Robot fetching confirmation method and device and robot |
CN115390663A (en) * | 2022-07-27 | 2022-11-25 | 合壹(上海)展览有限公司 | Virtual human-computer interaction method, system, equipment and storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103809733A (en) * | 2012-11-07 | 2014-05-21 | 北京三星通信技术研究有限公司 | Man-machine interactive system and method |
CN109815804A (en) * | 2018-12-19 | 2019-05-28 | 平安普惠企业管理有限公司 | Exchange method, device, computer equipment and storage medium based on artificial intelligence |
-
2019
- 2019-08-14 CN CN201910749386.9A patent/CN110442243A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103809733A (en) * | 2012-11-07 | 2014-05-21 | 北京三星通信技术研究有限公司 | Man-machine interactive system and method |
CN109815804A (en) * | 2018-12-19 | 2019-05-28 | 平安普惠企业管理有限公司 | Exchange method, device, computer equipment and storage medium based on artificial intelligence |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111136662A (en) * | 2020-02-25 | 2020-05-12 | 上海擎朗智能科技有限公司 | Robot fetching confirmation method and device and robot |
CN115390663A (en) * | 2022-07-27 | 2022-11-25 | 合壹(上海)展览有限公司 | Virtual human-computer interaction method, system, equipment and storage medium |
CN115390663B (en) * | 2022-07-27 | 2023-05-26 | 上海合壹未来文化科技有限公司 | Virtual man-machine interaction method, system, equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2877909B1 (en) | Multimodal interaction with near-to-eye display | |
CN108052079B (en) | Device control method, device control apparatus, and storage medium | |
JP5323770B2 (en) | User instruction acquisition device, user instruction acquisition program, and television receiver | |
US10248197B2 (en) | Systems and methodologies for real time eye tracking for electronic device interaction | |
KR102003588B1 (en) | System, method and computer program product for handling humanoid robot interaction with human | |
US20150331490A1 (en) | Voice recognition device, voice recognition method, and program | |
CN114391163A (en) | Gesture detection system and method | |
JP2003334389A (en) | Controller by gesture recognition, method thereof and recording medium | |
CN110442243A (en) | A kind of man-machine interaction method and system | |
WO2018103416A1 (en) | Method and device for detecting facial image | |
Mohammed | Efficient eye blink detection method for disabled-helping domain | |
KR20180096038A (en) | Crime prediction system based on moving behavior pattern | |
US9465981B2 (en) | System and method for communication | |
Pino et al. | Improving mobile device interaction by eye tracking analysis | |
Khilari | Iris tracking and blink detection for human-computer interaction using a low resolution webcam | |
Keskin et al. | A multimodal 3D healthcare communication system | |
Shashidhar et al. | Mouse Cusor Control Using Facial Movements-An HCI Application | |
CN113497912A (en) | Automatic framing through voice and video positioning | |
JP2015150620A (en) | robot control system and robot control program | |
Zhang et al. | Falling detection of lonely elderly people based on NAO humanoid robot | |
US10095308B2 (en) | Gesture based human machine interface using marker | |
Kapoor et al. | Light-Weight Seated Posture Guidance System with Machine Learning and Computer Vision | |
JP6859641B2 (en) | Evaluation system, information processing equipment and programs | |
JP2017041079A (en) | Operation recognition device | |
Elleuch et al. | Unwearable multi-modal gestures recognition system for interaction with mobile devices in unexpected situations |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20191112 |