CN110442243A - A kind of man-machine interaction method and system - Google Patents

A kind of man-machine interaction method and system Download PDF

Info

Publication number
CN110442243A
CN110442243A CN201910749386.9A CN201910749386A CN110442243A CN 110442243 A CN110442243 A CN 110442243A CN 201910749386 A CN201910749386 A CN 201910749386A CN 110442243 A CN110442243 A CN 110442243A
Authority
CN
China
Prior art keywords
human body
man
unit
posture information
interactive system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910749386.9A
Other languages
Chinese (zh)
Inventor
涂友冬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Zhiwei Intelligent Software Development Co Ltd
Original Assignee
Shenzhen Zhiwei Intelligent Software Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Zhiwei Intelligent Software Development Co Ltd filed Critical Shenzhen Zhiwei Intelligent Software Development Co Ltd
Priority to CN201910749386.9A priority Critical patent/CN110442243A/en
Publication of CN110442243A publication Critical patent/CN110442243A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality

Abstract

The present invention is applicable in human-computer interaction technique field, provides a kind of man-machine interaction method and system;Realtime graphic is acquired by image acquisition units and corresponding operational order is generated according to action message and posture information using the action message and posture information of image extraction unit extract real-time human body from image, and using instruction generation unit;To realize that movement and posture according to identification human body generate corresponding operational order, and also by instruction display unit display operation instruction, realization inputs operation instruction without using hardware, greatly improves work efficiency, hardware cost is reduced, the experience of user is improved.

Description

A kind of man-machine interaction method and system
Technical field
The invention belongs to human-computer interaction technique field more particularly to a kind of man-machine interaction method and systems.
Background technique
Man-machine interactive system is usually made of interactive device and controlled device, generates control instruction by interactive device, then Controlled device is sent by control instruction by wired or wireless mode.Traditional interactive device, such as mouse, touch screen or key Disk etc., user executes click or the operation such as mobile target object or mobile cursor by these interactive devices on interactive interface.
By taking mouse as an example, when the mouse key that is pressed then generates corresponding control instruction, then by corresponding control instruction It is sent to controlled device.
However, needing touch interaction equipment, such as mouse when interacting using traditional interactive device, need often The mechanical button on mouse is pressed, because frequently pressing, the reduced service life of mouse need more frequently to replace mouse, is increased Cost is added.
Summary of the invention
The purpose of the present invention is to provide a kind of man-machine interaction method and systems, it is intended to solve not mentioning due to the prior art For a kind of man-machine interaction method and system, lead to the problem that hardware cost is high, working efficiency is low, user experience is bad.
On the one hand, the present invention provides a kind of man-machine interactive systems, comprising:
Image acquisition units are used for real-time image acquisition;
Image extraction unit, action message and posture information for the extract real-time human body from described image;
Instruction generation unit, for generating corresponding operational order according to the action message and the posture information;
Display unit is instructed, for showing the operational order.
It is further preferred that described image acquisition unit includes multiple images sensor, multiple described image sensors are in The middle line at annular setting and the visual angle of two adjacent described image sensors is formed by angle and is equal;Multiple described images The visual angle of sensor is towards the inner ring of the annular.
Preferably, described image acquisition unit generates three-dimensional figure also according to multiple described image sensor acquired images Picture.
Preferably, action message and the posture letter of described image extraction unit extract real-time human body from the 3-D image Breath.
Preferably, the action message includes headwork, hand motion, arm action and the leg action of the human body And foot action;The posture information acts according to the waist of the human body, buttocks movement, leg action and headwork with And arm action generates.
It is further preferred that the man-machine interactive system further includes face identification unit, the face identification unit is from institute State the face feature data of human body described in extract real-time in image and be compared with preset face feature data, compare at The action message and posture information of power control described image extraction unit extract real-time human body from described image, comparison are unsuccessfully sent out Warning message is sent to carry out corresponding display to described instruction display unit.
Preferably, the face identification unit extracted from described image multiple human bodies face feature data, And it is extracted after succeeding with preset face feature comparing, according to preset two priority classes described image extraction unit excellent The action message and posture information of the high human body of first rank.
It is further preferred that described instruction generation unit also according to the sound of the human body generate corresponding operational order or And text information.
Preferably, the man-machine interactive system further include:
Customized unit is instructed, for customized with the action message of the human body or operation corresponding with posture information Instruction;
Controlled unit, for executing corresponding operation according to the operational order;
Audible alarm unit, for carrying out acoustic alarm.
It is further preferred that the man-machine interactive system further include:
Human eye detection unit, for obtaining the eye state of human body in real time from described image, at the eyes of the human body In state of opening, then controls described instruction generation unit corresponding operation is generated according to the action message and the posture information Instruction;The eyes of the human body are in closed state and then start timing, and the eyes closed time of the human body is greater than preset close The time is closed, then sends warning message and carries out corresponding display and audible alarm unit carry out sound report to described instruction display unit It is alert.
On the other hand, the present invention provides a kind of man-machine interaction method, the method includes the following steps:
Real-time image acquisition;
The action message and posture information of extract real-time human body from described image;
Corresponding operational order is generated according to the action message and the posture information;
Show the operational order.
It is of the invention beneficial to thinking that effect is to acquire realtime graphic by image acquisition units and uses image extraction unit The action message and posture information of extract real-time human body from image, and using instruction generation unit according to action message and appearance State information generates corresponding operational order;To realize that movement and posture according to identification human body generate corresponding operational order, And also by instruction display unit display operation instruction, realization is input operation instruction without using hardware, substantially increases work Efficiency reduces hardware cost, improves the experience of user.
Detailed description of the invention
Fig. 1 is the structural schematic diagram for the man-machine interactive system that the embodiment of the present invention one provides;
Fig. 2 is the flow chart of man-machine interaction method provided by Embodiment 2 of the present invention.
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, with reference to the accompanying drawings and embodiments, right The present invention is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, and It is not used in the restriction present invention.
Specific implementation of the invention is described in detail below in conjunction with specific embodiment:
Embodiment one:
The structure that Fig. 1 shows the man-machine interactive system of the offer of the embodiment of the present invention one illustrates only for ease of description Part related to the embodiment of the present invention.
A kind of man-machine interactive system, comprising:
Image acquisition units 101 are used for real-time image acquisition;
Image extraction unit 102, action message and posture information for the extract real-time human body from image;
Instruction generation unit 103, for generating corresponding operational order according to action message and posture information;
Display unit 104 is instructed, is instructed for display operation.
It is further preferred that image acquisition units 101 include multiple images sensor (not shown), multiple images are passed Sensor is arranged in a ring and the middle line at the visual angle of two adjacent imaging sensors is formed by angle and is equal;Multiple images pass Towards the inner ring of annular, i.e. annular inner ring is image acquisition region at the visual angle of sensor, realizes 360 degree of nothings in the same plane Dead angle acquires image;Multiple images sensor can also be arranged in multi-layer annular simultaneously, i.e., upper layer, middle layer, lower layer realize Image is acquired in the plane of three different heights, realizes comprehensive no dead angle;It can also be selected simultaneously according to specific application scenarios It is the imaging sensor of wide-angle or narrow angle with visual angle.
Preferably, image acquisition units 101 generate 3-D image also according to multiple images sensor acquired image;Just The action message and posture information of extraction human body that can be more accurate in image extraction unit 102.
Preferably, the action message and posture information of the extract real-time human body from 3-D image of image extraction unit 102;
In an embodiment of the present invention, action message include human body headwork (such as: shake the head, inclined head faces upward head, low Head etc.), hand motion (i.e. sign language acts, such as shows OK gesture, thumbs up gesture etc.), arm action (such as: in specific direction Stretch or and bending etc.) and leg action (such as: stretched in specific direction or and be bent etc.) and foot action (such as: foot Point lands, and heel lands, and single foot lands, and shows foot of standing at ease etc.);Posture information is according to the movement of the waist of human body, buttocks Movement, leg action and headwork and arm action generate (such as: do attention position, goose-step posture, squat appearance Gesture).
It is further preferred that man-machine interactive system further includes face identification unit 210, face identification unit 210 is from image The face feature data of middle extract real-time human body are simultaneously compared with preset face feature data, compare and successfully control image Extraction unit 102 action message of extract real-time human body and posture information, the subsequent operation processing of progress from image, compare and lose It loses transmission warning message and carries out corresponding display to instruction display unit 104, or warning message can also be sent to audible alarm list Member 240 is alarmed;The safety that human-computer interaction is improved equipped with face identification unit 210 prevents from arbitrarily being manipulated.
Preferably, face identification unit 210 extracted from image multiple human bodies face feature data and with it is default Face feature comparing success after, according to preset two priority classes image extraction unit 102 extract priority level it is high The action message and posture information of human body;It avoids the occurrence of operation confusion or crashes;
It is further preferred that instruction generation unit 103 generates corresponding operational order or and text also according to the sound of human body It is single can also to be sent to instruction display by word information (microphone for being provided with the sound of captured image pickup area) for text information Member 104 show or carries out corresponding text input with controlled unit 230;Improve intelligence degree, realization movement and language The comprehensive human-computer interaction of sound.
In an embodiment of the present invention, man-machine interactive system further include:
Customized unit 220 is instructed, for customized with the action message of human body or operation corresponding with posture information Instruction;It realizes and corresponding operational order is generated according to the action message and posture information of human body
Controlled unit 230, for executing corresponding operation according to operational order;Realize human-computer interaction;
Audible alarm unit 240, for carrying out acoustic alarm;Meet different use demands.
It is further preferred that man-machine interactive system further include:
Human eye detection unit 250, for obtaining the eye state of human body in real time from image, the eyes of human body, which are in, to be opened State, then control instruction generation unit 103 generate corresponding operational order according to action message and posture information;The eyes of human body Then start timing in closed state, the eyes closed time of human body be greater than preset closing time, then send warning message to Instruction display unit 104 carries out corresponding display and audible alarm unit 240 carries out acoustic alarm;Prevent user from carrying out human-computer interaction When sleeping (acoustic alarm is carried out by audible alarm unit 240, and then wakes up the user) or user when accident occurs (such as swoon Or die suddenly) it alarms so as to cause the attention of surrounding population.
Embodiment two:
Fig. 2 shows the implementation processes of man-machine interaction method provided by Embodiment 2 of the present invention, for ease of description, only show Go out part related to the embodiment of the present invention, details are as follows:
In step s 201, real-time image acquisition.
In step S202, the action message and posture information of extract real-time human body from image.
In step S203, corresponding operational order is generated according to action message and posture information.
In step S204, display operation instruction.
In an embodiment of the present invention: acquiring realtime graphic by image acquisition units 101 and use image extraction unit 102 from image extract real-time human body action message and posture information, and using instruction generation unit 103 according to movement believe Breath and posture information generate corresponding operational order;To realize that movement and posture according to identification human body generate corresponding operation Instruction, and also by instruction 104 display operation of display unit instruction, realization inputs operation instruction without using hardware, mentions significantly High working efficiency, reduces hardware cost, improves the experience of user.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all in essence of the invention Made any modifications, equivalent replacements, and improvements etc., should all be included in the protection scope of the present invention within mind and principle.

Claims (10)

1. a kind of man-machine interactive system, comprising:
Image acquisition units are used for real-time image acquisition;
Image extraction unit, action message and posture information for the extract real-time human body from described image;
Instruction generation unit, for generating corresponding operational order according to the action message and the posture information;
Display unit is instructed, for showing the operational order.
2. man-machine interactive system as described in claim 1, which is characterized in that described image acquisition unit includes that multiple images pass Sensor, multiple described image sensors are arranged in a ring and the middle line at the visual angle of two adjacent described image sensors is formed Angle be equal;The visual angle of multiple described image sensors is towards the inner ring of the annular.
3. man-machine interactive system as claimed in claim 2, which is characterized in that described image acquisition unit is also according to multiple described Imaging sensor acquired image generates 3-D image.
4. man-machine interactive system as claimed in claim 3, which is characterized in that described image extraction unit is from the 3-D image The action message and posture information of middle extract real-time human body.
5. man-machine interactive system as claimed in claim 4, which is characterized in that the action message includes the head of the human body Movement, hand motion, arm action and leg action and foot action;The posture information is dynamic according to the waist of the human body Work, buttocks movement, leg action and headwork and arm action generate.
6. man-machine interactive system as described in claim 1, which is characterized in that the man-machine interactive system further includes recognition of face Unit, the face identification unit from described image the face feature data of human body described in extract real-time and with preset face Portion's characteristic is compared, and compares the movement for successfully controlling described image extraction unit extract real-time human body from described image Information and posture information compare unsuccessfully transmission warning message and carry out corresponding to display to described instruction display unit.
7. man-machine interactive system as claimed in claim 6, which is characterized in that the face identification unit is mentioned from described image After getting the face feature data of multiple human bodies and succeeding with preset face feature comparing, according to preset Two priority classes described image extraction unit extracts the action message and posture information of the high human body of priority level;
Described instruction generation unit generates corresponding operational order or and text information also according to the sound of the human body.
8. man-machine interactive system as described in claim 1, which is characterized in that the man-machine interactive system further include:
Customized unit is instructed, is referred to for customized with the action message of the human body or operation corresponding with posture information It enables;
Controlled unit, for executing corresponding operation according to the operational order;
Audible alarm unit, for carrying out acoustic alarm.
9. man-machine interactive system as claimed in claim 8, which is characterized in that the man-machine interactive system further include:
Human eye detection unit, for obtaining the eye state of human body in real time from described image, the eyes of the human body, which are in, opens Open state then controls described instruction generation unit and generates corresponding operation according to the action message and the posture information and refers to It enables;The eyes of the human body are in closed state and then start timing, and the eyes closed time of the human body is greater than preset closure Time then sends warning message and carries out corresponding display and audible alarm unit carry out sound report to described instruction display unit It is alert.
10. a kind of method based on man-machine interactive system described in any one of claim 1-9, which is characterized in that the side Method includes the following steps:
Real-time image acquisition;
The action message and posture information of extract real-time human body from described image;
Corresponding operational order is generated according to the action message and the posture information;
Show the operational order.
CN201910749386.9A 2019-08-14 2019-08-14 A kind of man-machine interaction method and system Pending CN110442243A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910749386.9A CN110442243A (en) 2019-08-14 2019-08-14 A kind of man-machine interaction method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910749386.9A CN110442243A (en) 2019-08-14 2019-08-14 A kind of man-machine interaction method and system

Publications (1)

Publication Number Publication Date
CN110442243A true CN110442243A (en) 2019-11-12

Family

ID=68435430

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910749386.9A Pending CN110442243A (en) 2019-08-14 2019-08-14 A kind of man-machine interaction method and system

Country Status (1)

Country Link
CN (1) CN110442243A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111136662A (en) * 2020-02-25 2020-05-12 上海擎朗智能科技有限公司 Robot fetching confirmation method and device and robot
CN115390663A (en) * 2022-07-27 2022-11-25 合壹(上海)展览有限公司 Virtual human-computer interaction method, system, equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103809733A (en) * 2012-11-07 2014-05-21 北京三星通信技术研究有限公司 Man-machine interactive system and method
CN109815804A (en) * 2018-12-19 2019-05-28 平安普惠企业管理有限公司 Exchange method, device, computer equipment and storage medium based on artificial intelligence

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103809733A (en) * 2012-11-07 2014-05-21 北京三星通信技术研究有限公司 Man-machine interactive system and method
CN109815804A (en) * 2018-12-19 2019-05-28 平安普惠企业管理有限公司 Exchange method, device, computer equipment and storage medium based on artificial intelligence

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111136662A (en) * 2020-02-25 2020-05-12 上海擎朗智能科技有限公司 Robot fetching confirmation method and device and robot
CN115390663A (en) * 2022-07-27 2022-11-25 合壹(上海)展览有限公司 Virtual human-computer interaction method, system, equipment and storage medium
CN115390663B (en) * 2022-07-27 2023-05-26 上海合壹未来文化科技有限公司 Virtual man-machine interaction method, system, equipment and storage medium

Similar Documents

Publication Publication Date Title
EP2877909B1 (en) Multimodal interaction with near-to-eye display
CN108052079B (en) Device control method, device control apparatus, and storage medium
JP5323770B2 (en) User instruction acquisition device, user instruction acquisition program, and television receiver
US10248197B2 (en) Systems and methodologies for real time eye tracking for electronic device interaction
KR102003588B1 (en) System, method and computer program product for handling humanoid robot interaction with human
US20150331490A1 (en) Voice recognition device, voice recognition method, and program
CN114391163A (en) Gesture detection system and method
JP2003334389A (en) Controller by gesture recognition, method thereof and recording medium
CN110442243A (en) A kind of man-machine interaction method and system
WO2018103416A1 (en) Method and device for detecting facial image
Mohammed Efficient eye blink detection method for disabled-helping domain
KR20180096038A (en) Crime prediction system based on moving behavior pattern
US9465981B2 (en) System and method for communication
Pino et al. Improving mobile device interaction by eye tracking analysis
Khilari Iris tracking and blink detection for human-computer interaction using a low resolution webcam
Keskin et al. A multimodal 3D healthcare communication system
Shashidhar et al. Mouse Cusor Control Using Facial Movements-An HCI Application
CN113497912A (en) Automatic framing through voice and video positioning
JP2015150620A (en) robot control system and robot control program
Zhang et al. Falling detection of lonely elderly people based on NAO humanoid robot
US10095308B2 (en) Gesture based human machine interface using marker
Kapoor et al. Light-Weight Seated Posture Guidance System with Machine Learning and Computer Vision
JP6859641B2 (en) Evaluation system, information processing equipment and programs
JP2017041079A (en) Operation recognition device
Elleuch et al. Unwearable multi-modal gestures recognition system for interaction with mobile devices in unexpected situations

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20191112