CN103713738B - A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification - Google Patents

A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification Download PDF

Info

Publication number
CN103713738B
CN103713738B CN201310693150.0A CN201310693150A CN103713738B CN 103713738 B CN103713738 B CN 103713738B CN 201310693150 A CN201310693150 A CN 201310693150A CN 103713738 B CN103713738 B CN 103713738B
Authority
CN
China
Prior art keywords
high definition
photographic head
screen
gesture
coordinate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201310693150.0A
Other languages
Chinese (zh)
Other versions
CN103713738A (en
Inventor
何辉
李磊
刘凌志
汪志冰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Tuo Bao Science And Technology Co Ltd
Original Assignee
Wuhan Tuo Bao Science And Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Tuo Bao Science And Technology Co Ltd filed Critical Wuhan Tuo Bao Science And Technology Co Ltd
Priority to CN201310693150.0A priority Critical patent/CN103713738B/en
Publication of CN103713738A publication Critical patent/CN103713738A/en
Application granted granted Critical
Publication of CN103713738B publication Critical patent/CN103713738B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a kind of view-based access control model and follow the tracks of the man-machine interaction method with gesture identification, the present invention includes a varifocal high definition photographic head and multiple high definition photographic head.Apparatus of the present invention are arranged on above frame or the frame of screen.The information of wherein varifocal high definition camera collection vision, demarcate again through the position of Visual Observations Observations screen and the image of shooting, the identification of pupil is realized, thus realizing the first step visual tracking of the inventive method finally by the application specific processor of apparatus of the present invention.The second step of the present invention is after visual tracking, and multiple high definition photographic head realize the identification of gesture, i.e. finishing man-machine interaction in the region of Visual Observations Observations screen.Again through photographic head, screen is demarcated, gesture is carried out three-dimensional imaging, and judges position and the action of gesture touch point.Present invention achieves the gesture identification man-machine interactive system with visual tracking, it is provided that the contactless man-machine interaction mode of intelligence.

Description

A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification
Technical field
The invention belongs to Artificial technical field of intelligence, follow the tracks of the man-machine interaction method with gesture identification more particularly, to a kind of view-based access control model.
Background technology
The progress of technology makes to become closer to the mutual of natural way alternately between people and computer, " natural interaction " that namely people vigorously advocate.This man-machine interaction mode easily of touching technique has been pushed to numerous fields, except being applied in portable personal digital product, it is also widely used in the fields such as information household appliances, public information, electronic game, office automation devices and industrial equipment.Utilize touching technique, user to have only to use gesture the word on touching screen gently or that icon can be realized as with computer is mutual so that mutual more intuitively convenient between people and machine.
And existing touching technique must flow through human contact's screen thus the process of finishing man-machine interaction.The touching technique of this contact cannot realize the natural interaction with screen when people is away from screen, it is necessary to controls screen by devices such as remote controllers, it is impossible to providing good man-machine interaction experience, therefore above-mentioned touching technique exists limitation in artificial intelligence application.Visual Tracking utilizes the change of vision to replace staff motion on the touchscreen, still can arbitrary region on positioning screen when making people away from screen.Use touch screen by the motion of eyes, decrease many steps, accelerate development and the realization of intelligent man-machine interaction focusing on people.The application of the aspects such as current this technology is also only limitted to eye tracker, recognition of face, is not also applied in touching technique field.
Summary of the invention
Disadvantages described above or Improvement requirement for prior art, the invention provides a kind of view-based access control model and follow the tracks of the man-machine interaction method with gesture identification, its object is to, can arbitrarily have on the screen such as computer liquid crystal displayer of screen characteristics, common liquid crystals screen, projecting apparatus screen, giant display and realize visual tracking, and realize the man-machine interaction mode of Untouched control screen.
For achieving the above object, according to one aspect of the present invention, it is provided that a kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification, comprises the following steps:
(1) by infrared light supply, it is used for carrying out the varifocal high definition photographic head of visual tracking and multiple high definition photographic head for carrying out gesture identification is arranged on screen frame place;
(2) varifocal high definition camera collection facial image, and the facial image gathered is carried out face contour extraction;
(3) pixel coordinate (u of calculation procedure (2) obtains facial contour middle left and right pupil centereL,veL) and (ueR,veR);
(4) projection matrix Mel and the Mer of left and right pupil is calculated according to the pixel coordinate of left and right pupil center in facial contour and the coordinate at four angles of screen;
(5) projection matrix Mel and the Mer of the left and right pupil obtained by step (4) and the center pixel coordinate figure of left and right pupil calculate left and right pupil physical coordinates value on screen, and this region corresponding to physical coordinates value is the region that user performs gesture operation:
u e L v e L 1 = M e l X e l Y e l 1
u e R v e R 1 = M e r X e r Y e r 1
Wherein (Xer, Yer) represents right pupil physical coordinates value on screen, and (Xel, Yel) represents left pupil physical coordinates value on screen;
(6) according to the principle of binocular vision, the screen being placed with high definition photographic head is carried out parameter calibration, to obtain projection matrix Ml and the Mr of left and right high definition photographic head respectively;
(7) high definition camera collection user gesture touches the image of screen, and the image gathered is carried out pretreatment, to obtain the gesture of the user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F);
(8) according to the gesture operation of user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F) and the projection matrix Mr of the projection matrix Ml of left high definition photographic head and right high definition photographic head, and obtain the gesture of user three dimensional space coordinate (x on screen by below equationf,yf,zf), wherein this gesture operation is in the region that the physical coordinates value obtained in above-mentioned steps (5) is corresponding:
u 1 F v 1 F 1 = M l x f y f z f 1 , u 2 F v 2 F 1 = M r x f y f z f 1
(9) the coordinate z that step (8) obtains is judgedfWhether less than threshold values γ, if zfLess than γ, it can be determined that user's gesture generation click action, by USB interface by the three dimensional space coordinate (x of finger tipf,yf,zf) output, else process terminates.
Preferably, step (2) includes following sub-step:
(2-1) utilize varifocal high definition camera collection facial image, and with mask method to gather facial image denoising;
(2-2) utilize Sobel operator that the pixel on facial image does gradient conversion, to obtain facial contour.
Preferably, step (3) is specifically, using the left and right pixel coordinate value that Sobel operator obtains left pupil in the facial contour that step (2) obtains is uLeL、uHeL, the pixel coordinate value up and down of left pupil is vLeL、vHeL, the center pixel coordinate figure (u of left pupileL,veL) it is (uLeL+uHeL/2,vLeL+vHeL/ 2), the center pixel coordinate figure (u of right pupileR,veR) it is (uLeR+uHeR/2,vLeR+vHeR/ 2), wherein uLeR、uHeRFor the left and right pixel coordinate value of right pupil, vLeR、vHeRPixel coordinate value up and down for right pupil.
Preferably, step (6) is specifically, utilize Zhang Zhengyou to demarcate and screen is demarcated, to obtain demarcation thing pixel coordinate on the high definition photographic head of left and right, each demarcation thing pixel coordinate (u on the high definition photographic head of left and right1m,v1m)、(u2m,v2m), wherein m is the number of fixed point, and utilizes below equation to obtain the projection matrix Ml of left high definition photographic head and the projection matrix Mr of right high definition photographic head respectively:
u 1 m v 1 m 1 = M l x m y m z m 1 u 2 m v 2 m 1 = M r x m y m z m 1
Wherein (xm,ym,zm) for the physical coordinates of circle fixed point.
Preferably, step (7) specifically includes following sub-step:
(7-1) left and right high definition photographic head gathers user's gesture respectively and touches the image of screen, and is subtracted each other by pixel in the image corresponding point of the image collected and initialization frame, to form new image;
(7-2) the new image that step (7-1) is obtained carries out image denoising;
(7-3) utilize Sobel operator that the pixel on image does gradient conversion, to obtain edge detection graph;
(7-4) pixel on left and right high definition photographic head is carried out K curvature differentiation by the edge detection graph obtained according to step (7-3), to obtain the gesture of the user imager coordinate (u on left and right high definition photographic head1F,v1F) and (u2F,v2F)。
It is another aspect of this invention to provide that provide a kind of view-based access control model to follow the tracks of the man-machine interaction method with gesture identification, comprise the following steps:
(1) by infrared light supply, it is used for carrying out the varifocal high definition photographic head of visual tracking and multiple high definition photographic head for carrying out gesture identification is arranged on screen frame place;
(2) varifocal high definition camera collection facial image, and the facial image gathered is carried out face contour extraction;
(3) pixel coordinate (u of calculation procedure (2) obtains facial contour middle left and right pupil centereL,veL) and (ueR,veR);
(4) projection matrix Mel and the Mer of left and right pupil is calculated according to the pixel coordinate of left and right pupil center in facial contour and the coordinate at four angles of screen;
(5) projection matrix Mel and the Mer of the left and right pupil obtained by step (4) and the center pixel coordinate figure of left and right pupil calculate left and right pupil physical coordinates value on screen, and this region corresponding to physical coordinates value is the region that user performs gesture operation:
u e L v e L 1 = M e l X e l Y e l 1
u e R v e R 1 = M e r X e r Y e r 1
Wherein (Xer, Yer) represents right pupil physical coordinates value on screen, and (Xel, Yel) represents left pupil physical coordinates value on screen;
(6) according to the principle of binocular vision, the screen being placed with high definition photographic head is carried out parameter calibration, to obtain projection matrix Ml and the Mr of left and right high definition photographic head respectively;
(7) high definition camera collection user gesture touches the image of screen, and the image gathered is carried out pretreatment, to obtain the gesture of the user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F);
(8) when user slides touch screen, according to the gesture operation of user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F) and the projection matrix Mr of the projection matrix Ml of left high definition photographic head and right high definition photographic head, and gesture three dimensional space coordinate (x of the first frame finger tip on screen of user is obtained by below equationf1,yf1,zf1), wherein this gesture operation is in the region that the physical coordinates value obtained in above-mentioned steps (5) is corresponding:
u 1 F v 1 F 1 = M l x f 1 y f 1 z f 1 1 , u 2 F v 2 F 1 = M r x f 1 y f 1 z f 1 1
(9) step (8) is repeated, to obtain the three dimensional space coordinate (x of follow-up D-1 frame finger tip imagef2,yf2,zf2) ..., (xfD,yfD,zfD), wherein D represents that user slides the frame number of finger tip image collected when touching screen, thus obtaining gesture sliding trace on screen, is exported track by USB interface.
In general, by the contemplated above technical scheme of the present invention compared with prior art, it is possible to obtain following beneficial effect:
(1) present invention achieves and any screen (includes liquid crystal display screen, projecting apparatus screen or other screens etc.) realize that there is visual tracking location and the function of contactless touch;
(2) present invention uses simply, accurate positioning, it is simple to install.
Accompanying drawing explanation
Fig. 1 is the flow chart that view-based access control model of the present invention follows the tracks of the man-machine interaction method with gesture identification.
Fig. 2 is the schematic diagram of face contour detecting of the present invention.
Fig. 3 is the schematic diagram of visual tracking of the present invention.
Fig. 4 is the outline drawing of the device that gesture identification of the present invention uses.
Fig. 5 is the front view of the present invention.
Fig. 6 is the side view of screen of the present invention.
Fig. 7 is that the present invention demarcates thing schematic diagram.
Fig. 8 is that gesture of the present invention touches click schematic diagram.
Fig. 9 is gesture slip schematic diagram of the present invention.
Detailed description of the invention
In order to make the purpose of the present invention, technical scheme and advantage clearly understand, below in conjunction with drawings and Examples, the present invention is further elaborated.Should be appreciated that specific embodiment described herein is only in order to explain the present invention, is not intended to limit the present invention.As long as just can be mutually combined additionally, technical characteristic involved in each embodiment of invention described below does not constitute conflict each other.
As it is shown in figure 1, a kind of view-based access control model of the present invention is followed the tracks of and the man-machine interaction method of gesture identification, comprise the following steps:
(1) by infrared light supply, it is used for carrying out the varifocal high definition photographic head of visual tracking and multiple high definition photographic head for carrying out gesture identification is arranged on the optional position of screen frame;In the present embodiment, carrying out the varifocal high definition photographic head of visual tracking, its characteristic is 10 times of zooms, and resolution is 720P, and frame per second is 60 frames/second, and angle lens is at 110 °;Infrared light supply selects the infrared light of 800nm-1200nm wavelength;Carrying out the frame per second of the high definition photographic head of gesture identification be 60 frames/second, resolution is 720P, and angle lens is at 110 °;Screen is the screen of arbitrary dimension or form, including liquid crystal display screen, projecting apparatus screen or other screens) photographic head is placed in the optional position, left and right of screen frame, as shown in Figs. 4-6, exemplarily, choose the center installation infrared light source of screen upper side frame, varifocal high definition photographic head and high definition photographic head, in the present embodiment, use two high definition photographic head, and a varifocal high definition photographic head, it should be understood that the photographic head quantity of the present invention is not limited to this.The present invention passes through infrared light supply as secondary light source.
(2) varifocal high definition camera collection facial image, and the facial image gathered is carried out face contour extraction;As in figure 2 it is shown, this step includes following sub-step:
(2-1) utilize varifocal high definition camera collection facial image, and with mask method to gather facial image denoising;Specifically, the mask of a 3*3 is initially set up W = w 1 w 2 w 3 w 4 w 5 w 6 w 7 w 8 w 9 , Assume that on the facial image gathered, the pixel of certain point is aj,k, wherein j and k represents the position of point on image;Then there is aj,k=aj-1,k-1w1+aj-1,kw2+…+aj,kw5+…+aj+1,kw8+aj+1,k+1w9, thus obtaining new pixel aj,k, in the present embodiment, W = 1 9 1 1 1 1 1 1 1 1 1 ;
(2-2) facial image after denoising is carried out rim detection, namely utilize Sobel operator that the pixel on facial image does gradient conversion, to obtain facial contour;Specifically, M sets Sobel operator S h = - 1 - 2 - 1 0 0 0 1 2 1 For transverse gradients operator, S v = - 1 0 1 - 2 0 2 1 0 1 For longitudinal gradient operator, facial image is used respectively ShAnd SvCarry out convolution algorithm, with obtain this facial image at two horizontal and vertical on gradient map;
(3) pixel coordinate of left and right pupil center in the facial contour that calculation procedure (2) obtains;As it is shown on figure 3, this step is specifically, the Sobel operator that still calls in the facial contour that step (2) obtains in above-mentioned steps (2-2), the left and right pixel coordinate value obtaining left pupil is uLeL、uHeL, the pixel coordinate value up and down of left pupil is vLeL、vHeL, the therefore center pixel coordinate figure (u of left pupileL,veL) it is (uLeL+uHeL/2,vLeL+vHeL/ 2).In like manner can obtain the center pixel coordinate figure (u of right pupileR,veR) it is (uLeR+uHeR/2,vLeR+vHeR/ 2), wherein uLeR、uHeRFor the left and right pixel coordinate value of right pupil, vLeR、vHeRPixel coordinate value up and down for right pupil.
(4) projection matrix Mel and the Mer of left and right pupil is calculated according to the pixel coordinate of left and right pupil center in facial contour and the coordinate at four angles of screen, as shown in Figure 4, particularly as follows: first, when human eye vision watches the screen upper left corner attentively, (top left co-ordinate is (x to this stepA,yA, 0)), it is possible to trying to achieve left and right pupil center pixel coordinate on high definition photographic head by step (3) is (u1eL,v1eL), (u1eR,v1eR), in like manner try to achieve and watch the screen upper right corner attentively (upper right corner coordinate is (xB,yB, 0)) time, left and right pupil center is at the pixel coordinate respectively (u of high definition photographic head2eL,v2eL), (u2eR,v2eR);(lower left corner coordinate is (x to watch the screen lower left corner attentivelyC,yC, 0)) time, left and right pupil center is at the pixel coordinate respectively (u of high definition photographic head3eL,v3eL), (u3eR,v3eR);(lower right corner coordinate is (x to watch the screen lower right corner attentivelyD,yD, 0)) time, left and right pupil center is at the pixel coordinate respectively (u of high definition photographic head4eL,v4eL), (u4eR,v4eR);
Then, the principle according to binocular vision
u e L v e L 1 = M e l x y 1
Coordinate and top left co-ordinate according to above-mentioned four angles of screen are (xA,yA), upper right corner coordinate is (xB,yB), lower left corner coordinate is (xC,yC), lower right corner coordinate is (xD,yD) bring on the right of above-mentioned equation, the pixel coordinate of the left pupil that four angles of screen are corresponding is (u1eL,v1eL), (u2eL,v2eL), (u3eL,v3eL), (u4eL,v4eL) bring the above-mentioned equation left side, simultaneous solution equation into
Can calculate the projection matrix obtaining left pupil is
M e l = mel 11 mel 12 mel 13 mel 21 mel 22 mel 23 mel 31 mel 32 mel 33 ,
The projection matrix that in like manner can try to achieve right pupil is
M e r = mer 11 mer 12 mer 13 mer 21 mer 22 mer 23 mer 31 mer 32 mer 33 .
(5) projection matrix Mel and the Mer of the left and right pupil obtained by step (4) and the center pixel coordinate figure of left and right pupil calculate left and right pupil physical coordinates value on screen, and this region corresponding to physical coordinates value is the region that user performs gesture operation;Specifically, by the principle of following binocular vision
u e L v e L 1 = M e l X e l Y e l 1
u e R v e R 1 = M e r X e r Y e r 1
Calculating and obtain left and right pupil physical coordinates value on screen, wherein (Xer, Yer) represents right pupil physical coordinates value on screen;(Xel, Yel) represents left pupil physical coordinates value on screen.When vision invests different screen areas, showing broken box as shown in Figure 3, can complete location and the tracking of vision on screen, the region that the physical coordinates value of this step acquisition is corresponding is exactly the operating area of user's gesture in subsequent step (8).
(6) according to the principle of binocular vision, the screen being placed with high definition photographic head is carried out parameter calibration, to obtain projection matrix Ml and the Mr of left and right high definition photographic head respectively;Specifically, by the such as demarcation thing shown in figure (7), utilize Zhang Zhengyou to demarcate screen is demarcated, to obtain demarcation thing pixel coordinate on the high definition photographic head of left and right, each demarcation thing pixel coordinate (u on the high definition photographic head of left and right1m,v1m)、(u2m,v2m), wherein m is the number of fixed point, as figure (7) be shown with 9, (xm,ym,zm) for scheming the physical coordinates of the circle fixed point shown in (7).And utilize below equation to obtain the projection matrix Ml of left high definition photographic head and the projection matrix Mr of right high definition photographic head respectively:
u 1 m v 1 m 1 = M l x m y m z m 1 u 2 m v 2 m 1 = M r x m y m z m 1
The final projection matrix obtained is respectively
M l = ml 11 ml 12 ml 13 ml 14 ml 21 ml 22 ml 23 ml 24 ml 31 ml 32 ml 33 ml 34 , M r = mr 11 mr 12 mr 13 mr 14 mr 21 mr 22 mr 23 mr 24 mr 31 mr 32 mr 33 mr 34
(7) high definition camera collection user gesture touches the image of screen, the image gathered is carried out pretreatment, including image subtraction, image denoising, edge extracting, the finger tip differentiated based on K curvature or nib image recognition, to obtain the gesture of the user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F);As shown in Figure 8, this step specifically includes following sub-step:
(7-1) left and right high definition photographic head gathers user's gesture respectively and touches the image of screen, and is subtracted each other by pixel in the image corresponding point of the image collected and initialization frame, to form new image;
(7-2) the new image that step (7-1) is obtained carries out image denoising, and the process of image denoising is identical with above-mentioned steps (2-1), does not repeat them here;
(7-3) image after denoising is carried out rim detection, namely utilize Sobel operator that the pixel on image does gradient conversion, to obtain edge detection graph;The process of rim detection is identical with above-mentioned steps (2-2), does not repeat them here;
(7-4) pixel on left and right high definition photographic head is carried out K curvature differentiation by the edge detection graph obtained according to step (7-3), to obtain the gesture of the user imager coordinate on left and right high definition photographic head;Specifically, being the edge image that can extract gesture according to the edge detection graph obtained in (7-3), each edge coordinate point vector isIt is set to the K point that this is counted to for starting point by the clockwise direction at edgeThe K point counted to counterclockwise is set toThenK vector computing formula beWhen above-mentioned calculating α is more than 0 and more than setting threshold values β (its span as 0.5 to 1 between), then current vectorCorresponding pixel coordinate is the gesture of user imager coordinate (u on left high definition photographic head1F,v1F);The process of right photographic head is same as described above, and the gesture obtaining user is (u at the pixel coordinate of right high definition photographic head2F,v2F);
(8) according to the gesture operation of user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F) and the projection matrix Mr of the projection matrix Ml of left high definition photographic head and right high definition photographic head, and obtain the gesture of user three dimensional space coordinate (x on screen by below equationf,yf,zf), wherein this gesture operation is in the region that the physical coordinates value obtained in above-mentioned steps (5) is corresponding:
u 1 F v 1 F 1 = M l x f y f z f 1 , u 2 F v 2 F 1 = M r x f y f z f 1
Three dimensional space coordinate (the x that can obtain gesture is solved by above-mentioned two matrix equationf,yf,zf), namely complete three-dimensional imaging and the location of the gesture of user.
It should be noted that in this step, the mode that the gesture operation of user is click on touches screen.
(9) the coordinate z that step (8) obtains is judgedfWhether less than threshold values γ, wherein the span of γ and the length of screen are directly proportional, if zfLess than γ, it can be determined that user's gesture generation click action, by USB interface by the three dimensional space coordinate (x of finger tipf,yf,zf) output, else process terminates;
As shown in Figure 9, when user touches screen in a sliding manner, view-based access control model of the present invention location and the step included by gesture identification man-machine interaction method followed the tracks of are basic essentially identical with above-mentioned click mode, only difference is that above-mentioned steps (9) is replaced by:
Obtain the three dimensional space coordinate (x of continuous D frame finger tipf1,yf1,zf1), (xf2,yf2,zf2) ..., (xfD,yfD,zfD), wherein D represents that user slides the frame number of finger tip image collected when touching screen, and be positive integer, thus obtaining gesture sliding trace on screen, is exported by USB interface by track, thus realizing the identification of gesture slip.
Those skilled in the art will readily understand; the foregoing is only presently preferred embodiments of the present invention; not in order to limit the present invention, all any amendment, equivalent replacement and improvement etc. made within the spirit and principles in the present invention, should be included within protection scope of the present invention.

Claims (6)

1. a view-based access control model follows the tracks of the man-machine interaction method with gesture identification, it is characterised in that comprise the following steps:
(1) by infrared light supply, it is used for carrying out the varifocal high definition photographic head of visual tracking and multiple high definition photographic head for carrying out gesture identification is arranged on screen frame place;
(2) varifocal high definition camera collection facial image, and the facial image gathered is carried out face contour extraction;
(3) pixel coordinate (u of calculation procedure (2) obtains facial contour middle left and right pupil centereL,veL) and (ueR,veR);
(4) projection matrix Mel and the Mer of left and right pupil is calculated according to the pixel coordinate of left and right pupil center in facial contour and the coordinate at four angles of screen;
(5) projection matrix Mel and the Mer of the left and right pupil obtained by step (4) and the center pixel coordinate figure of left and right pupil calculate left and right pupil physical coordinates value on screen, and this region corresponding to physical coordinates value is the region that user performs gesture operation:
u e L v e L 1 = M e l X e l Y e l 1
u e R v e R 1 = M e r X e r Y e r 1
Wherein (Xer, Yer) represents right pupil physical coordinates value on screen, and (Xel, Yel) represents left pupil physical coordinates value on screen;
(6) according to the principle of binocular vision, the screen being placed with high definition photographic head is carried out parameter calibration, to obtain projection matrix Ml and the Mr of left and right high definition photographic head respectively;
(7) high definition camera collection user gesture touches the image of screen, and the image gathered is carried out pretreatment, to obtain the gesture of the user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F);
(8) according to the gesture operation of user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F) and the projection matrix Mr of the projection matrix Ml of left high definition photographic head and right high definition photographic head, and obtain the gesture of user three dimensional space coordinate (x on screen by below equationf,yf,zf), wherein this gesture operation is in the region that the physical coordinates value obtained in above-mentioned steps (5) is corresponding:
u 1 F v 1 F 1 = M l x f y f z f 1 , u 2 F v 2 F 1 = M r x f y f z f 1
(9) the coordinate z that step (8) obtains is judgedfWhether less than threshold values γ, if zfLess than γ, it can be determined that user's gesture generation click action, by USB interface by the three dimensional space coordinate (x of finger tipf,yf,zf) output, else process terminates.
2. man-machine interaction method according to claim 1, it is characterised in that step (2) includes following sub-step:
(2-1) utilize varifocal high definition camera collection facial image, and with mask method to gather facial image denoising;
(2-2) utilize Sobel operator that the pixel on facial image does gradient conversion, to obtain facial contour.
3. man-machine interaction method according to claim 2, it is characterised in that step (3) is specifically, using the left and right pixel coordinate value that Sobel operator obtains left pupil in the facial contour that step (2) obtains is uLeL、uHeL, the pixel coordinate value up and down of left pupil is vLeL、vHeL, the center pixel coordinate figure (u of left pupileL,veL) it is (uLeL+uHeL/2,vLeL+vHeL/ 2), the center pixel coordinate figure (u of right pupileR,veR) it is (uLeR+uHeR/2,vLeR+vHeR/ 2), wherein uLeR、uHeRFor the left and right pixel coordinate value of right pupil, vLeR、vHeRPixel coordinate value up and down for right pupil.
4. man-machine interaction method according to claim 1, it is characterized in that, step (6) is specially, utilize Zhang Zhengyou to demarcate screen is demarcated, to obtain demarcation thing pixel coordinate on the high definition photographic head of left and right, each demarcation thing pixel coordinate (u on the high definition photographic head of left and right1m,v1m)、(u2m,v2m), wherein m is the number of fixed point, and utilizes below equation to obtain the projection matrix Ml of left high definition photographic head and the projection matrix Mr of right high definition photographic head respectively:
u 1 m v 1 m 1 = M l x m y m z m 1 u 2 m v 2 m 1 = M r x m y m z m 1
Wherein (xm,ym,zm) for the physical coordinates of circle fixed point.
5. man-machine interaction method according to claim 1, it is characterised in that step (7) specifically includes following sub-step:
(7-1) left and right high definition photographic head gathers user's gesture respectively and touches the image of screen, and is subtracted each other by pixel in the image corresponding point of the image collected and initialization frame, to form new image;
(7-2) the new image that step (7-1) is obtained carries out image denoising;
(7-3) utilize Sobel operator that the pixel on image does gradient conversion, to obtain edge detection graph;
(7-4) pixel on left and right high definition photographic head is carried out K curvature differentiation by the edge detection graph obtained according to step (7-3), to obtain the gesture of the user imager coordinate (u on left and right high definition photographic head1F,v1F) and (u2F,v2F)。
6. a view-based access control model follows the tracks of the man-machine interaction method with gesture identification, it is characterised in that comprise the following steps:
(1) by infrared light supply, it is used for carrying out the varifocal high definition photographic head of visual tracking and multiple high definition photographic head for carrying out gesture identification is arranged on screen frame place;
(2) varifocal high definition camera collection facial image, and the facial image gathered is carried out face contour extraction;
(3) pixel coordinate (u of calculation procedure (2) obtains facial contour middle left and right pupil centereL,veL) and (ueR,veR);
(4) projection matrix Mel and the Mer of left and right pupil is calculated according to the pixel coordinate of left and right pupil center in facial contour and the coordinate at four angles of screen;
(5) projection matrix Mel and the Mer of the left and right pupil obtained by step (4) and the center pixel coordinate figure of left and right pupil calculate left and right pupil physical coordinates value on screen, and this region corresponding to physical coordinates value is the region that user performs gesture operation:
u e L v e L 1 = M e l X e l Y e l 1
u e R v e R 1 = M e r X e r Y e r 1
Wherein (Xer, Yer) represents right pupil physical coordinates value on screen, and (Xel, Yel) represents left pupil physical coordinates value on screen;
(6) according to the principle of binocular vision, the screen being placed with high definition photographic head is carried out parameter calibration, to obtain projection matrix Ml and the Mr of left and right high definition photographic head respectively;
(7) high definition camera collection user gesture touches the image of screen, and the image gathered is carried out pretreatment, to obtain the gesture of the user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F);
(8) when user slides touch screen, according to the gesture operation of user imager coordinate (u on left high definition photographic head1F,v1F) and imager coordinate (u on right high definition photographic head2F,v2F) and the projection matrix Mr of the projection matrix Ml of left high definition photographic head and right high definition photographic head, and gesture three dimensional space coordinate (x of the first frame finger tip on screen of user is obtained by below equationf1,yf1,zf1), wherein this gesture operation is in the region that the physical coordinates value obtained in above-mentioned steps (5) is corresponding:
u 1 F v 1 F 1 = M l x f 1 y f 1 z f 1 1 , u 2 F v 2 F 1 = M r x f 1 y f 1 z f 1 1
(9) step (8) is repeated, to obtain the three dimensional space coordinate (x of follow-up D-1 frame finger tip imagef2,yf2,zf2) ..., (xfD,yfD,zfD), wherein D represents that user slides the frame number of finger tip image collected when touching screen, thus obtaining gesture sliding trace on screen, is exported track by USB interface.
CN201310693150.0A 2013-12-17 2013-12-17 A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification Active CN103713738B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310693150.0A CN103713738B (en) 2013-12-17 2013-12-17 A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310693150.0A CN103713738B (en) 2013-12-17 2013-12-17 A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification

Publications (2)

Publication Number Publication Date
CN103713738A CN103713738A (en) 2014-04-09
CN103713738B true CN103713738B (en) 2016-06-29

Family

ID=50406777

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310693150.0A Active CN103713738B (en) 2013-12-17 2013-12-17 A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification

Country Status (1)

Country Link
CN (1) CN103713738B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016115872A1 (en) * 2015-01-21 2016-07-28 成都理想境界科技有限公司 Binocular ar head-mounted display device and information display method thereof
CN106406684B (en) * 2016-04-22 2019-07-02 法法汽车(中国)有限公司 Projection process method, apparatus and projector
CN107330746A (en) * 2016-09-18 2017-11-07 安徽华陶信息科技有限公司 A kind of purchase method and system based on VR technologies
CN106502418B (en) * 2016-11-09 2019-04-16 南京阿凡达机器人科技有限公司 A kind of vision follower method based on monocular gesture identification
CN106791399A (en) * 2016-12-22 2017-05-31 深圳市虚拟现实技术有限公司 Virtual reality zooming space localization method and system
CN108734066B (en) * 2017-04-20 2021-09-07 奥克斯空调股份有限公司 Gesture image acquisition equipment and method
CN107562210A (en) * 2017-10-17 2018-01-09 极鱼(北京)科技有限公司 interactive device and system
CN107967100A (en) * 2017-12-06 2018-04-27 Tcl移动通信科技(宁波)有限公司 Operation control process method and storage medium based on mobile terminal camera
CN109726646B (en) * 2018-12-14 2021-02-02 中国联合网络通信集团有限公司 Gesture recognition method and system and display method and system
CN111290575A (en) * 2020-01-21 2020-06-16 中国人民解放军空军工程大学 Multichannel interactive control system of air defense anti-pilot weapon
CN112380990A (en) * 2020-11-13 2021-02-19 咪咕文化科技有限公司 Picture adjusting method, electronic device and readable storage medium
CN112631422B (en) * 2020-12-10 2023-04-07 西安理工大学 Media interaction system based on human behavior triggering
CN115685616A (en) * 2022-10-27 2023-02-03 深圳市华弘智谷科技有限公司 Display device and eye movement tracking system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101441513A (en) * 2008-11-26 2009-05-27 北京科技大学 System for performing non-contact type human-machine interaction by vision
JP2009211563A (en) * 2008-03-05 2009-09-17 Tokyo Metropolitan Univ Image recognition device, image recognition method, image recognition program, gesture operation recognition system, gesture operation recognition method, and gesture operation recognition program
CN102426480A (en) * 2011-11-03 2012-04-25 康佳集团股份有限公司 Man-machine interactive system and real-time gesture tracking processing method for same
CN103246351A (en) * 2013-05-23 2013-08-14 刘广松 User interaction system and method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009211563A (en) * 2008-03-05 2009-09-17 Tokyo Metropolitan Univ Image recognition device, image recognition method, image recognition program, gesture operation recognition system, gesture operation recognition method, and gesture operation recognition program
CN101441513A (en) * 2008-11-26 2009-05-27 北京科技大学 System for performing non-contact type human-machine interaction by vision
CN102426480A (en) * 2011-11-03 2012-04-25 康佳集团股份有限公司 Man-machine interactive system and real-time gesture tracking processing method for same
CN103246351A (en) * 2013-05-23 2013-08-14 刘广松 User interaction system and method

Also Published As

Publication number Publication date
CN103713738A (en) 2014-04-09

Similar Documents

Publication Publication Date Title
CN103713738B (en) A kind of view-based access control model follows the tracks of the man-machine interaction method with gesture identification
US11609607B2 (en) Evolving docking based on detected keyboard positions
CN108052202B (en) 3D interaction method and device, computer equipment and storage medium
CN106598227B (en) Gesture identification method based on Leap Motion and Kinect
CN103488356B (en) A kind of touch identification method based on infrared camera three-dimensional imaging
EP2203982B1 (en) Detecting finger orientation on a touch-sensitive device
Hsieh et al. A real time hand gesture recognition system using motion history image
CN116724285A (en) Micro-gestures for controlling virtual and graphical elements
CN117178247A (en) Gestures for animating and controlling virtual and graphical elements
CN105637443B (en) Low-power mode in embedded unit
KR20120045667A (en) Apparatus and method for generating screen for transmitting call using collage
EP2966555A1 (en) Three-dimensional operation control method and device for touchscreen, and mobile terminal thereof
CN103677270A (en) Human-computer interaction method based on eye movement tracking
CN103135753A (en) Gesture input method and system
CN102880304A (en) Character inputting method and device for portable device
CN104808788A (en) Method for controlling user interfaces through non-contact gestures
CN103677240A (en) Virtual touch interaction method and equipment
WO2015153673A1 (en) Providing onscreen visualizations of gesture movements
TW201415376A (en) Gesture identification with natural images
CN106033250B (en) Object sensing device and method
CN104765454A (en) Human muscle movement perception based menu selection method for human-computer interaction interface
CN104035544A (en) Method for controlling electronic device and electronic device
Yang et al. 3D character recognition using binocular camera for medical assist
CN103902124B (en) 3D hologram interaction systems based on track identification and control method thereof
CN104375631A (en) Non-contact interaction method based on mobile terminal

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C53 Correction of patent of invention or patent application
CB02 Change of applicant information

Address after: 701, room 430075, building C2, future science and Technology City, East Lake hi tech Zone, Hubei, Wuhan, 999

Applicant after: WUHAN TURBO TECHNOLOGIES Corp.

Address before: 712, 2, 58, international headquarters, Optics Valley Road, East Lake hi tech Zone, Hubei, Wuhan, Optics Valley, 430074

Applicant before: WUHAN TURBO ELECTRONIC SYSTEM CO.,LTD.

COR Change of bibliographic data

Free format text: CORRECT: APPLICANT; FROM: WUHAN TURBO ELECTRONIC SYSTEM CO., LTD. TO: WUHAN TURBO TECHNOLOGY CO., LTD.

Free format text: CORRECT: ADDRESS; FROM: 430074 WUHAN, HUBEI PROVINCE TO: 430075 WUHAN, HUBEI PROVINCE

C14 Grant of patent or utility model
GR01 Patent grant
PE01 Entry into force of the registration of the contract for pledge of patent right
PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: A human-machine interaction method based on visual tracking and gesture recognition

Effective date of registration: 20230919

Granted publication date: 20160629

Pledgee: Wuhan Guanggu Branch of Hubei Bank Co.,Ltd.

Pledgor: WUHAN TURBO TECHNOLOGIES Corp.

Registration number: Y2023980057502