CN103376890B - The gesture remote control system of view-based access control model - Google Patents

The gesture remote control system of view-based access control model Download PDF

Info

Publication number
CN103376890B
CN103376890B CN201210121832.XA CN201210121832A CN103376890B CN 103376890 B CN103376890 B CN 103376890B CN 201210121832 A CN201210121832 A CN 201210121832A CN 103376890 B CN103376890 B CN 103376890B
Authority
CN
China
Prior art keywords
hands
scope
image
candidate
gesture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201210121832.XA
Other languages
Chinese (zh)
Other versions
CN103376890A (en
Inventor
王琪
范伟
谭志明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujitsu Ltd
Original Assignee
Fujitsu Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujitsu Ltd filed Critical Fujitsu Ltd
Priority to CN201210121832.XA priority Critical patent/CN103376890B/en
Publication of CN103376890A publication Critical patent/CN103376890A/en
Application granted granted Critical
Publication of CN103376890B publication Critical patent/CN103376890B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses the gesture remote control system of a kind of view-based access control model, the gesture remote control system of this view-based access control model includes: image capture device, for capturing a series of images of object;Gesture identification equipment, for the gesture of a series of images identification object captured from image capture device and recognition result is sent to operational order triggers equipment;And operational order triggers equipment, for triggering scheduled operation order according to the recognition result come from the transmission of gesture identification equipment.Described gesture identification equipment includes: hands detection part, the hands of the image detection object for being captured from image capture device;Hands tracking unit, for when the hands of object being detected when hands detection part in an image, follows the tracks of the hands of object in ensuing image;Gesture identification parts, the hands of the object traced into for the hands of object detected according to hands detection part and hands tracking unit determine the hands of object motion and according to determined by the motion of hands of object identify the gesture of object.

Description

The gesture remote control system of view-based access control model
Technical field
The present invention relates to image procossing, pattern recognition and Object tracking field, and relate more specifically to The gesture remote control system of view-based access control model.
Background technology
Along with life Computer and the numerous portable intelligent device current people becomes increasingly Indispensable, people are by more natural and more efficient mutual between people and computer of hope.So And, the man-machine interaction of traditional such as mouse/keyboard, remote controller, even touch screen etc (HCI) ancillary equipment (such as, in bathroom or kitchen, is being driven under some specific conditions In, etc.) it is inconvenient for a user, because local it is desirable that freely touch at these HCI.Therefore, in recent years, the gesture remote control system as one of potential solution obtains more coming The most concerns.
Substantially, gesture remote control system will be followed the tracks of hands and analyzed the expression of significant hands, if they Be identified as the one in predefined gesture, then corresponding operational order will be triggered to perform Predetermined operation.Owing in a lot of situations, gesture identification is the most complicated, so at gesture identification In reason, many different instruments are employed to solve this problem, such as HMM (Hidden Markov Models (HMM)), particle filter, finite state machine (FSM) and Neutral net.Most of gesture recognition systems require high computation complexity;Additionally, some of them tool Some is had to limit, for example, it is desired to extra equipment (if desired for wearing glove) or the instrument of precision (collecting depth information if desired for infrared camera) or can only be in good illumination environment and simple background Environment runs and (such as cannot distinguish between hands and the object of there is the colour of skin similar color, or can only identify quiet State gesture, etc.).
Accordingly, it would be desirable to the computation complexity of a kind of real time remote control is low and can be good in complex environment The gesture recognition system run.
Summary of the invention
According to an aspect of the present invention, the gesture remote control system of a kind of view-based access control model includes: image capturing Equipment, described image capture device is for capturing a series of images of object;Gesture identification equipment, institute State gesture identification equipment for a series of images identification object that captured from described image capture device Recognition result is also sent to operational order triggering equipment by gesture;And operational order triggers equipment, institute State operational order and trigger equipment for touching according to the recognition result come from the transmission of described gesture identification equipment Sending out scheduled operation order, wherein, described gesture identification equipment includes: hands detection part, and described hands is examined Survey the hands of the parts image detection object for being captured from described image capture device;Hands tracking portion Part, described hands tracking unit is at hands object being detected when described hands detection part in an image Time, ensuing image is followed the tracks of the hands of object;Gesture identification parts, described gesture identification parts For the hands of object detected according to described hands detection part and described hands tracking unit trace into right The hands of elephant determine the hands of object motion and according to determined by the motion of hands of object identify object Gesture.
In one embodiment, described hands detection part will be by being captured by described image capture device Image be transformed into gray level image, utilize cascade classifier based on local binary patterns from this gray scale The hands of image detection object.
In one embodiment, described hands tracking unit follows the tracks of the hands of object by following process: sharp It is used in previous image the scope of the hands detecting or tracing into and the broca scale picture of present image with previous Difference image between the broca scale picture of image, carrys out the use in the broca scale picture of original definition present image In the hunting zone following the tracks of hands;Perform template matching method to determine the hands as the present image traced into Scope, wherein, described template matching method includes:
In first candidate's hands scopes multiple defined in hunting zone, those the first candidate hands scopes have with The size that the size of To Template is identical, and, in second candidate's hands scope defined in difference image, This candidate's hands scope has the size identical with the size of To Template, and wherein, described To Template is The scope of the hands detected in previous image or trace into;
Following steps are performed until these multiple first candidates for the plurality of first candidate's hands scope circulation Hands scope is all till following matching judgment processes, so that it is determined that go out the time mated most with To Template Player's scope is as the scope of the hands traced in the broca scale picture of present image:
Calculate the meansigma methods of first candidate's hands scope and the absolute difference of each pixel of To Template As the first error;
If this first error is more than the first predetermined threshold, then it represents that this candidate's hands scope not with To Template mates, thus is excluded;
If this first error is less than the first predetermined threshold, then calculate the second error, the second error It it is the value that obtained of the meansigma methods of the value of each pixel by the first error being deducted second candidate's hands scope The value being multiplied by predetermined adjustment factor and obtain;
If this second error is less than the second predetermined threshold, it is determined that coupling, i.e. this first candidate hands Scope is determined to be in the scope of the hands traced in the skin color range of present image, and the second error Value as Second Threshold so that next first candidate's hands scope is carried out matching judgment.
In one embodiment, after the hunting zone determined in the broca scale picture of present image And before performing template matching method, revise the search model of original definition with reference to described difference image Enclose, and define the plurality of candidate's hands scope in hunting zone after the decrease and perform described template Joining method to determine the scope of the hands of present image, wherein, described correction includes: searching original definition Each limit of rope scope is gradually to inside contracting, and is more than the pixel of predetermined threshold when any side runs into pixel value Time, this limit stops to inside contracting.
In one embodiment, in described template matching method, for each first candidate's hands scope Determine that when mating with To Template, this coupling is verified, and described checking includes: calculate and be currently identified The first candidate's hands scope mated most is each with the scope of hands detected in described hands detection equipment The meansigma methods of the absolute difference of individual pixel is as the 3rd error;Judge that whether the 3rd error is more than the 3rd threshold Value, if the 3rd error more than the 3rd predetermined threshold, then judges currently to be identified first mated most Candidate's hands scope is not mated, thus gets rid of this first candidate hands scope.
In one embodiment, the broca scale picture used in the present invention can obtain as follows To: the value of the R component in the RGB component of each pixel of the image captured is deducted G component With the average of B component, to obtain a difference;And described difference is compared with predetermined threshold, if Described difference is less than described predetermined threshold, then the value of the respective pixel of broca scale picture takes 0, and if Described difference is bigger than described predetermined threshold, then the value of the respective pixel of broca scale picture takes described difference.
In one embodiment, described gesture identification parts according to described hands detection part and described hands with Track parts are in the position of the hands of the object that every two field picture detected or traced into, from the hands of every two field picture Hand position between adjacent two two field pictures that position calculation obtains is orientated and takes from adjacent two hand positions Direction of displacement to calculated adjacent two two field pictures determines the motion of the hands of object and will be determined The motion of hands of object right to identify with predefined gesture-hands movement trajectory map table carries out contrast The gesture of elephant.
According to a further aspect in the invention, a kind of gesture remote control method of view-based access control model is right including capturing The a series of images of elephant;Gesture from a series of images identification object captured;According to recognition result Trigger scheduled operation order, wherein, identify that the gesture of object includes: from the image detection captured The hands of object;When the hands of object being detected in an image, it is right to follow the tracks of in ensuing image The hands of elephant;And the hands according to the object detected and the hands of object traced into determine the hands of object Motion and according to determined by the motion of hands of object identify the gesture of object.
According to another aspect of the invention, a kind of side for RGB image being transformed into broca scale picture Method, the method includes: the value of the R component of the RGB component of each pixel of RGB image deducted G component and the average of B component, to obtain a difference;Described difference is compared with predetermined threshold, as The most described difference is less than described predetermined threshold, then the value of the respective pixel of broca scale picture takes 0, and such as The most described difference is bigger than described predetermined threshold, then the value of the respective pixel of broca scale picture takes described difference.
According to another aspect of the invention, a kind of for following the tracks of mesh calibration method, bag in image sequence Include: utilize scope and the broca scale of present image of the target detecting in previous image or tracing into Difference image between picture and the broca scale picture of previous image, carrys out the broca scale of original definition present image The hunting zone for following the tracks of target in Xiang;Perform template matching method to determine as working as of tracing into The scope of the target of front image, wherein, described template matching method includes:
At the first candidate target scopes multiple defined in described hunting zone, those the first candidate target models Enclose and there is the size identical with the size of To Template, and, defined in described difference image second Candidate target scope, this candidate target scope has the size identical with the size of To Template, its In, described To Template is the scope of the target detecting in previous image or tracing into;
Following steps are performed until these multiple first times for the plurality of first candidate target scope circulation Select target zone all till following matching judgment processes, mate most with To Template so that it is determined that go out Candidate target scope as the scope of the target traced in the broca scale picture of present image:
Calculate the average of a first candidate target scope and the absolute difference of each pixel of To Template Value is as the first error;
If this first error is more than the first predetermined threshold, then it represents that this candidate target scope is not Mate with To Template, thus be excluded;
If this first error is less than the first predetermined threshold, then calculate the second error, the second error It is that the meansigma methods of the value of each pixel by the first error deducts the second candidate target scope is obtained The value that value is multiplied by predetermined adjustment factor and obtains;
If this second error is less than the second predetermined threshold, it is determined that coupling, i.e. this first candidate mesh Mark scope is determined to be in the scope of the target traced in the skin color range of present image, and second The value of error as Second Threshold so that next first candidate target scope is carried out matching judgment.
The gesture recognition system of the view-based access control model according to the present invention, has and skin even if being moved across at hands In the case of some object of the color that color is similar, it is also possible to simply and efficiently from capture continuously Frame image sequence detects hands exactly, follows the tracks of hands movement and identify gesture.
Accompanying drawing explanation
Fig. 1 shows the schematic structure of the gestural control system of the present invention;
Fig. 2 shows the flow chart of the gesture recognition process performed by gesture identification equipment of the present invention;
Fig. 3 shows the example of the gray level image after captured RGB image and conversion;
Fig. 4 shows the example when successfully be detected hands in gray level image after the conversion, wherein hands Scope illustrated with rectangle frame;
Fig. 5 shows that the exemplary of broca scale picture utilized in the gesture recognition process of the present invention is shown Figure;
Fig. 6 shows the difference diagram between the broca scale picture of continuous two images moved right for hands Picture;
(a), (b) and (c) in Fig. 7 show defined in the gesture recognition process of the present invention for work as The broca scale picture of front image is followed the tracks of the exemplary diagram of the process of the hunting zone of hands;
Fig. 8 show the hunting zone of original definition be corrected after the exemplary diagram of result;
Fig. 9 shows the template matching algorithm improved in the gesture recognition process of the present invention by utilization Follow the tracks of the exemplary diagram of the process of the hands crossing over face;And
Figure 10 shows in the gesture identification equipment of the present invention for calculating hand position orientation, hands position Put orientation to change and the exemplary diagram in the defined hands movement direction of hand position orientation.
Detailed description of the invention
Below, by preferred embodiments of the present invention will be described in detail with reference to the annexed drawings.Note, in this specification With in accompanying drawing, it is denoted by the same reference numerals and there is the structural elements of substantially the same function and structure Part, and omit the repeated description to these structural details.
The invention provides the gesture remote control system of a kind of efficient and healthy and strong view-based access control model, this system Utilize common web camera work and in complex environment, show reliability, having only to relatively simultaneously Few amount of calculation.
Fig. 1 shows the diagram of the schematic construction of the gestural control system of the present invention.Such as Fig. 1 institute Showing, the gesture remote control system of the view-based access control model of the present invention mainly includes three parts: image capture device 101, for capturing a series of images of object;Gesture identification equipment 102, for setting from image capturing Recognition result is also sent to operational order by the gesture of standby 101 some the row image recognition objects captured Triggering equipment 103;And operational order triggers equipment 103, for according to setting from described gesture identification Standby 102 send the recognition result come triggers scheduled operation order.
In this manual, web camera is used as image capture device, but the present invention is not limited to This, any kind of known or following it will be appreciated that can capture video images acquisition equipment all It is used as this image capture device.In the present invention, gesture identification equipment is gestural control system Core, it identifies significant gesture and by the gesture motion identified and predefined hands Gesture compares, if the gesture motion identified matches with one of predefined gesture, then identifies Result is sent to operational order and triggers equipment to trigger predetermined operational order.This is described more fully below The content of the gesture recognition process of invention.
Before describing the gesture recognition process of the present invention, first provide this gesture recognition process and supported Predefined gesture, as shown in table 1.As known from Table 1, compared with static gesture, the present invention Gesture identification method also support dynamic gesture.Such as, when hands " short stay ", hands keeps quiet The shortest time (such as less than 1 second), when hands keeps " stop ", hands keeps static length Reaching more than 1 second, " 1 second " is merely illustrative and this value arbitrarily can be set by user here.Additionally, Such as, when hands motion (to the right, upwards, downwards) to the left, after hands movement track is short stay Mobile (to the right, upwards, downwards) to the left, and when hands is when brandishing, hands movement track be to the left-to Right-to the left-downwards (that is, →, ←, →, ←) mobile.Additionally, when hands turns clockwise, hands is transported Dynamic track is And when hands rotates counterclockwise, hands movement track is Here, hands movement rail Mark can be complete circle or more than half circle.If additionally, user does not make any significant Gesture, then arbitrary hands movement track is recorded.Although this giving definition of gesture as above, but Being that the present invention is not limited to this, user can the most at random be defined, and therefore, can be used for The gesture identified can be more natural.
Table 1: definition of gesture
Then, the detailed process of the gesture recognition process of the present invention will be described.This gesture recognition process from The image captured detects the hands of object, follows the tracks of the motion of hands, then according to the movement locus of hands with And predefined gesture carrys out what kind of significant gesture command of recognition expression, detailed process such as figure Shown in 2.
Fig. 2 shows the flow process of the gesture recognition process performed by gesture identification equipment of the present invention Figure.As in figure 2 it is shown, the gesture recognition process of the present invention by three main processing stage constitute: inspection Survey, follow the tracks of and identify.Specifically, a series of figures of the object for being captured by image capture device Every two field picture in Xiang, performs following process.First, at S201, image capture device caught The image of the object obtained is normalized, and being such as normalized is 160 pixel × 120 pixels, certainly this Bright being not limited to this, the big I being normalized arbitrarily is set by user.At S202, according to relevant hands The record that detection and hands are followed the tracks of determines whether detect in one's hands in previous frame image and whether trace into Hands, if being not detected by hands in previous frame, then performs the process of relevant hands detection, if upper One frame is followed the tracks of in one's hands, then performed the process that relevant hands is followed the tracks of.Such as, an embodiment In, mark isTrack is set for indicating whether detecting in one's hands, when detecting in one's hands, and isTrack Composed 1, and when being not detected by hands, isTrack is composed 0.It addition, in this embodiment, mark Will isDectec is set for indicating whether following the tracks of in one's hands, and when following the tracks of in one's hands, isDectec is composed 1, and when not following the tracks of in one's hands, isDectec is composed 0.Therefore, in this embodiment, by inspection Look into the value of mark isTrack and isDectec just can determine that whether detect in previous frame in one's hands and whether with Track is in one's hands.
In previous frame, it is not detected by hands and follows the tracks of in the case of in one's hands, in S202 being i.e. "No", performs the process of relevant hands detection.Specifically, as shown in Figure 2, in S203, Present image is transformed into gray level image from RGB image, certainly, if the image captured is just Be gray level image, then this step will be omitted.After Fig. 3 shows captured RGB image and conversion The example of gray level image.In S204, utilize training in advance good based on LBP (local binary mould Formula) cascade classifier gray level image after the conversion gamut in inspection side hands.Fig. 4 illustrates and works as Successfully be detected example during hands in gray level image after the conversion, wherein the scope of hands is by with rectangle frame Illustrate.As previously described, if be detected that hands, in S206, isDectec is composed 1, without Detect in one's hands, then in S207, isDectec is composed 0, and these records are saved for next frame Use when image processes.
When gray level image after the conversion is not detected by the case of hands, the gesture identification of the present invention Method returns to S201 to process next frame image.On the other hand, when ash after the conversion Degree image successfully be detected in the case of hands, in S208, current RGB image is transformed into the colour of skin Image, this is easy to separate area of skin color with background, thus reduce next frame image is carried out hands with The impact that track is caused due to background when processing.More specifically, used in the present invention by RGB The method that image is transformed into broca scale picture is as follows: assume that the value of the pixel in broca scale picture is represented by s, The R component of RGB component of each pixel in RGB image, G component, the value difference of B component Represented by r, g, b, and set an interim intermediate variable Temp, then the value of s is fixed by equation below Justice:
Temp=r-((g+b)/2));
Temp=MAX (0, Temp);
S=Temp > 140?0:Temp;
That is, first, the value of the R component in the RGB component of each pixel of RGB image is deducted G Component and the average value of B component, to obtain a difference;Then, by this difference and predetermined threshold ratio Relatively, if this difference is less than predetermined threshold, then the value of the respective pixel of broca scale picture takes 0, and such as Really this difference is bigger than predetermined threshold, then the value of the respective pixel of broca scale picture takes this difference.Such as, figure 5 show captured RGB image showing by the broca scale picture after converting with predetermined threshold 140 Example, wherein rectangle frame represents that hands is detected, as it has been described above, area of skin color separates with background.Although There has been described and as above convert, from RGB image, the method obtaining broca scale picture, but user Can also be as desired to use other colour of skin segregation methods to obtain broca scale picture.Additionally, currently scheme The broca scale picture of picture is reserved for the difference image meter when next image carrying out hands tracking and processing Calculate, and represent that the parameter of the scope of the hands detected also is reserved for carrying out next image Template matching when hands tracking processes and error evaluation.
In a upper image, have been detected by hands or follow the tracks of in the case of in one's hands, in S202 being i.e. "Yes", performs the process that relevant hands is followed the tracks of.The hands tracking of the present invention processes and uses the template improved Joining method, this template matching method calculates faster and can tolerate some of hands hand in motor process Change.Specifically, in S209, present image is transformed into broca scale picture from RGB image, becomes Change method as described above.In S210, calculate present image broca scale picture with preserved before Difference image between the broca scale picture of one image.In the calculating of difference image, only along setting about fortune The difference in dynamic direction is retained, and the value in opposite direction with hands movement is dropped.Specifically, first, The value of each pixel of the broca scale picture of calculating present image and the corresponding picture of the broca scale picture of previous image The difference of the value of element.Then, this difference is compared with 0, if this difference is bigger than 0, then difference diagram The value of the respective pixel of picture takes this difference, if this difference is less than 0, then and the respective pixel of difference image Value take 0.Fig. 6 show the broca scale of continuous two images moved right for hands as between Difference image.
Then, in S210, according to the hands detecting in previous image or tracing into and currently scheme Difference image between broca scale picture and the broca scale picture of previous image of picture carrys out the colour of skin at present image Image is followed the tracks of hands.Specifically, first, the model of the hands detecting in previous image or tracing into is utilized Enclose the hunting zone for following the tracks of hands in the broca scale picture of original definition present image.Such as, exist The scope of hands is by with in the example shown in rectangle frame, and hunting zone can define according to equation below:
Range.x=MAX (0, Target.x-15);
Range.y=MAX (0, Target.y-15);
Range.width=MIN (a-Range.x, Target.width+30);
Range.height=MIN (b-Range.y, Target.height+30),
Wherein, a left side for the rectangle frame of the hands detected during Target.x, Target.y represent previous image or trace into The horizontal stroke on the summit at upper angle, vertical coordinate, Target.width, Target.height represent inspection in previous image The width of the rectangle frame of hands surveyed or trace into and height, and Range.x, Range.y represent current The horizontal stroke on summit in the upper left corner of the hunting zone in the broca scale picture of image, vertical coordinate, Range.width, Range.height represent the width of the hunting zone in the broca scale picture of present image And highly, and a and b represents the horizontally and vertically pixel count of image, and " 15 " respectively, The numerical value of " 30 " is set based on experience value and according to circumstances may be such that initial setting Hunting zone is as close possible to other values of the hands scope followed the tracks of.
Next step perform the amount of calculation in template matching method in order to reduce and improve accuracy, as above After described original definition hunting zone, utilize the broca scale picture of present image and the skin of previous image Difference image between color image revises the hunting zone of original definition, will be in described difference image Area-of-interest is as revised hunting zone.As shown in Figure 7, it is shown that in (a) previous image The scope of the hands detected or trace into, the broca scale picture of (b) present image and the colour of skin of previous image Difference image between image, and (c) is at the search model defined in the broca scale picture of present image Enclosing, wherein, in figure, the label in the upper left corner exemplarily represents the frame number of image.Further, Fig. 8 shows The diagram that the hunting zone of the example as hunting zone correction reduces, wherein, original definition are gone out The four edges of hunting zone is by gradually to inside contracting, and when any bar limit runs into pixel value more than predetermined threshold When being worth the pixel of (such as 5), then this limit stops to inside contracting.
After hunting zone defined in the broca scale picture of present image, in S212, perform mould Plate matching method is to determine the scope of the hands as the present image traced into.Here template matching method refers to Using the scope of hands that detects or trace in a upper image as To Template, by candidate's hands scope with This template compares, if error is less than certain value, it is determined that coupling.In order to reduce when hands is crossed over There is tracking error during some objects (such as, face, other hands) of the colour of skin, above-mentioned template Join method and consider the movable information from difference image, particularly as follows:
In first candidate's hands scopes multiple defined in hunting zone, those the first candidate hands scopes have with The size that the size of To Template is identical, and, in second candidate's hands scope defined in difference image, This candidate's hands scope has the size identical with the size of To Template;
Following steps are performed until these multiple first candidates for the plurality of first candidate's hands scope circulation Hands scope is all till following matching judgment processes, so that it is determined that go out the time mated most with To Template Player's scope is as the scope of the hands traced in the broca scale picture of present image: calculate one first The meansigma methods of the absolute difference of each pixel of candidate's hands scope and To Template is as the first error;If should First error is more than the first predetermined threshold, then it represents that this candidate's hands scope is not mated with To Template, Thus be excluded;If this first error less than the first predetermined threshold, then calculates the second error, second Error is that the meansigma methods of the value of each pixel by the first error deducts second candidate's hands scope is obtained Value be multiplied by predetermined adjustment factor;If this second error is less than the second predetermined threshold, it is determined that Join, i.e. this first candidate hands scope is determined to be in the hands traced in the skin color range of present image Scope, and the value of the second error as Second Threshold so that next first candidate's hands scope is entered Row matching judgment.
As it has been described above, because this template matching method is by looking for step by step with reference to previous candidate's hands scope Going out the hands scope mated most, therefore, this makes it possible to the change of the hand tolerating hands in motion continuously Change.But, in continuous print tracking processes, this result also in the accumulation of matching error.Therefore, may be used To limit the matching error of history by arranging extra verification process, this verification process includes: meter Calculate and be currently identified the first candidate's hands scope mated most and hands detected in hands detection equipment The meansigma methods of the absolute difference of each pixel of scope is as the 3rd error;Judge that the 3rd error is the biggest In the 3rd threshold value, if the 3rd error more than the 3rd predetermined threshold, then judges currently to be identified The first candidate's hands scope joined is not mated, thus gets rid of this first candidate hands scope.Fig. 9 shows Go out the tracking result when hands is move across face, in order to be shown more clearly that effect, optionally shown Go out the picture frame at interval, and wherein rectangle frame has represented the hands traced into.
Then, in S213, it is judged that whether successfully track in one's hands in S212.If successfully tracked In one's hands, then in S214, indicate that isTrack is composed 1, otherwise indicates in S215 as mentioned above IsTrack composed 0 and method flow forward to S203 with to present image perform hands detection process.
In S214, indicate that isTrack is composed in the case of 1, i.e. follow the tracks of in one's hands in present image In the case of, at S216, the colour of skin of present image is preserved and processes for next frame image Time difference image calculate, and would indicate that the hands scope traced in present image parameter preserve with For for the template matching of next frame image is processed, and the position of hands that will trace in present image The horizontal stroke put, vertical coordinate are respectively defined as in previous image the abscissa of the scope of the hands detecting or tracing into With the half of width and and previous image in the vertical coordinate of the scope of hands that detects or trace into high The sum of the half of degree, if (x y) represents the position of the hands traced in present image with HandPos Coordinate, then it is represented as equation below:
HandPos (x, y)=(Target.x+Target.width/2, Target.y+Target.height/2)
Then, at S217, to all figures in a series of images captured by image capture device After being carried out the process that hands detects or hands is followed the tracks of, the multiple hand positions passing through to be recorded can get The movement locus of hands.By analyzing this movement locus and with reference to defined gesture-movement locus mapping table May recognize that the gesture of object.Specifically, first, calculate in every two adjacent images and detect or follow the tracks of Between the position of the hands arrived hand position orientation Orient, i.e. calculate by present image detection or with Track to hands position and previous image in the straight line that constitutes of the position of hands detecting or trace into and level The angle in direction
Then, direction of displacement DeltaOrient between two adjacent images is calculated.Specifically, will be current Image is orientated Orient relative to the hand position of previous image and deducts latter image relative to present image Hand position is orientated LastOrient and obtains difference.If the absolute value of this difference is more than 180, then exist In the case of described difference is more than 0, described difference deducts 360 and obtains value as picture displacement direction DeltaOrient;In the case of described difference is less than 0, described difference obtains value and makees plus 360 For picture displacement direction DeltaOrient.If the absolute value of this difference is not more than 180, the most described difference Value is as picture displacement direction DeltaOrient.
Then, for 8 kinds of significant gestures (" arbitrarily " is used for recording real time kinematics track, and And " short stay " is provided as the preliminary activities for up/down/left/right gesture, stop Stay, up/down/left/right, wave as one group, they values based on Orient;CW/CC As another group, they values based on DeltaOrient and the accumulation of a period of time and, especially, right In the gesture waved, it includes the alternating movement of motion to the left and to the right), gesture identification is as follows:
Stop
In order to identify the gesture of stop, need that holding in one's hands is detected for continuous STAY_NUM frame quiet Only, i.e. for continuous STAY_NUM frame, always have the position of the hands traced at present image with The position of the hands detecting in previous image or tracing into is identical, and wherein STAY_NUM is fixed in advance The frame number of justice.
Up/down/left/right
In order to identify the gesture of up/down/left/right, first, should detect that preliminary activities is " short Suspend and stay ", then at ensuing DIREC_NUM frame, the value of Orient should be protected in one direction Holding within the specific limits, wherein, DIREC_NUM is predefined frame number.As example, false Fixed " short stay " action is to keep static for 3 frames.Figure 10 (b) show for up/down/ The value scope of the Orient in left/right direction.Such as, if 46≤Orient≤134, then it is assumed that be Upward direction moves.
Wave:
In order to identify the gesture waved, need to detect 4 motor segments of continuous print: to the right-to the left-to the right- To the left.In order to detect each motor segment, need Orient value should be maintained at accordingly for N continuous frame Value in the range of (MIN_SEG_NUM≤N <=MAX_SEG_NUM), wherein MIN_SEG_NUM and MAX_SEG_NUM is predefined frame number threshold value.
Turn clockwise:
In order to identify the gesture turned clockwise, need in some sequential frame images, DeltaOrient Value just remain, the absolute value of DeltaOrient be maintained at certain limit (such as, more than 10 and Scope less than or equal to 50) and DeltaOrient absolute value and reached predetermined threshold CIRCLE_DEGREE。
Rotate counterclockwise:
In the judgement rotated counterclockwise, in addition to the value of DeltaOrient remains and bears, other are equal Identical with the judgement turned clockwise.
After gesture described above is identified, all enumerators for corresponding gesture candidate will be by Reset.Additionally, system will be " frozen " a few frame, i.e. nonrecognition gesture in these several frames.This machine It is the wrong identification in order to avoid being not intended to gesture for some that system is introduced into.Such as, if the user desired that Constantly making gesture to the right, hands can be retracted after gesture to the right completing one by naturally, Make next one gesture to the right the most again.If not utilizing this kind " freezing " mechanism, the most this " shifting Return " gesture being misidentified as to the left will be likely to.
So, the gesture of object is identified.
Then, in S219, it is exported to operational order as the gesture of recognition result and triggers and set Standby.
Then, equipment is triggered in the operational order as the interface between gesture and corresponding operating order In, send the gesture identification result come by checking predefined gesture according to from gesture identification equipment Corresponding operational order is triggered with the mapping table of operational order.Gesture and the mapping table of operational order As shown in Table 2 below.This table 2 gives and 8 kinds of corresponding looking into for windows picture of gesture See the mapping between the operation of device software.Although given here be 8 kinds of gestures corresponding for Mapping between the operation of windows Photo Viewer software, but the invention is not restricted to this, more Mapping between gesture and the corresponding operating planted can be defined, and operation not only can relate to various types of The operation of the software of type, it is possible to relating to the operation of difference in functionality to various electronic equipments, this is by user As desired to definition.Such as, the present invention can be applied to gesture-mouse control.
Gesture Operational order
Short stay Carriage return
To the left If previous/picture amplifies, represent and be moved to the left
To the right If the next one/picture amplifies, represent and move right
Upwards If picture amplifies, represent and move up
Downwards If picture amplifies, represent and move down
Wave Exit
Turn clockwise Amplify
Rotate counterclockwise Reduce
Table 2: gesture-operational order mapping table
The gesture recognition system of the view-based access control model according to the present invention, has and skin even if being moved across at hands In the case of some object of the color that color is similar, it is also possible to simply and efficiently from capture continuously Frame image sequence detects hands exactly, follows the tracks of hands movement and identify gesture.
Described above is each parts of the gesture identification remote control system of the view-based access control model according to the present invention and each Corresponding operating in parts, however can the situation of the purport without departing from the present invention make multiple change and Amendment, these change and amendment also falls within the scope of the present application.

Claims (9)

1. a gesture remote control system for view-based access control model, including:
Image capture device, described image capture device is for capturing a series of images of object;
Gesture identification equipment, described gesture identification equipment is for being captured from described image capture device Recognition result is also sent to operational order and triggers equipment by the gesture of a series of images identification object;And
Operational order triggers equipment, and described operational order triggers equipment for according to from described gesture identification Equipment sends the recognition result come and triggers scheduled operation order, wherein, described gesture identification equipment bag Include:
Hands detection part, described hands detection part is for being captured from described image capture device The hands of image detection object;
Hands tracking unit, described hands tracking unit is for when described hands detection part is at an image In when the hands of object being detected, ensuing image is followed the tracks of the hands of object;
Gesture identification parts, described gesture identification parts are for detecting according to described hands detection part To the hands of object and the hands of object that traces into of described hands tracking unit to determine the motion of the hands of object And according to determined by the motion of hands of object identify the gesture of object,
Wherein, described hands tracking unit is by processing the hands following the tracks of object as follows:
Utilize scope and the broca scale picture of present image of the hands detecting in previous image or tracing into And the difference image between the broca scale picture of previous image, carrys out the broca scale picture of original definition present image In for following the tracks of the hunting zone of hands;
Perform template matching method to determine the scope of the hands as the present image traced into,
Wherein, described template matching method includes:
In first candidate's hands scopes multiple defined in hunting zone, those the first candidate hands scopes have with The size that the size of To Template is identical, and, in second candidate's hands scope defined in difference image, This candidate's hands scope has the size identical with the size of To Template, and wherein, described To Template is The scope of the hands detected in previous image or trace into;
Following steps are performed until these multiple first candidates for the plurality of first candidate's hands scope circulation Hands scope is all till following matching judgment processes, so that it is determined that go out the time mated most with To Template Player's scope is as the scope of the hands traced in the broca scale picture of present image:
Calculate the meansigma methods of first candidate's hands scope and the absolute difference of each pixel of To Template As the first error;
If this first error is more than the first predetermined threshold, then it represents that this candidate's hands scope not with To Template mates, thus is excluded;
If this first error is less than the first predetermined threshold, then calculate the second error, the second error It it is the value that obtained of the meansigma methods of the value of each pixel by the first error being deducted second candidate's hands scope The value being multiplied by predetermined adjustment factor and obtain;
If this second error is less than the second predetermined threshold, it is determined that coupling, i.e. this first time Player's scope is determined to be in the scope of the hands traced in the skin color range of present image, and second The value of error as Second Threshold so that next first candidate's hands scope is carried out matching judgment.
The gesture remote control system of view-based access control model the most according to claim 1, wherein, described hands is examined Survey parts to pass through the image captured by described image capture device is transformed into gray level image, utilize Cascade classifier based on local binary patterns is from the hands of this gray level image detection object.
The gesture remote control system of view-based access control model the most according to claim 1, wherein, is determining After hunting zone in the broca scale picture of present image and before performing template matching method, reference Described difference image defines in revising the hunting zone of original definition, and hunting zone after the decrease The plurality of candidate's hands scope also performs described template matching method to determine the scope of the hands of present image, Wherein, described correction includes: by each limit of the hunting zone of original definition gradually to inside contracting, and works as When any side runs into the pixel that pixel value is more than predetermined threshold, this limit stops to inside contracting.
The gesture remote control system of view-based access control model the most according to claim 1, wherein, at described mould In plate matching method, determine for each first candidate's hands scope mate with To Template time, this coupling Being verified, described checking includes:
Calculate and be currently identified that the first candidate's hands scope mated most is examined with in described hands detection equipment The meansigma methods of the absolute difference of each pixel of the scope of the hands measured is as the 3rd error;
Judge whether the 3rd error is more than the 3rd threshold value, if the 3rd error is more than the 3rd predetermined threshold Value, then judge currently to be identified that the first candidate's hands scope mated most is not mated, thus get rid of This first candidate hands scope.
The gesture remote control system of view-based access control model the most according to any one of claim 1 to 4, its In, broca scale seems to obtain as follows:
The value of the R component in the RGB component of each pixel of the image captured is deducted G component With the average of B component, to obtain a difference;And
Described difference is compared with predetermined threshold, if described difference is less than described predetermined threshold, then skin The value of the respective pixel of color image takes 0, and if described difference bigger than described predetermined threshold, then skin The value of the respective pixel of color image takes described difference.
The gesture remote control system of view-based access control model the most according to claim 1, wherein, described gesture Identification component is detected according to described hands detection part and described hands tracking unit at every two field picture or is followed the tracks of To the position of hands of object, adjacent two two field pictures that obtain from the position calculation of the hands of every two field picture it Between hand position orientation and the position of adjacent two two field pictures that obtains from adjacent two hand position orientation calculation Move direction determine object hands motion and by determined by the motion of hands of object with predefined Gesture-hands movement trajectory map table carries out contrasting the gesture identifying object.
7. a gesture remote control method for view-based access control model, including:
The a series of images of capture object;
Gesture from a series of images identification object captured;
Scheduled operation order is triggered according to recognition result,
Wherein, identify that the gesture of object includes:
Hands from the image detection object captured;
When the hands of object being detected in an image, ensuing image is followed the tracks of object Hands;And
Hands according to the object detected and the hands of object traced into determine the motion of the hands of object also The motion of the hands of object determined by according to identifies the gesture of object,
Described gesture remote control method also includes:
Utilize scope and the broca scale picture of present image of the hands detecting in previous image or tracing into And the difference image between the broca scale picture of previous image, carrys out the broca scale picture of original definition present image In for following the tracks of the hunting zone of hands;
Perform template matching method to determine the scope of the hands as the present image traced into,
Wherein, described template matching method includes:
In first candidate's hands scopes multiple defined in hunting zone, those the first candidate hands scopes have with The size that the size of To Template is identical, and, in second candidate's hands scope defined in difference image, This candidate's hands scope has the size identical with the size of To Template, and wherein, described To Template is The scope of the hands detected in previous image or trace into;
Following steps are performed until these multiple first candidates for the plurality of first candidate's hands scope circulation Hands scope is all till following matching judgment processes, so that it is determined that go out the time mated most with To Template Player's scope is as the scope of the hands traced in the broca scale picture of present image:
Calculate the meansigma methods of first candidate's hands scope and the absolute difference of each pixel of To Template As the first error;
If this first error is more than the first predetermined threshold, then it represents that this candidate's hands scope not with To Template mates, thus is excluded;
If this first error is less than the first predetermined threshold, then calculate the second error, the second error It it is the value that obtained of the meansigma methods of the value of each pixel by the first error being deducted second candidate's hands scope The value being multiplied by predetermined adjustment factor and obtain;
If this second error is less than the second predetermined threshold, it is determined that coupling, i.e. this first candidate hands Scope is determined to be in the scope of the hands traced in the skin color range of present image, and the second error Value as Second Threshold so that next first candidate's hands scope is carried out matching judgment.
The gesture remote control method of view-based access control model the most according to claim 7, wherein, broca scale picture Obtain as follows:
The value of the R component of the RGB component of each pixel of the image captured is deducted G component with The average of B component, to obtain a difference;
Described difference is compared with predetermined threshold, if described difference is less than described predetermined threshold, then skin The value of the respective pixel of color image takes 0, and if described difference bigger than described predetermined threshold, then skin The value of the respective pixel of color image takes described difference.
9. for following the tracks of a mesh calibration method in image sequence, including:
Utilize scope and the broca scale of present image of the target detecting in previous image or tracing into Difference image between picture and the broca scale picture of previous image, carrys out the broca scale of original definition present image The hunting zone for following the tracks of target in Xiang;
Execution template matching method is to determine the scope of the target as the present image traced into, wherein, Described template matching method includes:
At the first candidate target scopes multiple defined in described hunting zone, those the first candidate target models Enclose and there is the size identical with the size of To Template, and, defined in described difference image second Candidate target scope, this candidate target scope has the size identical with the size of To Template, its In, described To Template is the scope of the target detecting in previous image or tracing into;
Following steps are performed until these multiple first times for the plurality of first candidate target scope circulation Select target zone all till following matching judgment processes, mate most with To Template so that it is determined that go out Candidate target scope as the scope of the target traced in the broca scale picture of present image:
Calculate the average of a first candidate target scope and the absolute difference of each pixel of To Template Value is as the first error;
If this first error is more than the first predetermined threshold, then it represents that this candidate target scope is not Mate with To Template, thus be excluded;
If this first error is less than the first predetermined threshold, then calculate the second error, the second error It is that the meansigma methods of the value of each pixel by the first error deducts the second candidate target scope is obtained The value that value is multiplied by predetermined adjustment factor and obtains;
If this second error is less than the second predetermined threshold, it is determined that coupling, i.e. this first candidate mesh Mark scope is determined to be in the scope of the target traced in the skin color range of present image, and second The value of error as Second Threshold so that next first candidate target scope is carried out matching judgment.
CN201210121832.XA 2012-04-16 2012-04-16 The gesture remote control system of view-based access control model Active CN103376890B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210121832.XA CN103376890B (en) 2012-04-16 2012-04-16 The gesture remote control system of view-based access control model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210121832.XA CN103376890B (en) 2012-04-16 2012-04-16 The gesture remote control system of view-based access control model

Publications (2)

Publication Number Publication Date
CN103376890A CN103376890A (en) 2013-10-30
CN103376890B true CN103376890B (en) 2016-08-31

Family

ID=49462114

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210121832.XA Active CN103376890B (en) 2012-04-16 2012-04-16 The gesture remote control system of view-based access control model

Country Status (1)

Country Link
CN (1) CN103376890B (en)

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101534742B1 (en) * 2013-12-10 2015-07-07 현대자동차 주식회사 System and method for gesture recognition of vehicle
CN104050454B (en) * 2014-06-24 2017-12-19 深圳先进技术研究院 A kind of motion gesture track acquisition methods and system
CN104318218A (en) * 2014-10-29 2015-01-28 百度在线网络技术(北京)有限公司 Image recognition method and device
CN105807783A (en) * 2014-12-30 2016-07-27 览意科技(上海)有限公司 Flight camera
CN106325485B (en) * 2015-06-30 2019-09-10 芋头科技(杭州)有限公司 A kind of gestures detection recognition methods and system
CN106491071A (en) * 2015-09-06 2017-03-15 中兴通讯股份有限公司 A kind of method for giving a test of one's eyesight and terminal
CN105223957B (en) * 2015-09-24 2018-10-02 北京零零无限科技有限公司 A kind of method and apparatus of gesture manipulation unmanned plane
CN105242614A (en) * 2015-11-17 2016-01-13 广州新科佳都科技有限公司 Platform screen door safety protection control method and system
CN106934333B (en) * 2015-12-31 2021-07-20 芋头科技(杭州)有限公司 Gesture recognition method and system
CN105657260A (en) * 2015-12-31 2016-06-08 宇龙计算机通信科技(深圳)有限公司 Shooting method and terminal
CN105677039B (en) * 2016-02-16 2020-06-09 北京博研智通科技有限公司 Method and device for detecting driving state based on gesture and wearable device
CN106022211B (en) * 2016-05-04 2019-06-28 北京航空航天大学 A method of utilizing gesture control multimedia equipment
CN106020448B (en) * 2016-05-06 2019-03-29 深圳市国华识别科技开发有限公司 Man-machine interaction method and system based on intelligent terminal
CN106094861B (en) * 2016-06-02 2024-01-12 零度智控(北京)智能科技有限公司 Unmanned aerial vehicle, unmanned aerial vehicle control method and unmanned aerial vehicle control device
CN106041912B (en) * 2016-06-16 2018-06-22 深圳先进技术研究院 Master-slave mode snake-shaped robot system and its position control method
CN106920251A (en) * 2016-06-23 2017-07-04 阿里巴巴集团控股有限公司 Staff detecting and tracking method and device
CN106295531A (en) * 2016-08-01 2017-01-04 乐视控股(北京)有限公司 A kind of gesture identification method and device and virtual reality terminal
CN108062773B (en) * 2016-11-07 2021-05-28 深圳光启合众科技有限公司 Image processing method and device and robot
CN106951871B (en) * 2017-03-24 2020-07-28 北京地平线机器人技术研发有限公司 Motion trajectory identification method and device of operation body and electronic equipment
CN109558000B (en) * 2017-09-26 2021-01-22 京东方科技集团股份有限公司 Man-machine interaction method and electronic equipment
CN110163055A (en) * 2018-08-10 2019-08-23 腾讯科技(深圳)有限公司 Gesture identification method, device and computer equipment
CN110163068A (en) * 2018-12-13 2019-08-23 腾讯科技(深圳)有限公司 Target object tracking, device, storage medium and computer equipment
CN110276292B (en) * 2019-06-19 2021-09-10 上海商汤智能科技有限公司 Intelligent vehicle motion control method and device, equipment and storage medium
CN110458095B (en) * 2019-08-09 2022-11-18 厦门瑞为信息技术有限公司 Effective gesture recognition method, control method and device and electronic equipment
CN110716648B (en) * 2019-10-22 2021-08-24 上海商汤智能科技有限公司 Gesture control method and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101853071A (en) * 2010-05-13 2010-10-06 重庆大学 Gesture identification method and system based on visual sense
CN102200830A (en) * 2010-03-25 2011-09-28 夏普株式会社 Non-contact control system and control method based on static gesture recognition
CN102339125A (en) * 2010-07-23 2012-02-01 夏普株式会社 Information equipment and control method and system thereof

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102200830A (en) * 2010-03-25 2011-09-28 夏普株式会社 Non-contact control system and control method based on static gesture recognition
CN101853071A (en) * 2010-05-13 2010-10-06 重庆大学 Gesture identification method and system based on visual sense
CN102339125A (en) * 2010-07-23 2012-02-01 夏普株式会社 Information equipment and control method and system thereof

Also Published As

Publication number Publication date
CN103376890A (en) 2013-10-30

Similar Documents

Publication Publication Date Title
CN103376890B (en) The gesture remote control system of view-based access control model
Ji et al. Advances in view-invariant human motion analysis: A review
Choi et al. A general framework for tracking multiple people from a moving camera
CN108717531B (en) Human body posture estimation method based on Faster R-CNN
Raheja et al. Real-time robotic hand control using hand gestures
Ibraheem et al. Survey on various gesture recognition technologies and techniques
CN107357427A (en) A kind of gesture identification control method for virtual reality device
CN106030610B (en) The real-time 3D gesture recognition and tracking system of mobile device
CN104850219A (en) Equipment and method for estimating posture of human body attached with object
CN102831439A (en) Gesture tracking method and gesture tracking system
CN103135753A (en) Gesture input method and system
CN105589553A (en) Gesture control method and system for intelligent equipment
CN104063677A (en) Equipment used for estimating human body posture and method thereof
JP6487642B2 (en) A method of detecting a finger shape, a program thereof, a storage medium of the program, and a system for detecting a shape of a finger.
JP2011186576A (en) Operation recognition device
Chen et al. Vision-based skeleton motion phase to evaluate working behavior: case study of ladder climbing safety
Viraktamath et al. Comparison of YOLOv3 and SSD algorithms
CN111767932B (en) Action determination method and device, computer equipment and computer readable storage medium
Jain et al. Gestarlite: An on-device pointing finger based gestural interface for smartphones and video see-through head-mounts
Gritai et al. Matching trajectories of anatomical landmarks under viewpoint, anthropometric and temporal transforms
CN102136139B (en) Targeted attitude analytical equipment and targeted attitude analytical approach thereof
Zhang et al. Robot-person tracking in uniform appearance scenarios: A new dataset and challenges
KR102227494B1 (en) Apparatus and method for processing an user input using movement of an object
Kim et al. Visual multi-touch air interface for barehanded users by skeleton models of hand regions
KR101171239B1 (en) Non-touch data input and operating method using image processing

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant