CN103677274B - A kind of interaction method and system based on active vision - Google Patents

A kind of interaction method and system based on active vision Download PDF

Info

Publication number
CN103677274B
CN103677274B CN201310724304.8A CN201310724304A CN103677274B CN 103677274 B CN103677274 B CN 103677274B CN 201310724304 A CN201310724304 A CN 201310724304A CN 103677274 B CN103677274 B CN 103677274B
Authority
CN
China
Prior art keywords
skin
colour
training
color
prior probability
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201310724304.8A
Other languages
Chinese (zh)
Other versions
CN103677274A (en
Inventor
沈三明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vtron Group Co Ltd
Original Assignee
Vtron Technologies Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vtron Technologies Ltd filed Critical Vtron Technologies Ltd
Priority to CN201310724304.8A priority Critical patent/CN103677274B/en
Publication of CN103677274A publication Critical patent/CN103677274A/en
Application granted granted Critical
Publication of CN103677274B publication Critical patent/CN103677274B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a kind of interaction method and system based on active vision, its method is that projection arrangement will need interactive image projection to arbitrary plane, and use the operation information of the filming apparatus collection user interactive picture to being projected out, the information of shooting is transferred to processing means be analyzed processing simultaneously, realizes man-machine interaction with the operation obtaining user corresponding.The processing means of the present invention uses the mode of off-line training so that detection efficiency is improved.When training data, it is proposed that a kind of new training mechanism, decrease substantial amounts of manual operation.When detection, have employed the model of real-time update, the colour of skin of the most current complexion model only several frames in front is relevant, so can effectively remove the illumination impact on the colour of skin, and effectively improve efficiency of algorithm, fully achieve real-time requirement.

Description

A kind of interaction method and system based on active vision
Technical field
The present invention relates to technical field of computer vision, more particularly, to a kind of interaction based on active vision Projecting method and system.
Then the picture to required interaction performs corresponding user operation, and by projection device to projection Region is watched for user;
Background technology
Although the multimodal human-computer interaction technology of integrated use vision, audition, sense of touch, olfactory sensation, the sense of taste etc. is more come Be applied manyly, but, both hands as action important in virtual reality system and perception relational model, It still plays the most replaceable effect in virtual reality system.At present, touch system is as a kind of up-to-date electricity Brain input equipment, it is man-machine interaction mode the simplest, convenient, a kind of.It give many matchmakers Body, with brand-new looks, is extremely attractive brand-new multimedia interactive equipment.Along with the progress of science and technology, projection The use of instrument is more and more easier, and any one plane can be become a display screen by it, and projector widely should For training conference, classroom instruction, cinema etc..
At present, camera and projection have gradually entered into the life of ordinary citizen, and carry out hands by projection and camera Automatically the identification of gesture becomes current study hotspot, by the automatic identification of gesture, has reached more preferable man-machine friendship Mutually so that the use of projection is convenient.The projection interactive system being currently based on vision is substantially based on auxiliary Light projection or based on finger tip location, but not strong in current existing scheme mostly real-time, and robustness is the highest.
Patent 200910190517.0 discloses a kind of projection interactive method based on finger, and hands is passed through in this invention The information such as CF in video file, extract the profile of finger, and record the movement locus of finger, Then by the movement locus of finger and the instruction contrast in the instruction database pre-defined, affiliated operation rail is judged Which kind of operational order mark belongs to, and reaches the purpose of man-machine interaction.Patent 200910197516.9 discloses one Finger identification method in interactive demonstration system, for interactive demonstration system based on video camera projection In determined the operation behavior of user by the identification of finger.This invention is based on image processing techniques, according to User paint operation time finger gesture feature, with its geometric space positional information for mainly identifying information, to taking the photograph As the image of head shooting is analyzed and processes, thus identify finger.In interactive projection system, in order to right Gesture is identified, and first has to segmentation from picture and obtains hand region.And at present about the segmentation of hand region An always difficult point.In projection interactive system, due to the irradiation of projector light, the arm of people may Presenting different colors, be also possible to comprise staff in projected picture, this brings difficulty all to segmentation arm.
Summary of the invention
For overcoming at least one defect (not enough) described in above-mentioned prior art, present invention firstly provides a kind of knowledge Other efficiency and the high interaction method based on active vision of accuracy.
A further object of the present invention is to propose a kind of interactive projection system based on active vision.
To achieve these goals, technical scheme is as follows:
A kind of interaction method based on active vision, projection arrangement will need interactive image projection to arbitrarily Plane, and use the operation information of the filming apparatus collection user interactive picture to being projected out, simultaneously by shooting Information is transferred to processing means and is analyzed processing, and realizes man-machine interaction with the operation obtaining user corresponding,
The process of described processing means Treatment Analysis information is: a selected color space to illumination-insensitive, Use Bayes' assessment to extract hand region, and this hand region be tracked and the location of fingertip location, Judge whether finger contacts with projection screen by range finding again, to determine whether to carry out man-machine interaction;
Using Bayes' assessment to extract hand region, its concrete mode is:
By off-line training acquisition probability P (s), P (c), P (c | s) and P (s | c), wherein P (s) represents training During the prior probability of the colour of skin, P (c) represents the prior probability of each color during training, P (c | s) table Show the prior probability that pixel is the colour of skin that pixel value is c, and P (s | c) represent that pixel value is c's after training Pixel is the probability of the colour of skin;
P (s | c)=P (c | s) P (s)/P (c) (1)
Hysteresis threshold T is obtained during off-line trainingmax, Tmax=P (s | c);It is to obtain according to probability distribution graph , namely probability distribution is interval;
Adaptive method is used to carry out Face Detection to extract hand region, according to the colour of skin district of nearest w frame identification Territory judges the area of skin color point detected at present, and the prior probability of nearest w frame is Pw(s), Pw(c), Pw(c | s), Wherein PwS () represents the prior probability of the colour of skin, P in nearest w framewC () represents each color in nearest w frame Prior probability, Pw(c | s) represent that in nearest w frame, pixel value is the prior probability that pixel is the colour of skin of c, separately examines The probability P that survey is area of skin color ' (s | c) be:
P'(s | and c)=γ P (s | c)+(1-γ) Pw(s|c) (2)
Wherein Pw(s | c) obtained by formula (1), i.e. Pw(s | c)=Pw(c|s)Pw(s)/PwC (), γ is one Control coefrficient, this control coefrficient is relevant with the training set of training stage;
Work as P'(s | c) > Tmax, detect whether the most at present into area of skin color, so far to extract hand region, otherwise It is not belonging to area of skin color.
Having a lot about hand region extracting method at present, the most frequently used having is based on hand color, based on hand Shape and the method demarcated based on color and physics.The color space that Face Detection is used at present mainly has The most several, RGB, normalized RGB, HSV, YCrCb, YUV etc..In the method, carry out During Face Detection, use color space to illumination-insensitive, thus so that the detection robustness of the colour of skin The highest.After a given color space, the way that the simplest judgement colour of skin which color is made up of is right Selected space adds a restrictive condition, i.e. afterwards threshold value Tmax.This hysteresis threshold TmaxDrawn by experience, As at given a series of area of skin color image, so can be obtained by the distribution of area of skin color.
This method uses Bayesian learning method based on off-line when extracting hand region, uses non-pattra leaves subsequently This framework is tracked.With existing algorithm is compared, this method has the advantage that (1) passes through off-line The mode of study obtains the distribution of the colour of skin, and so can be greatly increased detection is speed.(2) have employed in real time The complexion model of change, i.e. this complexion model only several frames in front are relevant, therefore, although do not have the colour of skin mould of complexity Type, under the light conditions even in change, also can robust, effectively identify area of skin color.(3) side The efficiency of algorithm that method is used is the highest, can reach real-time process.
A kind of system being applied to described interaction method based on active vision, including for needing mutually Dynamic image projection is to the projection arrangement of arbitrary plane, for gathering the operation of user's interactive picture to being projected out The filming apparatus of information and carry out the processing means of information process analysis,
The process of described processing means Treatment Analysis information is: a selected color space to illumination-insensitive, Use Bayes' assessment to extract hand region, and this hand region be tracked and the location of fingertip location, Judge whether finger contacts with projection screen by range finding again, to determine whether to carry out man-machine interaction;
Using Bayes' assessment to extract hand region, its concrete mode is:
By off-line training acquisition probability P (s), P (c), P (c | s) and P (s | c), wherein P (s) represents training During the prior probability of the colour of skin, P (c) represents the prior probability of each color during training, P (c | s) table Show the prior probability that pixel is the colour of skin that pixel value is c, and P (s | c) represent that pixel value is c's after training Pixel is the probability of the colour of skin;
P (s | c)=P (c | s) P (s)/P (c) (1)
Hysteresis threshold T is obtained during off-line trainingmax, Tmax=P (s | c);
Adaptive method is used to carry out Face Detection to extract hand region, according to the colour of skin district of nearest w frame identification Territory judges the area of skin color point detected at present, and the prior probability of nearest w frame is Pw(s), Pw(c), Pw(c | s), Wherein PwS () represents the prior probability of the colour of skin, P in nearest w framewC () represents each color in nearest w frame Prior probability, Pw(c | s) represent that in nearest w frame, pixel value is the prior probability that pixel is the colour of skin of c, separately examines The probability P that survey is area of skin color ' (s | c) be:
P'(s | and c)=γ P (s | c)+(1-γ) Pw(s|c) (2)
Wherein Pw(s | c) obtained by formula (1), i.e. Pw(s | c)=Pw(c|s)Pw(s)/PwC (), γ is one Control coefrficient, this control coefrficient is relevant with the training set of training stage;
Work as P'(s | c) > Tmax, detect whether the most at present into area of skin color, so far to extract hand region, otherwise It is not belonging to area of skin color.
Compared with prior art, technical solution of the present invention provides the benefit that: the present invention uses the side of off-line training Formula so that detection efficiency is improved.When training data, it is proposed that a kind of new training mechanism, decrease Substantial amounts of manual operation.When detection, have employed the model of real-time update, the most current complexion model only front The colour of skin of several frames is relevant, so can effectively remove the illumination impact on the colour of skin.And the calculation that this patent proposes Method efficiency is the highest, fully achieves real-time requirement.
Accompanying drawing explanation
Fig. 1 is the interactive projection system schematic diagram of the present invention.
Fig. 2 is the flow chart of processing means of the present invention analyzing and processing information.
Fig. 3 is for using triangulation measuring principle figure.
Detailed description of the invention
Accompanying drawing being merely cited for property explanation, it is impossible to be interpreted as the restriction to this patent;
In order to the present embodiment is more preferably described, some parts of accompanying drawing have omission, zoom in or out, and do not represent reality The size of border product;
To those skilled in the art, in accompanying drawing, some known features and explanation thereof may be omitted is to manage Solve.
It is illustrated in figure 1 the optical projection system schematic diagram of the present invention, including projector, two cameras, calculating equipment. The main effect of projector is exactly that picture is thrown into arbitrary plane.The effect of camera is the picture that shooting projects away Face, is then communicated in calculating equipment.The data of collected by camera are mainly analyzed by calculating equipment, its point The flow chart that analysis processes is as shown in Figure 2.
First, it is from complex background, extract staff.Extracting staff from complex background is exactly from entire image The middle staff part by correspondence extracts, and it relates to the segmentation of image and asks staff area judging two Topic.Segmentation image typically belongs to be the feature extraction of low level, mainly make use of the geological information of staff, color Information and movable information.Wherein, geological information includes the shape of staff, profile etc.;Movable information refers to The movement locus of staff.The most accurate location fingertip location that is extracted as in staff region is laid a good foundation, generally The methods such as grey relevant dynamic matrix, edge detection operator method, calculus of finite differences can be used to realize.In the present invention, in order to Remove the impact that projection ray irradiates, use method based on Bayesian Estimation when arm extracts.Finding hands After the position in portion, start to keep the tracking to hand region, simultaneously need to position finger tip accurately.And about The localization method of finger tip also has a lot, has special marking method, edge analysis, Hough circle transformation etc..
Finally, need calculating camera to the distance of finger, judge whether finger contacts with projection screen.This is The main method using triangulation of system, i.e. based on binocular vision three-dimensional rebuilding method calculates this distance.
In the present embodiment, the method for " Bayesian Estimation " is used to carry out skin cluster:
Detection method about hand region has a lot at present, and the most frequently used having is based on hand color, based on hands Portion's shape and the method demarcated based on color and physics.It is to illumination that current algorithm exists maximum problem The most sensitive, secondly it is exactly the efficiency of algorithm.The color space that at present Face Detection is used mainly have with Under several, RGB, normalized RGB, HSV, YCrCb, YUV etc..When carrying out Face Detection, The color space to illumination-insensitive can be paid the utmost attention to, thus so that the detection robustness of the colour of skin compares High.After a given color space, the way that the simplest judgement colour of skin which color is made up of is to selected Space add restrictive condition, i.e. a hysteresis threshold.This hysteresis threshold is drawn by experience, as given one The area of skin color image of series, so can be obtained by the distribution of area of skin color.The present embodiment uses based on off-line Bayesian learning method detect area of skin color, use non-Bayesian frame to be tracked subsequently.With existing Algorithm compare, the present invention has an advantages below: (1) obtain by the way of off-line learning the colour of skin point Cloth, so can be greatly increased detection is speed.(2) have employed the complexion model of real-time change, i.e. this skin The color model only several frames in front are relevant, therefore, although do not have the complexion model of complexity, even in the illumination of change In the case of, also can robust, effectively identify area of skin color.(3) efficiency of algorithm of the present invention The highest, real-time process can be reached.
1.1 Face Detection
Face Detection mainly includes following components: (a) estimates certain pixel and belong to the probability of the colour of skin; B () obtains hysteresis threshold T according to probability distribution graphmax;Face Detection have employed bayes method, mainly wraps Include the off-line training process of iteration and adaptive detection process.
A. training and testing mechanism
Firstly the need of the given a series of picture including area of skin color, and need manual choosing in picture Going out flesh tone portion, employing color space is YUV4:2:2.The when that but the present embodiment specifically being trained and identifies, Do not use Y-component, mainly have following two reason: (1) Y-component is relevant with the brightness of pixel, Therefore remove Y-component, can effectively reduce the illumination impact on detection.(2) after removing Y-component, phase For YUV, the dimension of image reduces, so the efficiency of whole process can be greatly promoted.
Assuming that a pixel is that (x, y), its pixel value is that (x, y), training process mainly calculates following several c=c to I Individual value: prior probability P (s) of (1) colour of skin;(2) prior probability that each color occurs in training data P(c);(3) pixel value is the prior probability P that pixel is the colour of skin (c | s) of c;Through training after, pixel The probability P that pixel is the colour of skin (s | c) that value is c can be obtained by bayesian criterion:
P (s | c)=P (c | s) P (s)/P (c) (1)
I.e. can determine that hysteresis threshold Tmax, Tmax=P (s | c), in being embodied as, when a certain pixel belongs to skin The probability of color is more than hysteresis threshold Tmax, then this point is the colour of skin.
A. the off-line training after simplifying
Training mainly completes in the case of off-line, asks so it does not interferes with online detection efficiency etc. Topic.But, the training data that will obtain abundance is a job the most time-consuming.In order to solve this problem, The present embodiment uses a kind of adaptive training process.First, it is trained with one group of little data set, Then carry out the identification of area of skin color by substantial amounts of data and hysteresis threshold, and real-time renewal priori is general Rate P (s), P (c) and P (c | s).The most continuous update after threshold value by the area of skin color in picture and Non-area of skin color is separately.If classifier training obtains the result of mistake, then need manpower intervention to correct this Mistake, but the method still can complete the work required for major part.If it is intended to more accurate result, can To input more training data, if the result of training has reached demand, can deconditioning immediately.
B. the self-adapting detecting colour of skin
Even if having employed UV model, in the case of illumination is continually changing, still can obtain some mistakes Recognition result.In order to solve this problem, need according to the area of skin color of former frame identifications judge current Area of skin color point.Therefore, this patent have employed two groups of prior probabilitys: i.e. off-line training prior probability P (s), P (c) With P (c | s).The prior probability P of w frame recentlyw(s), Pw(c), Pw(c|s).The priori of the most nearest w frame Probability can reflect current colour of skin state, and also can preferably adapt to current light conditions.The colour of skin is then Defined by following formula:
P'(s | and c)=γ P (s | c)+(1-γ) Pw(s|c) (2)
Wherein Pw(s | c) and Pw(s | c) all can be obtained by (1) formula, the priori of the most whole training set is general Rate and the prior probability of nearest w frame.γ is then a control coefrficient, and this control coefrficient is with the training of detection-phase Collect relevant.Work as P'(s | c) > Tmax, detect whether the most at present into area of skin color, so far to extract hand region, Otherwise it is not belonging to area of skin color.
1.2 colours of skin are followed the tracks of
Owing to being tracked just for hand region, i.e. monotrack, the present embodiment have employed Camshift Algorithm, what this algorithm was the most highly developed at present applies in motion tracking and image segmentation, and it is by video All frame numbers make meanShift computing, and meanShift algorithm is the gradient ascent algorithm of a variable step, and will The size of the result of previous frame, i.e. search window and center, as at the beginning of next frame meanshift algorithm search window Initial value.With this, iteration continues, so that it may realizes the tracking to target.Algorithmic procedure is as follows:
(1) search window is initialized;(2) color probability distribution of search window is calculated;(3) meanShift is run Algorithm, it is thus achieved that the size and location of search window;(4) in next frame video image, the initial of step (3) is used Value reinitializes the size and location of search window, then jumps to step (2) and proceed.
1.3 find finger tip
After taking out arm, task below is intended to find finger tip position in the picture, the method finding finger tip Have a variety of, such as the approximation K curvature by calculating profile, take extreme value according to K curvature at finger tip point and obtain finger Point position. the present invention is by following steps searching finger tip:
1) find out largest contours and fill this profile, thus having obtained not having noisy arm foreground image;
2) convex closure of profile is calculated;
3) calculate arm position of centre of gravity, convex closure is found out several candidate points of maximum curvature;
4) using candidate point farthest for distance position of centre of gravity as finger tip.
1.4 range finding
After finding fingertip location, need the distance obtaining finger tip to camera.According to solid geometry principle, first The demarcation of projector and video camera is carried out during the work to be done.Accurately and simple calibration process be projector- Video camera constitutes active vision system and carries out the key point of three-dimensional measurement, and calibrating camera method becomes the most very much Ripe, the most commonly used is the scaling method of Zhang Zhengyou.Calibrating camera has only to use a plane chessboard Lattice.As long as finding the projection just can be in the hope of as the corresponding relation between the two-dimensional points of three-dimensional coordinate point and projection picture Solve inner parameter and the external parameter of projector.
Following steps are used to carry out labeling projection instrument in the present embodiment:
S1. calibrating camera;
S2. prepare a blank, blank posts a papery gridiron pattern;
S3. control projector projects to go out a gridiron pattern and be radiated on blank;
Extract two tessellated angle points the most respectively;
S5. blank place plane is calculated according to papery X-comers;
S6. the video camera demarcated is used to calculate the three-dimensional coordinate of projection X-comers;
The original image of the projection S7. combining angle point three-dimensional coordinate point and projector calculates outside the internal participation of projector Portion's parameter;
The inside that can obtain projector through above step participates in external parameter.Just triangulation can be utilized Method calculates the finger tip distance to camera, as shown in Figure 3.Wherein P is observation station, Ol、OrFor camera, T Being the distance between two cameras, Z is two cameras vertical dimensions to observation station, and f is the focal length of camera, xr And xlRepresent observation station horizontal coordinate in the picture respectively.Represent two camera shootings respectively The horizontal coordinate of picture centre.
Utilize similar triangles can be easy to derive Z value, as shown in Figure 3:
T - ( x l - x r ) Z - f = T Z
⇒ Z = f T x l - x r - - - ( 4 )
Utilize the principle of triangulation, it is also possible to calculate the distance of camera distance projection screen easily, thus The distance of finger distance screen can be extrapolated, if the distance of finger distance screen is less than a certain specific threshold, It is considered as click event to occur.By finger tip position in screen, and geometric calibration before, permissible At cursor of mouse location to finger tip, and analog mouse clicks on event.Realize interpersonal alternately, thus reach Any one projection plane is become the purpose of a touch screen.
The present invention uses the method for " Bayesian Estimation " to separate arm, and motion tracking system is joined interaction In optical projection system so that the efficiency of algorithm is improved.One new data training mechanism is proposed again;And During Face Detection, real-time renewal complexion model so that robustness and the efficiency of algorithm are all greatly improved.Simultaneously The present invention is by the checking of test of many times, and result shows that this system processing speed is fast, reliable and stable.
Obviously, the above embodiment of the present invention is only for clearly demonstrating example of the present invention, and not It it is the restriction to embodiments of the present invention.For those of ordinary skill in the field, in described above On the basis of can also make other changes in different forms.Here without also cannot be to all of enforcement Mode gives exhaustive.All any amendment, equivalent and improvement made within the spirit and principles in the present invention Deng, within should be included in the protection domain of the claims in the present invention.

Claims (10)

1. an interaction method based on active vision, projection arrangement arrives needing interactive image projection Arbitrary plane, and use the operation information of the filming apparatus collection user interactive picture to being projected out, will clap simultaneously The information taken the photograph is transferred to processing means and is analyzed processing, and realizes man-machine interaction with the operation obtaining user corresponding,
It is characterized in that, the process of described processing means Treatment Analysis information is: select one to illumination-insensitive Color space, use Bayes' assessment to extract hand region, and this hand region be tracked and finger tip The location of position, then judge whether finger contacts with projection screen by range finding, man-machine to determine whether to carry out Alternately;
Using Bayes' assessment to extract hand region, its concrete mode is:
By off-line training acquisition probability P (s), P (c), P (c | s) and P (s | c), wherein P (s) represents training During the prior probability of the colour of skin, P (c) represents the prior probability of each color during training, P (c | s) table Show the prior probability that pixel is the colour of skin that pixel value is c, and P (s | c) represent that pixel value is c's after training Pixel is the probability of the colour of skin;
P (s | c)=P (c | s) P (s)/P (c) (1)
Hysteresis threshold T is obtained during off-line trainingmax, Tmax=P (s | c);
Adaptive method is used to carry out Face Detection to extract hand region, according to the colour of skin district of nearest w frame identification Territory judges the area of skin color point detected at present, and the prior probability of nearest w frame is Pw(s), Pw(c), Pw(c | s), Wherein PwS () represents the prior probability of the colour of skin, P in nearest w framewC () represents each color in nearest w frame Prior probability, Pw(c | s) represent that in nearest w frame, pixel value is the prior probability that pixel is the colour of skin of c, separately examines The probability P that survey is area of skin color ' (s | c) be:
P'(s | and c)=γ P (s | c)+(1-γ) Pw(s|c) (2)
Wherein Pw(s | c) obtained by formula (1), i.e. Pw(s | c)=Pw(c|s)Pw(s)/PwC (), γ is one Control coefrficient, this control coefrficient is relevant with the training set of training stage;
Work as P'(s | c) > Tmax, detect whether the most at present into area of skin color, so far to extract hand region, otherwise It is not belonging to area of skin color.
Interaction method based on active vision the most according to claim 1, described color space is adopted Use YUV4:2:2.
Interaction method based on active vision the most according to claim 2, described YUV4:2:2 Color space does not use Y-component.
Interaction method based on active vision the most according to claim 1, described off-line training Detailed process is: selects one group of a small amount of data to be trained, then uses mass data and hysteresis threshold to enter The identification of row area of skin color, and real-time renewal prior probability P (s), P (c) and P (c | s), the most constantly By the hysteresis threshold after updating, the area of skin color in picture and non-area of skin color are separated.
Interaction method based on active vision the most according to claim 1, described to hand region The implementation being tracked is: use Camshift algorithm, will the result of previous frame, i.e. search window Size and center, as the initial value of next frame meanshift algorithm search window;With this, iteration continues, it is achieved Tracking to target.
The most according to claim 5, interaction method based on active vision, described enter hand region The specific implementation of line trace is:
1) search window is initialized;
2) color probability distribution of search window is calculated;
3) meanShift algorithm is run, it is thus achieved that the size and location of search window;
4) by step 3 in next frame video image) initial value reinitialize size and the position of search window Put, then jump to step 2) proceed.
Interaction method based on active vision the most according to claim 1, finger tip position, described location The mode put is:
A) find out largest contours and fill this profile, to have obtained not having noisy arm foreground image;
B) convex closure of profile is calculated;
C) calculate arm position of centre of gravity, convex closure is found out several candidate points of maximum curvature;
D) using candidate point farthest for distance position of centre of gravity as finger tip.
Interaction method based on active vision the most according to claim 1, described range finding is to measure Finger tip to the distance of camera, concrete mode is: labeling projection instrument and video camera, the wherein scaling method of projector For:
S1. calibrating camera;
S2. prepare a blank, blank posts a papery gridiron pattern;
S3. control projector projects to go out a gridiron pattern and be radiated on blank;
Extract two tessellated angle points the most respectively;
S5. blank place plane is calculated according to papery X-comers;
S6. the video camera demarcated is used to calculate the three-dimensional coordinate of projection X-comers;
S7. the original image combining the three-dimensional coordinate point of angle point and the projection of projector calculates the internal participation of projector External parameter;
S8. the method utilizing triangulation calculates finger tip distance Z to camera;When distance Z is less than a certain specific Threshold value, then it is assumed that have click event to occur.
9. the interaction method based on active vision being applied to described in any one of claim 1 to 8 System, including for the interactive image projection projection arrangement to arbitrary plane will be needed, is used for gathering user To the filming apparatus of the operation information of the interactive picture being projected out and the processing means that carries out information process analysis,
It is characterized in that, the process of described processing means Treatment Analysis information is: select one to illumination-insensitive Color space, use Bayes' assessment to extract hand region, and this hand region be tracked and finger tip The location of position, then judge whether finger contacts with projection screen by range finding, man-machine to determine whether to carry out Alternately;
Using Bayes' assessment to extract hand region, its concrete mode is:
By off-line training acquisition probability P (s), P (c), P (c | s) and P (s | c), wherein P (s) represents training During the prior probability of the colour of skin, P (c) represents the prior probability of each color during training, P (c | s) table Show the prior probability that pixel is the colour of skin that pixel value is c, and P (s | c) represent that pixel value is c's after training Pixel is the probability of the colour of skin;
P (s | c)=P (c | s) P (s)/P (c) (1)
Hysteresis threshold T is obtained during off-line trainingmax, Tmax=P (s | c);
Adaptive method is used to carry out Face Detection to extract hand region, according to the colour of skin district of nearest w frame identification Territory judges the area of skin color point detected at present, and the prior probability of nearest w frame is Pw(s), Pw(c), Pw(c | s), Wherein PwS () represents the prior probability of the colour of skin, P in nearest w framewC () represents each color in nearest w frame Prior probability, Pw(c | s) represent that in nearest w frame, pixel value is the prior probability that pixel is the colour of skin of c, separately examines The probability P that survey is area of skin color ' (s | c) be:
P'(s | and c)=γ P (s | c)+(1-γ) Pw(s|c) (2)
Wherein Pw(s | c) obtained by formula (1), i.e. Pw(s | c)=Pw(c|s)Pw(s)/PwC (), γ is one Control coefrficient, this control coefrficient is relevant with the training set of training stage;
Work as P'(s | c) > Tmax, detect whether the most at present into area of skin color, so far to extract hand region, otherwise It is not belonging to area of skin color.
System the most according to claim 9, it is characterised in that described filming apparatus is two cameras.
CN201310724304.8A 2013-12-24 2013-12-24 A kind of interaction method and system based on active vision Active CN103677274B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310724304.8A CN103677274B (en) 2013-12-24 2013-12-24 A kind of interaction method and system based on active vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310724304.8A CN103677274B (en) 2013-12-24 2013-12-24 A kind of interaction method and system based on active vision

Publications (2)

Publication Number Publication Date
CN103677274A CN103677274A (en) 2014-03-26
CN103677274B true CN103677274B (en) 2016-08-24

Family

ID=50315082

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310724304.8A Active CN103677274B (en) 2013-12-24 2013-12-24 A kind of interaction method and system based on active vision

Country Status (1)

Country Link
CN (1) CN103677274B (en)

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104978012B (en) 2014-04-03 2018-03-16 华为技术有限公司 One kind points to exchange method, apparatus and system
CN104202547B (en) * 2014-08-27 2017-10-10 广东威创视讯科技股份有限公司 Method, projection interactive approach and its system of target object are extracted in projected picture
CN104331158B (en) * 2014-10-29 2018-05-25 山东大学 The man-machine interaction method and device of a kind of gesture control
CN104778460B (en) * 2015-04-23 2018-05-04 福州大学 A kind of monocular gesture identification method under complex background and illumination
CN105371784A (en) * 2015-12-24 2016-03-02 吉林大学 Machine vision based holographic man-machine interaction system for automotive inspection
CN105869166B (en) * 2016-03-29 2018-07-10 北方工业大学 A kind of human motion recognition method and system based on binocular vision
CN106204604B (en) * 2016-04-29 2019-04-02 北京仁光科技有限公司 Project touch control display apparatus and its exchange method
CN107343184A (en) * 2016-05-03 2017-11-10 中兴通讯股份有限公司 Projecting apparatus processing method, device and terminal
CN106991417A (en) * 2017-04-25 2017-07-28 华南理工大学 A kind of visual projection's interactive system and exchange method based on pattern-recognition
JP6898021B2 (en) * 2018-03-07 2021-07-07 Necソリューションイノベータ株式会社 Operation input device, operation input method, and program
CN109190357B (en) * 2018-08-30 2021-08-06 袁精侠 Gesture verification code implementation method for man-machine verification by only utilizing cache resources
CN109410274B (en) * 2018-10-08 2022-03-15 武汉工程大学 Method for positioning typical non-cooperative target key points in real time under high frame frequency condition
CN109934105B (en) * 2019-01-30 2022-12-16 华南理工大学 Virtual elevator interaction system and method based on deep learning
CN109683719B (en) * 2019-01-30 2021-10-22 华南理工大学 Visual projection interaction method based on YOLOv3

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101430760A (en) * 2008-11-18 2009-05-13 北方工业大学 Human face super-resolution processing method based on linear and Bayesian probability mixed model
CN102298694A (en) * 2011-06-21 2011-12-28 广东爱科数字科技有限公司 Man-machine interaction identification system applied to remote information service

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101430760A (en) * 2008-11-18 2009-05-13 北方工业大学 Human face super-resolution processing method based on linear and Bayesian probability mixed model
CN102298694A (en) * 2011-06-21 2011-12-28 广东爱科数字科技有限公司 Man-machine interaction identification system applied to remote information service

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
A Bayesian skin/non-skin color classifier using non-parametric density estimation;Chai D,etc;《proceedings of the 2003 international symposium on》;20030628;全文 *
一种肤色干扰下的变形手势跟踪方法;刘玉进等;《计算机工程与应用》;20091211(第35期);全文 *
基于机器视觉的运动目标检测技术的研究;王兴杰;《中国优秀硕士学位论文全文数据库》;20130715;全文 *

Also Published As

Publication number Publication date
CN103677274A (en) 2014-03-26

Similar Documents

Publication Publication Date Title
CN103677274B (en) A kind of interaction method and system based on active vision
US8401248B1 (en) Method and system for measuring emotional and attentional response to dynamic digital media content
CN103941866B (en) Three-dimensional gesture recognizing method based on Kinect depth image
CN104639840B (en) Image processing apparatus and image processing method
CN108446585A (en) Method for tracking target, device, computer equipment and storage medium
CN109684925B (en) Depth image-based human face living body detection method and device
CN111539273A (en) Traffic video background modeling method and system
US20120287122A1 (en) Virtual apparel fitting system and method
Niu et al. Tactic analysis based on real-world ball trajectory in soccer video
CN102831382A (en) Face tracking apparatus and method
Tan et al. Dynamic hand gesture recognition using motion trajectories and key frames
KR20110013200A (en) Identifying method of human attitude and apparatus of the same
CN108198201A (en) A kind of multi-object tracking method, terminal device and storage medium
CN103598870A (en) Optometry method based on depth-image gesture recognition
CN102831439A (en) Gesture tracking method and gesture tracking system
CN104202547A (en) Method for extracting target object in projection picture, projection interaction method and system thereof
CN109145803A (en) Gesture identification method and device, electronic equipment, computer readable storage medium
CN104821010A (en) Binocular-vision-based real-time extraction method and system for three-dimensional hand information
CN111161313A (en) Multi-target tracking method and device in video stream
CN109934127A (en) Pedestrian's recognition and tracking method based on video image and wireless signal
CN111400423B (en) Smart city CIM three-dimensional vehicle pose modeling system based on multi-view geometry
CN113435336A (en) Running intelligent timing system and method based on artificial intelligence
EP4136570A1 (en) Artificial intelligence and computer vision powered driving-performance assessment
CN110517285B (en) Large-scene minimum target tracking based on motion estimation ME-CNN network
CN103761011A (en) Method, system and computing device of virtual touch screen

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CP03 Change of name, title or address

Address after: Kezhu road high tech Industrial Development Zone, Guangzhou city of Guangdong Province, No. 233 510670

Patentee after: Wei Chong group Limited by Share Ltd

Address before: 510663 No. 6, color road, hi tech Industrial Development Zone, Guangdong, Guangzhou, China

Patentee before: Guangdong Weichuangshixun Science and Technology Co., Ltd.

CP03 Change of name, title or address